We provide IT Staff Augmentation Services!

Sr. Data Architect/data Modeler Resume

5.00/5 (Submit Your Rating)

Seattle, WA

SUMMARY

  • Over 9+ years of IT experienced in Data Architecture, Data Modeling, designing and Data Analysis with Conceptual, Logical and Physical Modeling for Online Transaction Processing (OLTP), Online Analytical Processing (OLAP) and Data Warehousing. 
  • Experienced in setting up connections to different Databases like Oracle, SQL Server, DB2, NoSQL databases, Teradata and NETEZZA according to users requirement. 
  • Experienced in Data Modeling techniques employing Data warehousing concepts like star/snowflake schema and Extended Star.
  • Experienced in Big Data Hadoop Ecosystem in ingestion, storage, querying, processing and analysis of bigdata.
  • Experience in Data Warehousing and Database Management. Areas of specialization include Data Architecture, Data Modeling (Both Dimensional and Relational Models), Data Analysis, Database Design, Data Federation, Data Integration (ETL), Metadata/Semantic/Universe Design, Static and OLAP/Cube Reporting and Testing
  • Excellent knowledge and experience in OLTP/OLAP System Study with focus on Oracle Hyperion Suite of technology, developing Database Schemas like Star schema and Snowflake schema (Fact Tables, Dimension Tables) used in relational, dimensional and multidimensional modeling, physical and logical Data modeling using Erwin9.6, ER Studio.
  • Experienced in big data analysis and developing data models using Hive, PIG, and Map reduce, SQL with strong data architecting skills designing data - centric solutions.
  • Excellent skills on Oracle, Netezza, Teradata15/14, SQL Server, and DB2 database architecture Netezza, Stored procedure and UDX development with Normalization (1NF, 2NF, 3NF and BCNF) and De-normalization techniques for improved database performance in OLTP and Data Warehouse/Data Mart environments.
  • Have experience on AWS Redshift db design and development and AWS S3 development experience. 
  • Experienced in SQL, PL/SQL package, function, stored procedure, triggers, and materialized view, to implement business logics of oracle database. 
  • Experienced in importing and exporting data from different relational databases like MySQL6.x, Netezza, Oracle into HDFS and Hive using Sqoop
  • Experienced in SAS interacts with third party software such as databases (DB2, SQL SERVER, ORACLE), web application servers, strong knowledge on SAS HPA architecture etc.
  • Experienced in Teradata RDBMS using Fast load, Fast Export, Multi load, T pump, and Teradata SQL Assistance and BTEQ Teradata utilities.
  • Experience in BI/DW solution (ETL, OLAP, Data mart), Informatica, BI Reporting tool like Tableau and Qlikview and also experienced leading the team of application, ETL, BI developers, Testing team.
  • Excellent knowledge of industry standard methodologies like Software Development Life Cycle (SDLC), Iterative Software Development Life Cycle Process, and Agile Scrum methodology.
  • Good experienced in Normalization for OLTP and De-normalization of Entities for Enterprise Data Warehouse. 
  • Extensive experience on usage of ETL & Reporting tools like SQL Server Integration Services (SSIS), SQL Server Reporting Services (SSRS) 
  • Expertise in Data Analysis, Data Migration, Data Profiling, Data Cleansing, Transformation, Integration, Data Import, and Data Export through the use of multiple ETL tools such as Informatica Power Center.
  • Expert knowledge and experience in authoring, managing and deploying reporting solutions using SQL Server Reporting Services (SSRS).
  • Experienced in testing, data validation and writing SQL and PL/SQL statements - Stored Procedures, Functions, Triggers and packages.
  • Extensively experience on EXCEL PIVOT tables to run and analyze the result data set and perform UNIX scripting.
  • Experience with various Online Analytical Processing tools (OLAP) like designing Data Marts and Data Warehouse using Star Schema and Snowflake Schema in implementing Decision Support Systems, fact and dimension tables modeling of data at all the three levels: view, logical and physical. 
  • Experienced in Logical and physical database designing like Tables, Constraints, Index, etc. using Erwin, ER Studio, Power Designer and SQL Modeler.
  • Experienced with SQL Server and T-SQL in constructing Temporary Tables, Table variables, Triggers, user functions, views, Stored Procedures.
  • Heavy use of Access queries, V-Lookup, formulas, Pivot Tables, etc.

TECHNICAL SKILLS:

Analysis and Modeling Tools: Erwin r9.6/r9.5/r9.1/r8.x, Sybase Power Designer, Oracle Designer, BP win ER/Studio, MS Access 2000, Oracle, Star-Schema, Snowflake-Schema Modeling, and FACT and dimension tables, Pivot Tables.

OLAP Tools: Tableau, SAP BO, SSAS, Business Objects, and Crystal Reports 9.

Oracle: Oracle 9i, 10g, 11g, R2 database servers with RAC, ASM, Data Guard, Grid Control and Oracle Golden Gate(Oracle Enterprise Manager),Oracle Data Guard, SQL* Net, SQL Loader and SQL*PLUS, AWR,ASH, ADDM, Explain Plan.

ETL Tools: SSIS, Pentaho, Informatica Power Center 9.7/9.6/9.5/9.1 etc.

Programming Languages: Java, Base SAS and SAS/SQL, SQL, T-SQL, HTML, Java Script, CSS, UNIX shells scripting, PL/SQL.

Database Tools: Microsoft SQL Server 2014/2012/2008/2005 , Teradata, Oracle 12c/11g/10g/9i, and MS Access, Poster SQL, Netezza, SQL Server, Oracle. 

Web technologies: HTML, DHTML, XML, JavaScript

Reporting Tools: Business Objects, Crystal Reports, Tableau

Operating Systems: Microsoft Windows 9x / NT / 2000/XP / Vista/7 and UNIX Windows 95, 98, Windows NT, Windows XP, 7.

Tools & Software: TOAD, MS Office, BTEQ, Teradata 15/14.1/14/13.1/13 , SQL Assistant

Big Data: Hadoop, HDFS 2, Hive, Pig, HBase, Sqoop, Flume.

Other tools: TOAD, SQL PLUS, SQL LOADER, MS Project, MS Visio and MS Office, Have worked on C++, UNIX, PL/SQL etc.

WORK EXPERIENCE

Sr. Data Architect/Data Modeler

Confidential, Seattle, WA

Responsibilities:

  • Assumed leadership role in various divisions of Data Warehouse group such as the Business Analysis (group that defines the data transformation rules), the database architecture (the group that defines the logical and physical architecture), the ETL (with Data stage as the platform) and Business Intelligence (Reporting). This has provided me with the unique experience to acquire all aspects of data management such as data acquisition and transformation, maintenance and delivery.
  • Played key role in defining all aspects of Data Governance - data architecture, data security, master data management, data archival & purging and metadata. 
  • Involved in reviewing, and validating data requirements, including, Data integration (ETL, EAI, and web services), Meta data, Modelling (Conceptual, Logical, and Physical), Data delivery, and Business intelligence. Participate in research and method selection for solutions, being actively involved in design meetings with stakeholders.
  • Working with the architecture and development teams to help choose data-related technologies, design architectures, and model data in a manner that is efficient, scalable, and supportable.
  • Worked in Dimension Data modeling concepts like Star Join Schema Modeling, Snow-Flake Modeling, FACT and Dimensions Tables, Physical and Logical Data Modeling.
  • Worked on NoSQL databases including HBase, Mongo DB, and Cassandra. Implemented multi-datacenter and multi-rack Cassandra cluster. 
  • Developed logical & physical data model using data warehouse methodologies, including Star schema - Star-joined schemas, conformed dimensions data architecture, early/late binding techniques, data modeling, designing & developing ETL applications using Informatica Power Center.
  • Collecting large amounts of log data using Apache Flume and aggregating using PIG/HIVE in HDFS for further analysis. 
  • Created mappings using pushdown optimization to achieve good performance in loading data into Netezza and implemented Physical and Logical standby database and schema replication using Oracle12c streams.
  • Worked with delivery of Data & Analytics applications involving structured and un-structured data on Hadoop based platforms on AWS EMR 
  • Involved in performing extensive Back-End testing by writing SQL queries and PL/SQL stored procedures to extract the data from SQL Database.
  • Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL-like access on Hadoop data 
  • Worked on TERADATA15 and utility domains. Optimization of Queries in a Teradata database environment and worked in using Teradata tools like Fast Load, Multi Load, T Pump, Fast Export, Teradata Parallel Transporter (TPT) and BTEQ. 
  • Used Erwin Data Modeler and Erwin Model Manager to create Conceptual, Logical and Physical data models and maintain the model versions in Model Manager for further enhancements.
  • Worked in Relational Database Concepts, Entity Relation Diagrams, Normalization and De-normalization Concepts. 
  • Successfully loaded files to hive and HDFS from Oracle and Involved in loading data from UNIX file system to HDFS and involved in the validation of the OLAP Unit testing and System Testing of the OLAP Report Functionality and data displayed in the reports.
  • Designed and Developed Oracle12c, PL/SQL Procedures and UNIX Shell Scripts for Data Import/Export and Data Conversions.
  • Used ETL methodology for supporting data extraction, transformations and loading processing, in a complex MDM using Informatica.
  • Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop/Big Data concepts. 
  • Worked on advanced statistical methods of analysis and software packages like Base SAS, SAS/MACROS, SAS/STAT, SAS ENTERPIRSE GUIDE AND MINER in helping and solving different problems and Ad-hoc requests in the Organization.
  • Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Involved in Netezza to Oracle Shell script for loading tables which are required by QA Tools from Netezza in to the Oracle.
  • Involved in documentation of Data Mapping and ETL specifications for Data warehouse and Interacted with QA team in their testing of Data warehouse. 
  • Build and maintain SQL scripts, indexes, and complex queries for data analysis and extraction.
  • Worked extensively on Oracle 12c R2 RMAN cloning for promoting data to lower environments from production. 

Environment: - Erwin 9.6, DB2, Oracle 12c, SQL, PL/SQL, DB2, Teradata15, SQL Server2014, Informatica MDM, UNIX, Netezza, Teradata, BTEQ, MongoDB, Cassandra, SSRS, SSIS, Tableau, Hive, Pig, Mapreduce, T-SQL, Excel, MS Access, Metadata, Aginity, Teradata SQL Assistance, HBase, Hadoop, Big Data.

Sr. Data Architect/ Data Modeler

Confidential, Malvern PA

Responsibilities:

  • Developing strategies for data acquisitions, archive recovery, and implementation of databases and working in a data warehouse environment, which includes data design, database architecture, and Metadata and repository creation.
  • Involved in reviewing business requirements and analyzing data sources form Excel/Oracle SQL Server2012 for design, development, testing, and production rollover of reporting and analysis projects within Tableau Desktop to Netezza database. 
  • Involved in translating business needs into long-term architecture solutions and reviewing object models, data models and metadata.
  • Worked in Implementation of full lifecycle in Data Modeler/Data Analyst, Data warehouses and Data marts with Star Schemas, Snowflake Schemas, and SCD& Dimensional Modeling Erwin r9.5. 
  • Handled importing of data from various data sources, performed transformations using Hive, Map Reduce, loaded data into HDFS and Extracted the data from My SQL into HDFS using Sqoop 
  • Worked with DBA's to create a best fit Physical Data Model from the Logical Data Model using Erwin9.5.
  • Involve in the Analysis, design, testing and Implementation of Business Intelligence solutions using Enterprise Data Warehouse, ETL, OLAP, Client/Server applications.
  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, Hbase database and Sqoop. 
  • Worked in using the IBM Info sphere MDM and loaded the data Used ETL process to Extract, Transform and Load the data into stage area and data warehouse.
  • Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP. 
  • Involved in designing data warehouses and data lakes on regular (Oracle, SQL Server) high performance on big data (Hadoop - Hive and HBase) databases. 
  • Involved in Data Modeling using data Modelers, Star Schema/Snow flake schema, FACT & Dimensions tables, Physical & logical data modeling.
  • Developed Map Reduce jobs in java for data cleaning and preprocessing and developed UDFs in Java as and when necessary to use in PIG and HIVE queries. 
  • Conducted Design discussions and meetings to come out with the appropriate Data Warehouse at the lowest level of grain for each of the Dimensions involved. 
  • Involved in mapping spreadsheets that will provide the Data Warehouse Development (ETL) team with source to target data mapping. 
  • Worked with SQL, SQL PLUS, Oracle PL/SQL Stored Procedures, Triggers, SQL queries and loading data into Data Warehouse/Data Marts.
  • Identified and tracked the slowly changing dimensions (SCD I and II) and determined the hierarchies in Time and Branch Dimensions. 
  • Developed SQL, PL/SQL and ETL scripts on UNIX platform. Ran workflows and monitored jobs using Informatica tools. Performed OLAP and OLTP data modeling, used ERWin9.5.
  • Developed Teradata BTEQ scripts to Load data from Teradata14.1 Staging area to Data Warehouse, Data Warehouse to data marts for specific reporting requirements.
  • Involved in Normalization (3rd normal form), De-normalization (Star Schema for Data Warehousing. 
  • Involved in Communicated with Infrastructure and other teams for better flow of data in SAS through Oracle, MySQL, Hadoop, and raw files etc. 
  • Stored Procedures, User Defined Functions, Views, TSQL scripting for complex business logic. Extensively used Joins and sub-queries for complex queries involving multiple tables from different databases. 
  • Automated the process of reporting which indicated metrics on profitability of different customer groups, profitability of different using SAS/BASE, SAS/MACRO, SAS/STAT, and SAS/GRAPH. 
  • Designed and Developed Hadoop cluster and different Big Data analytic tools including Pig, Hive, HBase, Oozie, ZooKeeper, SQOOP, flume, Spark, Impala, Cassandra with Cloudera Distribution. 
  • Deployed reports to the Report server in Native Mode and on SharePoint Server and implemented Report delivery using Subscription in SSRS
  • Generated comprehensive analytical reports by running SQL queries against current databases to conduct data analysis. 
  • Extensively using Agile methodology as the Organization Standard to implement the data Models. 
  • Used reverse engineering to create Graphical Representation (E-R diagram) and to connect to existing database.
  • Transferred data from various OLTP data sources, such as Oracle, MS Access, MS Excel, Flat files, CSV files into SQL Server 2012. 
  • Extensively worked on UNIX Shell scripting and BTEQs to extract the data from the warehouse.

Environment: - Erwin9.5, UNIX, Teradata14.1, MDM/Activators, Star and Snow Flake, SAS/BASE, SAS/MACRO, SAS/STAT, and SAS/GRAPH, PL/SQL, ETL, DB2, SQL Server2012, Data Stores, DB2, SSRS, SSIS, Tableau, Oracle11g and Netezza, Informatica Power Centre, Hadoop, Hive, Pig, MapReduce, Agile, MongoDB, BigData, MDM, Metadata etc.

Sr. Data Modeler/Data Analyst

Confidential, Southfield, MI

Responsibilities:

  • Involved in data management including Data Modeling, Metadata, Data Analysis, Data mapping and Data Dictionaries, Erwin9.1 and involved in Data Modeling (Oracle/MySQL/Netezza), Data Characterization, Workflow design and implementation 
  • Used Erwin Data Modeler and Erwin Model Manager to create Conceptual, Logical and Physical data models and maintain the model versions in Model Manager for further enhancements. 
  • Worked on logical and physical modeling of various data marts as well as data warehouse using Taradata14
  • Created and maintained Logical Data Model (LDM) for the project. Includes documentation of all entities, attributes, data relationships, primary and foreign key structures, allowed values, codes and glossary terms in accordance with the Corporate Data Dictionary etc. 
  • Implemented Bulk Load Process by converting existing Triggers to Oracle11g packages to improve the Data loading process. 
  • Designed and Developed Oracle11g, PL/SQL Procedures and UNIX Shell Scripts for Data Import/Export and Data Conversions. 
  • Maintained warehouse metadata, naming standards and warehouse standards. 
  • Involved in the validation of the OLAP Unit testing and System Testing of the OLAP Report Functionality and data displayed in the reports. 
  • Worked in using Teradata14.1 tools like Fast Load, Multi Load, T Pump, Fast Export, Teradata Parallel Transporter (TPT) and BTEQ. 
  • Designed the overall ETL solution including analyzing data, preparation of high level, detailed design documents, test plans and deployment strategy.
  • Strong knowledge of Entity-Relationship concept, Facts and dimensions tables, slowly changing dimensions and Dimensional Modeling (Star Schema and Snow Flake Schema).
  • Created Software Development Life Cycle (SDLC) document and excelled in the process of Change Management, Release management and Configuration processes. 
  • Prepared Data Visualization reports for the management using
  • Documented logical, physical, relational and dimensional data models. Designed the Data Marts in dimensional data modeling using star and snowflake schemas. 
  • Involved in Normalization and De-Normalization of existing tables for faster query retrieval. 
  • Create MDM base objects, Landing and Staging tables to follow the comprehensive data model in MDM. 
  • Used ETL methodology for supporting data extraction, transformations and loading processing, in a complex MDM using Informatica
  • Performed Data Validation and Data Cleaning using PROC SORT, PROC FREQ and through various SAS formats. 
  • Designed Data Staging Area and Data Warehouse to integrate the data from various sources including Flat Files to facilitate management to make more fact based decisions. 
  • Created jobs, alerts to run SSIS, SSRS packages periodically. Created the automated processes for the activities such as database backup processes and SSIS, SSRS Packages run sequentially using SQL Server Agent job and windows Scheduler.
  • Perform reverse engineering of physical data models from databases and SQL scripts. 
  • Involved in Normalization (3rd normal form), De-normalization (Star Schema for Data Warehousing
  • Used SSIS to create ETL packages to validate, extract, transform and load data to data warehouse databases, data mart databases to OLAP databases. 
  •   Implemented slowly changing dimensions Type2 and Type3 for accessing history of reference data changes. 
  • Extensively used SAS to query and subset data, summarize and present data, combine tables using joins and merges and created and modified tables. 
  • Utilized SDLC and Agile methodologies such as SCRUM. 
  • Worked in PL/SQL Programming (Stored procedures, Triggers, Packages) using Oracle (SQL, PL/SQL), SQL Server2008 and UNIX shell scripting to perform job scheduling.

Environment: ERwin9.1, Teradata14, Oracle10g, PL/SQL, UNIX, Agile, TIDAL, MDM, ETL, BTEQ, SQL Server2008, Netezza, DB2, SAS, Tableau, UNIX, SSRS, SSIS, T-SQL, MDM, Informatica, SQL.

Sr. Data Modeler/Data Analyst

Confidential, Providence, RI

Responsibilities: 

  • Created physical and logical models and used Erwin9.1 for Dimensional Data Modeling.
  • Designed and Developed Oracle PL/SQL Procedures and UNIX Shell Scripts for Data Import/Export and data Conversions.
  • Performed legacy application data cleansing, data anomaly resolution and developed cleansing rule sets for ongoing cleansing and data synchronization. 
  • Extensively used Star Schema methodologies in building and designing the logical data model into Dimensional Models. 
  • Involved in project cycle plan for the data warehouse, source data analysis, data extraction process, transformation and loading strategy designing. 
  • Worked on Conceptual, Logical Modeling and Physical Database design for OLTP and OLAP systems. 
  • Designed a STAR schema for sales data involving shared dimensions (Conformed) using Erwin Data Modeler. 
  • Designed and build the OLAP cubes using Star schema and Snow Flake Schema using native OLAP Service Manager. 
  • Extensively used Teradata utilities (BTEQ, Fast load, Multiload, TPUMP) to import/export and load the data from oracle, flat files.
  • Performed Data Analysis tasks on warehouses from several sources like Oracle, DB2, and XML etc and generated various reports and documen6ts. 
  • Created Database Maintenance Plans for the performance of SQL Server which covers Database Integrity checks, update database Statistics and Re-indexing. 
  • Involved in workflows and monitored jobs using Informatica tools. Developed SQL, PL/SQL and ETL scripts on UNIX. 
  • Migrated data from SAS environment to SQL Server 2008 via SQL Integration Services (SSIS
  • Used External Loaders like Multi Load, T Pump and Fast Load to load data into Teradata Database, Involved in analysis, development, testing, implementation and deployment. 
  • Used SSIS to create ETL packages to Validate, Extract, Transform and Load data into Data Warehouse and Data Mart. 
  • Merging SAS datasets, Macro facility, SAS Graphs, SAS Formats, SAS Functions, storing and managing data in SAS files 
  • Developed stored procedures on Netezza and SQL server for data manipulation and data warehouse population
  • Actively involved in Normalization (3NF) & De-normalization of database. 
  • Involved in implementing of SQL Server 2008. 
  • Developed multiple processes for Daily Data Ingestion from Client associated data vendors and Production Team, Client site employees using SSIS and SSRS
  • Resolved the data type inconsistencies between the source systems and the target system using the mapping documents and analyzing the database using SQL queries. 
  • Extensively used Extract Transform Loading (ETL) tool of SQL Server to populate data from various data sources 
  • Extensively used SQL Loader to load data from the Legacy systems into Oracle databases using control files and used Oracle External Tables feature to read the data from flat files into Oracle staging tables. 

Environment:- ERwin9.1, Data Modeling, Informatica Power Center9.6, Taradata SQL, PL/SQL, BTEQ, DB2, Oracle, Agile, ETL, Tableau, Cognos, Business Objects, UNIX, SQL Server2008, TOAD, SAS, SSRS, SSIS, T-SQL etc.

Data Analyst/Data Modeler

Confidential

Responsibilities:

  • Data analysis and reporting using MY SQL, MS Power Point, MS Access and SQL assistant.
  • Involved in MY SQL, MS Power Point, MS Access Database design and design new database on Netezza which will have optimized outcome.
  • Used DB2 Adapters to integrate between Oracle database and Microsoft SQL database in order to transfer data 
  • Designed the data marts using the Ralph Kimball's Dimensional Data Mart modeling methodology using ER Studio. 
  • Involved in writing T-SQL, working on SSIS, SSRS, SSAS, Data Cleansing, Data Scrubbing and Data Migration.
  • Used Normalization methods up to 3NF and De-normalization techniques for effective performance in OLTP systems. 
  • Initiated and conducted JAD sessions inviting various teams to finalize the required data fields and their formats. 
  • Involved in designing and implementing the data extraction (XML DATA stream) procedures. 
  • Created base tables, views, and index. Built a complex Oracle procedure in PL/SQL for extract, loading, transforming the data into the warehouse via DBMS Scheduler from the internal data. 
  • Involved in writing scripts for loading data to target data Warehouse using Bteq, Fast Load, MultiLoad 
  • Create ETL scripts using Regular Expressions and custom tools (Informatica, Pentaho, and Sync Sort) to ETL data. 
  • Developed SQL Service Broker to flow and sync of data from MS-I to Microsoft's master database management (MDM). 
  • Extensively involved in Recovery process for capturing the incremental changes in the source systems for updating in the staging area and data warehouse respectively 
  • Strong knowledge of Entity-Relationship concept, Facts and dimensions tables, slowly changing dimensions and Dimensional Modeling (Star Schema and Snow Flake Schema).
  • Involved in loading data between Netezza tables using NZSQL utility. 
  • Worked on Data modeling using Dimensional Data Modeling, Star Schema/Snow Flake schema, and Fact & Dimensional, Physical & Logical data modeling. 
  • Generated Stats pack/AWR reports from Oracle database and analyzed the reports for Oracle8.x wait events, time consuming SQL queries, table space growth, and database growth.

Environment: ER Studio, MY SQL, MS Power Point, MS Access, MY SQL, MS Power Point, MS Access, Netezza, DB2, T-SQL, DTS, SSIS, SSRS, SSAS, ETL, MDM, 3NF and De-normalization, Teradata, Oracle8.x, (Star Schema and Snow Flake Schema) etc.

We'd love your feedback!