We provide IT Staff Augmentation Services!

Data Architect / Modeler Resume

4.00/5 (Submit Your Rating)

Atlanta, GA

SUMMARY

  • Highly effective Data Architect with over 10+ years of experience specializing in working with big data, cloud, data and analytics platforms.
  • Experience in Architecture, Design and Development of large Enterprise Data Warehouse (EDW) and Data - marts for target user-base consumption.
  • Excellent knowledge in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
  • Performed data analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata.
  • Excellent knowledge in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
  • Excellent experience on Teradata SQL queries, Teradata Indexes, Utilities such as Mload, Tpump, Fast load and Fast Export.
  • Expert in writing SQL queries and optimizing the queries in Oracle, SQL Server 2008 and Teradata.
  • Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
  • Strong experience in using Excel and MS Access to dump the data and analyze based on business needs.
  • Expertise lies in Data Modeling, Database design and implementation of Oracle, AWS Redshift databases and Administration, Performance tuning etc.
  • Experience in analyzing data using Hadoop Ecosystem including HDFS, Hive, Spark, Spark Streaming, Elastic Search, Kibana, Kafka, HBase, Zookeeper, PIG, Sqoop, Flume.
  • Experienced working with Excel Pivot and VBA macros for various business scenarios.
  • Strong experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export
  • Data Transformation using Pig scripts in AWS EMR, AWS RDS.
  • Experience working with data modeling tools like Erwin, Power Designer and ER Studio.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and from RDBMS to HDFS.
  • Experience in data analysis using Hive, PigLatin, Impala.
  • Well versed in Normalization / De-normalization techniques for optimum performance in relational and dimensional database environments.
  • Good understanding of AWS, big data concepts and Hadoop ecosystem.
  • Experienced in various Teradata utilities like Fastload, Multiload, BTEQ, and Teradata SQL Assistant.
  • Expert in writing SQL queries and optimizing the queries in Oracle, SQL Server 2008 and Teradata.
  • Develop and manage SQL, Python and R code bases for data cleansing and data analysis using Git version control
  • Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
  • Extensive ETL testing experience using Informatica 8.6.1/8.1 (Power Center/ Power Mart) (Designer, Workflow Manager, Workflow Monitor and Server Manager)
  • Excellent in creating various artifacts for projects which include specification documents, data mapping and data analysis documents.
  • An excellent team player& technically strong person who has capability to work with business users, project managers, team leads, architects and peers, thus maintaining healthy environment in the project.

PROFESSIONAL EXPERIENCE

Data Architect / Modeler

Confidential - Atlanta, GA

Responsibilities:

  • Involved in Business requirements and gathering requirements for designing data models, and developing new database.
  • Worked with Project advisors and solution engineers for planning and designing the model.
  • Created Logical and Physical models according to requirements using Power Designer.
  • Conducted and participated in model reviews to validate the upgrades.
  • Used Normal forms to develop logical model.
  • Converted data from old database to newly modeled database.
  • Created Views, stored procedures to generate required reports and dashboard.
  • Developed queries, views and procedures to send out alerts for various informatica based environments through SQL Server Agent.
  • Utilized SQL Server Management tool for generating new database and loading data.
  • Worked in Data Analysis, data profiling and identifying SourceData, Source Meta Data, Data Definitions and Data Formats.
  • Worked closely with Informatica ETL team to explain the complex Data Transformation.
  • Used forward engineering to generate DDL from the Physical Data Model.
  • Apart from Data Modeler experience, worked on Linux and Unix environments to install Oracle client for informatica utilities.
  • Worked on SNOW to extract data and create tickets and incidents for installation purposes.

Environment: SQL Server 2014, Power Designer, Informatica, Teradata Utilities, Tableau, MS Office 2015, MS Excel, SQL Agent, Linux,SNOW.

Data Architect / Modeler

Confidential - Dover, NH

Responsibilities:

  • Collaboratively worked with the Data modeling architects and other data modelers in the team to design the Enterprise Level Standard Data model.
  • Integrate all data in the data management platform enabling accessibility.
  • Involve in data cleaning procedure by removing old, corrupted or irrelevant data in consultation with the teams.
  • Worked with Big Data Hadoop Ecosystem in ingestion, storage, querying, processing and analysis of big data and conventional RDBMS.
  • Use database design and database modeling concepts to ensure data accessibility and security
  • Worked very close with Data Architectures and DBA team to implement data model changes in database in all environments.
  • Responsible for delivering and coordinating data - profiling, data-analysis, data-governance, data-models (conceptual, logical, physical), data-mapping, data-lineage and reference data management.
  • Worked on Data Stage admin activities like creating ODBC connections to various Data sources, Server Start up and shut down, Creating Environmental Variables, Creating Data Stage projects.
  • Participated in all phases of project including Requirement gathering, Architecture, Analysis, Design, Coding, Testing, Documentation and warranty period.
  • Involved in Relational and Dimensional Data modeling for creating Logical and Physical Design of Database and ER Diagrams with all related entities and relationship with each entity based on the rules provided by the business manager using ER Studio.
  • Worked on Normalization and De-normalization concepts and design methodologies like Ralph Kimball and Bill Inmon's Data Warehouse methodology.
  • Worked with Data governance, Data quality, data lineage, Data architect to design various models and processes.
  • Involved in Data Profiling, Data cleansing and make sure the data is accurate and analyzed when it is transferring from OLTP to Data Marts and Data Warehouse.
  • Worked on SQL Server concepts SSIS (SQL Server Integration Services), SSAS (Analysis Services) and SSRS (Reporting Services).
  • Generated and DDL ( Data Definition Language) scripts using ER Studio and assisted DBA in Physical Implementation of Data Models.
  • Extensively worked on creating the migration plan to Amazon web services (AWS).
  • Extracted Mega Data from AWS, and Elastic Search engine using Sql Queries to create reports.
  • Completed enhancement for MDM (Master data management) and suggested the implementation for hybrid MDM (Master Data Management).
  • Exported data from HDFS environment into RDBMS using Sqoop for report generation and visualization purpose.
  • Generated comprehensive analytical reports by running SQL queries against current databases to conduct Data Analysis.
  • Performed Data Analysis, Data Migration and data profiling using complex SQL on various sources systems including Oracle and Teradata
  • Designed and documented Use Cases, Activity Diagrams, Sequence Diagrams, OOD (Object Oriented Design) using UML and Visio.
  • Designed both 3NF data models for ODS, OLTP systems and dimensional data models using Star and Snow Flake Schemas.
  • Used forward engineering to generate DDL from the Physical Data Model and handed it to the DBA.
  • Integrated Spotfire visualization into client's Salesforce environment.
  • Involved in Normalization and De-Normalization of existing tables for faster query retrieval and designed both 3NF data models for ODS, OLTP systems and dimensional data models using star and snow flake Schemas.
  • Involved in Planning, Defining and Designing database using ER Studio on business requirement and provided documentation.
  • Worked with BTEQ to submit Sql statements, import and export data, and generate reports in Teradata.
  • Developed Full life cycle of Data Lake, Data Warehouse with Big data technologies like Spark and Hadoop.
  • Created data masking mappings to mask the sensitive data between production and test environment.
  • Responsible for all metadata relating to the EDW's overall data architecture, descriptions of data objects, access methods and security requirements.
  • Used Agile Methodology of Data Warehouse development using Kanbanize.
  • Worked with DBA group to create Best-Fit Physical Data Model from the Logical Data Model using Forward Engineering.
  • Worked with NoSQL databases like HBase in creating HBase tables to load large sets of semi-structured data coming from various sources.
  • Development of Data stage design concepts, execution, testing and deployment on the client server
  • Developed Linux Shell scripts by using Nzsql/Nzload utilities to load data from flat files to Netezza database.
  • Validated the data of reports by writing SQL queries in PL/SQL Developer against ODS.
  • Involved in user training sessions and assisting in UAT (User Acceptance Testing).

Environment: ER Studio, AWS, OLTP, Teradata r15, Sqoop 1.4, Cassandra 3.11, MongoDB 3.6, HDFS, Linux, Shell, scripts, NoSQL, SSIS, SSAS, HBase 1.2, MDM

Sr. Data Modeler

Confidential - Seattle, WA

Responsibilities:

  • Interacting with business users to analyze the business process and requirements and transforming requirements into.
  • Requirement gathering from the users by conducting a series of meeting with the Business System Users to gather the requirements for reporting.
  • Identification of Risks in schedule and marking them in the Use Cases.
  • Extensively used Star Schema Methodologies in building and designing the logical data model into Dimensional Models.
  • Worked with DBA Group to create Best - Fit Physical Data Model from the Logical Data Model using Forward Engineering using ERWin.
  • Worked with DBA's to create a best-fit physical data model from the logical data model.
  • UsedMongodb for high write load.
  • Created Design Fact & Dimensions Tables using ERWIN tool.
  • Created a Normalized logical model for the business to understand the relationships and at the deeperlevel of granularity.
  • Forward Engineering the Physical Model to generate the DDL scripts to implement on Oracle 11g database.
  • Reverse engineering of the existing reports to identify the key data elements and data entities required to design the data warehouse.
  • Stored data using JSON style documents using Mongodb.
  • Developed Star and Snowflake schemas based dimensional model to develop the data warehouse
  • Worked on slowly changing dimension tables and hierarchies in dimensions.
  • Good working knowledge in normalizing the tables up to 1NF, 2NF & 3NF.
  • Conducted performance tuning of the database that included creating indexes, optimizing SQLstatements.
  • Regularly conducted and participated in weekly status meetings.

Environment: Teradata, SQL Server 2000/ 2005, Erwin 7.5.2, Mongo db, Windows XP, XML, JSON,Excel, Access, Visio, ER Studio 9.7, Tableau 9.03, AWS, Teradata 15, MDM, GIT, Unix, Python 3.5.2, MLlib, SAS, regression, logistic regression.

Data Modeler/Analyst

Confidential - Fremont, CA

Responsibilities:

  • Involved in designing, modelling, maintaining and deploying complex data models, and data integrations utilizing SQL server (SSIS) based ETL processes for new and existing systems
  • Involved in the documentation of Data Modeler/ Data Analyst and ETL specifications for Data warehouse
  • Worked on setting up AWS, DMS and SNS for data transfer and replication
  • Database level Update for any change to Domain values used in SQL Rules, Update the tables
  • Involved in OLAP model based on Dimension and FACTS for efficient loads of data based on Star Schema structure on levels of reports using multi - dimensional models
  • Executed Hive queries on Parquet tables stored in Hive to perform data analysis to meet the business requirements and worked on importing and exporting data from Oracle and DB2 into HDFS and HIVE
  • Performed Extracting, Transforming and Loading (ETL) data from Excel, Flat file, Oracle to MS SQL Server by using SSIS services

We'd love your feedback!