Sr. Data Architect/data Modeler Resume
Washington, DC
SUMMARY
- Over 10+ years of Industry experience as a DataAnalyst with solid understanding of DataModeling, Evaluating DataSources and strong understanding of DataWarehouse/DataMartDesign, ETL, BI, OLAP, Client/Server applications
- Expert in writing SQL queries and optimizing the queries in Oracle, SQLServer, Netezza&Teradata.
- Experienced in handling Big Data using HADOOP eco system components like SQOOP, PIG and HIVE.
- Excellent experience with AWS EC2, EMR, AWS Redshift, AWS S3, AWS Lambda & Microsoft Azure
- Excellent knowledge inDataAnalysis,DataValidation,DataCleansing,DataVerification and identifyingdatamismatch
- Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling
- Enterprisein DataManagementDataGovernance,DataArchitecture,DataModeling, Warehousing,DataIntegration, Meta - data, ReferenceDataand MDM
- Good understanding and hands on experience in setting up and maintaining NoSQL Databases like Cassandra, MongoDB, and HBase.
- Performeddataanalysis anddataprofiling using complex SQL on various sources systems including Oracle, Netezza and Teradata.
- Excellent experience on TeradataSQLqueries, TeradataIndexes, Utilities such as Mload, Tpump, Fastload and FastExport.
- Experienced in SQL, PL/SQL package, function, stored procedure, triggers, and materialized view, to implement business logics of oracle database.
- Strong experience in Normalization(1NF, 2NF, 3NF and BCNF) and Denormalization techniques for effective and optimum performance
- Expertise in designing Star schema, Snowflake schema forDataWarehouse, ODS architecture by using tools like Erwindatamodeler, PowerDesigner, and E-RStudio.
- Well experienced indefining, designing, integrating and re-engineeringthe EnterpriseData warehouseandData Martsin different environments likeTeradata, Oraclewith multipleTerabytes of sizeand various levels of complexity.
- Highly proficient inDataModeling retaining concepts of RDBMS, Logical and PhysicalDataModeling until 3NormalForm (3NF) and MultidimensionalDataModelingSchema (Star schema, Snow-Flake Modeling, Facts and dimensions).
- Expert in developing transactional enterprisedatamodels that strictly meet normalization rules, as well as EnterpriseDataWarehouses using Kimball and InmonDataWarehouse methodologies.
- Solid experience in building ODS, EDW, Staging, data mart, semantic layer and maintaining metadata repository.
- Excellent experience in writing SQLqueries to validatedatamovement between different layers indatawarehouse environment.
- Strong experience in using Excel and MSAccess to dump thedataand analyze based on business needs.
- Experience in generating DDLscripts and creating indexing strategies.
TECHNICAL SKILLS
Analysis and Modeling Tools: Erwin9.6/9.5, Sybase Power Designer
ETL Tools: Informatica Power Center 10.0/9.6.1/9.5 , Ab Initio, Talend, Pentaho
OLAP Tools: MS SQL Analysis Manager, DB2 OLAP, CognosPowerplay, Tableau, TibcoSpotfire
Big Data: Hadoop, HDFS 2, Hive, Pig, HBase, Sqoop, Flume, Spark.
Languages: SQL, PL/SQL, T-SQL, UNIX Shell Scripting, AWK
Databases: MS SQL Server 2016/2015, Oracle12c/11g, DB2 V9.5, Teradata R15/R14, MS Access 2007
Operating Systems: Windows 8/7/XP, UNIX (Sun Solaris 10)
Project Execution Methodologies: Ralph Kimball and Bill Inmon data warehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)
Tools: & Software: TOAD, MS Office, BTEQ, Teradata SQL Assistant, MS-Office suite (Word, Excel, MS Project and Outlook), VSS
Data warehouse Tools: SAP Business Objects XIR3.1/XIR2, Web Intelligence, Talend, Tableau, Pentaho
Reporting Tools: Crystal reports XI, Business Intelligence and SSRS, Business Objects 5.x / 6.x,Cognos7.0/6.0.
PROFESSIONAL EXPERIENCE
Confidential, Washington DC
Sr. Data Architect/Data Modeler
Responsibilities:
- Implemented logical and physical relational database and maintained Database Objects in the data model using Erwin9.6
- Worked with delivery of Data & Analytics applications involving structured and un-structured data on Hadoop based platforms on AWS EMR
- Worked with Sqoop, Flume, Pig, Spark, Oozie and Hive
- Worked with data lineage, data profiler and metadata management tools
- Implemented Cloud Data Solutioning and data stream design, worked with Hadoop, Storm, Hive, Pig, AWS (EMR, Redshift, S3, etc) and Data Lake Design
- Built and managed hosted big data architecture, toolkit such as Hadoop with Oozie, Sqoop, Pig, Hive, Flume, HBase, Avro, HBase, Parkequet, Storm, Spark, NiFi
- Involved in Data Architecture, Data profiling, Data analysis, data mapping and Data architecture artifacts design.
- Interaction with Business Analyst, SMEs and other DataArchitects to understanding Business needs and functionality for various project solutions
- Loaded data into Hive Tables from HadoopDistributedFileSystem (HDFS) to provide SQL access on Hadoopdata
- Designed and Developed Oracle12c, PL/SQL Procedures and UNIXShellScripts for DataImport/Export and Data Conversions.
- Produced and enforced data standards and maintain a repository of data architecture artifacts and procedures.
- Responsible for the data architecturedesign delivery, data model development, review, approval and Datawarehouse implementation.
- Developed and implemented data cleansing, data security, data profiling and data monitoring processes.
- Experience in Datamining with querying and mining large datasets to discover transition patterns and examine financial reports.
- Responsible for technical datagovernance, enterprise wide data modeling and databasedesign.
- Connected to AmazonRedshift through Tableau to extract live data for real time analysis.
- Used StarSchema and SnowflakeSchema methodologies in building and designing the Logical Data Model into Dimensional Models.
Environment: Erwin9.6, Oracle 12c, Hive, ODS, OLAP, OLTP, Hadoop, MapReduce, HDFS, MDM, NoSQL, Business Objects, Agile, Unix, Spark, Cassandra, OLAP, SAP,MS Access, MS Excel.
Confidential, Keller TX
Sr. Data Architect/Data Modeler
Responsibilities:
- Translated the business requirements into logical and physical datamodelusing Infosphere, analyze data across data sources to support requirements, identify integration and security requirements and develop data architecture for supporting IT solution.
- Designed and implemented new jobs in Infosphere to extract, load, and transfer data
- Designed and maintained existing Infosphere jobs to process and analyze quantities of data.
- Involved ineffectively diagnose, isolate, and resolve complex problems pertaining to data infrastructure.
- Provide data architecture support to enterprise data management efforts, such as the development of the enterprise data model and master and reference data, as well as support to projects, such as the development of physical data models, data warehouses and data marts.
- Managing stakeholder’s expectations, and working with business users to gather requirements, resolve business rule requirements, and perform joint conceptual data model reviews.
- DesigningLogical Data Models and Physical Data Models for Relational and multi-dimensional environments.
- Performed or guidedthe team to perform key architectural and development tasks for data modeling, database performance, troubleshooting, ETL specifications, and reporting and security implementation.
- Understandingcomplex data flows, data dependencies and relations, and develop a database design which can support business needs and program and tune database functions, SQL queries, stored procedures, triggers.
- Translate business requirements into conceptual, logical/ canonical and physical data model, model transactional and Data ware housing Models and define and design data integration interface and work with database administrator to create physical database designs
- Develop ascalable architecture for staging and loadingdata using Teradata database
- Maintained effective working relationships with functional owners, peers and vendors for implementation of IT projects
- Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.
- Responsible for modeling of BI solutions including dimensional data marts and operational reporting databases and establish and maintain processes to support and grow data modeling practices
- Conducting strategy and architecture sessions and deliver artifacts such as MDM strategy (Current state, Interim State and Target state) and MDM Architecture (Conceptual, Logical and Physical) at detail level.
- Extensively used AginityNetezzawork bench to perform various DML, DDL etc. operations on Netezzadatabase.
- Utilize a range of existing big data and other DW technologies and data modeling techniques such as star-schema, snowflake, hybrid, and new design techniques optimized for data acquisition, storage and visualization, and evangelizing the data models and techniques.
- TransformedLogical Data ModeltoPhysical Data Modelensuring thePrimary keyandforeign key relationships in PDM, Consistency of definitions of Data Attributes and Primary Index considerations.
- Performeddataanalysis anddataprofiling using complexSQL on various sources systems including Oracle, Teradata and Netezza.
- Good experience in Capital market and Portfolio management.
- Done reporting through the use of the Excel Data mining add-in as cubes are developed, and with R when available in the 2016 SQL Server release.
- Performed Reverse Engineering of the legacy application using DDL scripts in Erwin, and developed Logical and Physical data models for Central Model consolidation.
Environment: IBM Infosphere 11.x, Informatica 9.5, SQL Server 2016, SQL, Oracle 12c, MS Excel, MS Visio, Rational Rose, Netezza, Teradata R15, CSV files, XML files, Cognos, UNIX, AWK, Aginity, Teradata SQL Assistant.
Confidential, Overland Park, KS
Sr. Data Analyst/Data Modeler
Responsibilities:
- Worked ondataanalysis,dataprofiling,datamodeling,datamapping and testing for a given task
- Worked on Creating a Function to Generate the Dynamic WHERE Clause with VDP
- Wrote use test cases and tested Trillium match merged the process
- Worked with DBA to create the physical model and tables.
- Performed various data analysis at the source level and determined the key attributes for designing of Fact and Dimension tables using star schema for an effective Data Warehouse and Data Mart.
- Involved with impact assessment in terms of schedule changes, dependency impact,code changes for various change requests on the existing Data Warehouse applications that running in Production environment.
- Created logical and physical data models using Erwin and reviewed these models with business team and data architecture team.
- Extensively completed data quality management using information steward and did extensive data profiling.
- Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.
- Designed flow diagrams which explain the process flow ofSemanticdevelopment.
- Supported the DBA to physically implement the tables in both oracle and db2 databases.
- Established and maintained comprehensivedatamodel documentation including detailed descriptions of businessentities, attributes, anddatarelationships.
- Developed mapping spreadsheets for ETL team with source to targetdatamapping with physical naming standards,datatypes, volumetric, domain definitions, Transformation Rules and corporate meta-datadefinitions.
- Compiled and maintains the Enterprise Schema across all applications. Enforces principles of good canonical data design.
- Identify data collection, data transformation, normalization, rule administration, error detection and correction, data consolidation, data storage, data distribution, data classification, taxonomy services, item master creation, schema mapping, product codification, data enrichment and data governance through MDM.
- Helped the BI, ETLDevelopers, Project Manager and end users in understanding theDataModel,dataflow and the expected output for each model created
- Gained Comprehensive knowledge and experience in process improvement, normalization/de-normalization,dataextraction,datacleansing,datamanipulation
- Created, documented and maintained logical and physical database models in compliance with enterprise standards and maintained corporate metadata definitions for enterprisedatastores within a metadata repository.
- Perform semantic data modeling for the ELDM, domain logical data models, and physical data model formats.
- Extensively used Erwinfrom ER Studio as the main tool for modeling along with Visio
- Used SQL on a wide scale for analysis, performance tuning and testing
Environment: PDM, SAP, JD Edwards, Teradata 13.10, Microsoft SQL Server 2012, SQL Manager, SAP Logon, Erwin 9.6/8.0, Visio, T-SQL, RDF,Informatica, DB2, Business Objects XI, Teradata SQL Assistant
Confidential, Chicago,IL
Sr. Data Analyst/Data Modeler
Responsibilities:
- Used Erwin, created Conceptual, Logical and Physicaldatamodels.
- Indulge in Power designer tool, Informatica (Repository Manager, Designer and Workflow Manager.
- Worked on using Curator API on Elasticsearch to data back up and restoring.
- Translated logicaldatamodels into physical database models, generated DDLs for DBAs
- Worked withdatainvestigation, discovery and mapping tools to scan every singledatarecord from many sources.
- Performed cross-platform database migration for development & Production databases running on Sun OS to LINUX using utilities such as db2move and db2look.
- Transformed Logical Data Model to Physical Data Model ensuring the Primary Key and Foreign Key relationships in PDM, Consistency of definitions of Data Attributes and Primary Index Considerations.
- Translated business requirements into conceptual, logical/canonicaland physicaldatamodel.
- Identified measures and metrics from requirements, cataloged them in a Report Design Workbook and mapped them to thesemanticlayer anddatastores.
- Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data.
- Implemented Data Archiving strategies to handle the problems with large volumes of data by moving inactive data to another storage location that can be accessed easily.
- Worked closely hand in hand with the Business Analytics manager, who was also a part of the design/data modeling team.
- Implemented performance tuning of SQL queries and MicroStrategy reports.
- Designed both 3NFdatamodels for ODS, OLTP systems and dimensionaldatamodels using star and snowflakeSchemas.
- Collaborated with thedatawarehousing team, ensuring thatdatainfrastructure supports the needs of analytics team and validatingdataquality.
- Written SQLscripts to test the mappings and Developed Traceability Matrix of Business Requirements mapped to Test Scripts to ensure any Change Control in requirements leads to test case update.
- Used CAErwinDataModeler(Erwin) fordatamodeling (datarequirements analysis, database design etc.) of custom developed information systems, including databases of transactional systems anddatamarts.
- Tested the ETL process for both beforedatavalidation and afterdatavalidation process. Tested the messages published by ETLtool anddataloaded into various databases
Environment: Informatica9.5,DataFlux, Erwin 8, Quality Center 9.2, SQL, TOAD, PL/SQL, Flat Files, Teradata, dataFlux, Oracle 12c, SQL, T-SQL, Data Steward, DB2.
Confidential, Menomonee Falls, WI
Sr. Data Analyst/Data Modeler
Responsibilities:
- Designed & Created Test Cases based on the Business requirements (Also referred Source to Target Detailed mapping document & Transformation rules document)
- Customized reports in OBIEE answers and integrated them into the Dashboards according to the Business requirements.
- Involved in Migrating thedatamodel from one database to Teradata database and prepared a Teradata staging model.
- Developed ER and DimensionalModels using PowerDesigner advanced features. Created physical and logicaldatamodels using Power Designer.
- Successfully created and managed a conversion testing effort which included adataquality review, two system test cycles, and user acceptance testing.
- Collecting, aggregating, matching, consolidating, quality-assuring, persisting and distributing such data throughout an organization to ensure consistency trough MDM.
- Wrote complex SQLqueries for validating thedataagainst different kinds of reports generated by Business ObjectsXIR2.
- Worked in importing and cleansing ofdatafrom various sources like Teradata, Oracle, flat files, SQLServer2005 with high volumedata
- Involved in extensiveDatavalidation by writing several complex SQL queries and Involved in back-end testing and worked withdataquality issues.
- Tested the database to check field size validation, check constraints, stored procedures and cross verifying the field size defined within the application with metadata.
- Developed regression test scripts for the application and Involved in metrics gathering, analysis and reporting to concerned team and tested the testing programs.
- Participated in meetings, reviews, and user group discussions as well as communicating with stakeholders and business groups.
Environment: Quality Center 8.6, MS Excel 2007, PL/SQL, T-SQL, Java, Business Objects XIR2, ETL Tools Informatica 9.1, SSIS, Oracle 11G, RDF,Teradata R13, Teradata SQL Assistant
Confidential, Charlotte, NC
Sr. Data Analyst/Data Modeler
Responsibilities:
- Interacting with business users to understand the business requirements.
- Coordinating with Business analysts to understand the business challenges anddatafor reporting.
- Designed Starschema for Risk Retail reporting for credit card portfolio subject area.
- Created and maintained Logical and Physicalmodels for thedatamart, which supports the Credit, Fraud and Risk Retail reporting for credit card portfolio.
- Created partitions and indexes for the tables in thedatamart.
- Maintained metadata (datadefinitions of table structures) and version controlling for thedatamodel.
- Conducted design reviews with business analysts, Enterprisedataarchitect and solution lead to create proof of concept for the reports.
- Developed SQLscripts for creating tables, Sequences, Triggers, views and materialized views.
- Used Ab Initio DQE for data quality solution for enterprise-level data processing and data management systems.
- Worked in importing and cleansing of data from various sources like DB2, Oracle, flat files onto SQL Server 2005 with high volume data
- Developed SQLscripts for loadingdatafrom staging area to target tables.
- Conducted performance analysis and created partitions, indexes and Aggregate tables.
- Performed unit testing, system integrated testing for the aggregate tables.
- Performeddataanalysis on the target tables to make sure thedataas per the business expectations.
- Studied the existing BIarchitecture and analyzed the existing systems to propose the alternate solutions.
- Worked over ETL methodology for supporting data analysis in a corporate-wide-ETL solution using Ab Initio, Informatica.
- Proposed BI solution by migrating the existing reports to Business Objects (crystal reports/Universe design).
- Performed theDataMapping,Datadesign (DataModeling) to integrate thedataacross the multiple databases in to EDW.
- Migrated critical reports using PL/SQL&UNIX packages.
- Involved in the creation of reporting universes in Business Objects.
Environment: Oracle, Toad, Toad Data Modeler 3.4, T-SQL, SQL Plus, Erwin Data Modeler, Erwin Model Manager 7.3, MS Visio, DB2, Source Offsite (SOS), MDM, Steward, Windows XP, QC Explorer, Business Objects, Share point workspace 2010, Oracle SQL Developer
Confidential
Data Analyst / Data Modeler
Responsibilities:
- Designed the dimensional Data Model of the data warehouse.
- Responsible for DimensionalDataModeling and Modeling Diagrams using ERWIN.
- Createddatamodel report for consumers
- Worked with DBA to create Physical Model and tables from the logical model.
- Tuned SQLqueries for better performance.
- Physicaldatamodeling for proposed OLAP and ERWIN.
- Developed conceptual, logical, and physicaldatamodels using PowerDesigner for the Client'sData Layer.
- Involved in preparing Logical and PhysicalDataModels and conducted controlled brain-storming sessions with project focus groups.
- Designed and documented physicaldatamodels using PowerDesigner and Oracle Designer for numerous upgrade releases.
- Conducted JAD sessions with management, vendors, users and other stakeholders for open and pending issues to develop specifications.
- Defined the calculations based on the reporting needs and the sourcedataavailable
- Worked with DBA group to create Best-Fit PhysicalDataModel from the LogicalDataModel using Forward engineering using Erwin
- Worked with cross-functional teams and prepared detailed design documents for production phase of current customer database application.
- Interacting with business users to analyze the business process and requirements and transforming requirements into Conceptual, logical and PhysicalDataModels, designing database, documenting and rolling out the deliverables.
Environment: Erwin 7.5, InformaticaPower Center 7.1.2, Oracle 10g, Sybase, Windows, Toad, PL/SQL, Oracle 9i, SQL Server 2000, SQL*Loader, UNIX, Windows 2000