Sr. Data Architect/data Modeler Resume
Harrisburg, PA
SUMMARY
- Over 10 year of Senior Data Architect/Modeler/Analyst with IT professional experienced in Data Analysis, Data Modeling, Data Architecture, designing, developing, and implementing data models for enterprise - level applications and systems.
- Experienced in integration of various relational and non-relational sources such as DB2, Teradata, Oracle, Netezza, SQL Server, NoSQL, COBOL, XML and Flat Files, to Netezza database.
- Experienced in designing Star Schema, Snowflake schema for Data Warehouse, by using tools like Erwindatamodeler, Power Designer and Embarcadero E-R Studio.
- Experienced in big data analysis and developing data models using Hive, PIG, and Map reduce, SQL with strong data architecting skills designing data-centric solutions.
- Experienced inDatamodeling forDataMart/DataWarehouse development including conceptual,logicaland physicalmodeldesign, developing Entity Relationship Diagram (ERD), reverse/forward engineer (ERD) with CA ERwindatamodeler.
- Experienced in Netezza tools and Utilities NzLoad, NzSql, NzPL/SQL, Sqltoolkits, Analytical functions etc.
- Extensived in Relational and DimensionalDatamodeling for creating Logical and Physical Design of Database and ER Diagrams using multipledatamodeling tools like Erwin, ER Studio.
- Experienced in importing and exporting the data using Sqoop from HDFS to Relational Database systems/mainframe and vice-versa.
- Excellent Knowledge of RalphKimball and BillInmon's approaches toDataWarehousing.
- Experienced in development and support knowledge on Oracle, SQL, PL/SQL,T-SQL queries
- Experienced in LogicalDataModel(LDM) and PhysicalDataModels(PDM) using Erwindatamodeling tool.
- Experienced in migration of Data from Excel, Flat file, Oracle to MS SQL Server by using SQL Server SSIS.
- Experienced in process improvement,Normalization/De-normalization, data extraction, data cleansing, and data manipulation.
- Experienced in ETL design, development and maintenance using Oracle SQL, PL/SQL, TOAD SQL Loader, and Database Management System (RDBMS).
- Experienced in Designed and developed Data models for Database (OLTP), the Operational Data Store (ODS), Data warehouse (OLAP), and federated databases to support client enterprise Information Management Strategy.
- Experienced in data from various sources like Oracle Database, Flat Files, and CSV files and loaded to target warehouse.
- Experienced in Transform, and Load data from heterogeneous data sources to SQL Server using SQL Server Integration Services (SSIS) Packages.
- Hands on experience on tools like R, SQL, SAS and Tableau
- Good knowledge Developing Informatica Mappings, Mapplets, Sessionss, Workflows and Worklets for data loads from various sources such as Oracle, Flat Files, DB2, SQL Server etc.
- Excellent understanding and working experience of industry standard methodologies like System Development Life Cycle (SDLC), as per Rational Unified Process (RUP), AGILE Methodologies.
- Experienced in MDM (Master Data management) in removing duplicates, standardizing data, and to eliminate incorrect data.
- Experienced in Extracting, Transforming and Loading (ETL) data using SSIS/DTS creating.
TECHNICAL SKILLS
Analysis and Modeling Tools: Erwin r9.6/r9.5/r9.1/r8.x, Sybase Power Designer, Oracle Designer, BP win ER/Studio, .1, MS Access 2000, Star-Schema, Snowflake-Schema Modeling, FACT and dimension tables, Pivot Tables.
OLAP Tools: Tableau, SAP BO, SSAS, Business Objects, and Crystal Reports 9.
Oracle: Oracle12c/11g/10g/9i/8.x, R2 database servers with RAC, ASM, Data Guard, Grid Control and Oracle Golden Gate(Oracle Enterprise Manager),Oracle Data Guard, SQL, Net, SQL Loader and SQL PLUS, AWR,ASH, ADDM, Explain Plan.
ETL Tools: SSIS, Pentaho, Informatica Power Center 9.7/9.6/9.5/9.1 etc.
Programming Languages: Java, Base SAS, SSIS and SAS/SQL, SQL, T-SQL, HTML/ XHTML/ HTML4.0.1/ HTML3.2, Java Script, CSS3/CSS2/CSS1, UNIX shells scripting, PL/SQL.
Database Tools: Microsoft SQL Server 2014/2012/2008/2005 , Teradata, and MS Access, Poster SQL, Netezza, SQL Server, Oracle.
Web technologies: Python, HTML, XHTML, DHTML, XML, JavaScript
Reporting Tools: Business Objects, Crystal Reports
Operating Systems: Microsoft Windows 9x / NT / 2000/XP / Vista/7 and UNIX Windows 95, 98, Windows NT, Windows XP, 7.
Tools: & Software: TOAD, MS Office, BTEQ, Teradata 15/14.1/14/13.1/13 , SQL Assistant
Big Data: Hadoop, HDFS 2, Hive, Pig, HBase, Sqoop, Spark.
Other tools: TOAD, SQL PLUS, SQL LOADER, MS Project, MS Visio and MS Office, C++, UNIX, PL/SQL etc.
PROFESSIONAL EXPERIENCE
Sr. Data Architect/Data modeler
Confidential, Harrisburg PA
Responsibilities:
- Provided a consultative approach with business users, asking questions to understand the business need and deriving the data flow, logical, and physical data models based on those needs.
- Developed long term data warehouse roadmap and architectures, designs and builds the data warehouse framework per the roadmap.
- Specifies overall Data Architecture for all areas and domains of the enterprise, including Data Acquisition, ODS, MDM, Data Warehouse, Data Provisioning, ETL, and BI.
- Responsible for leading the architectural development of project/metadata and reporting framework.
- Defining and enabling innovative and advanced technologies to organize, manage and enable complex business analytics capabilities.
- Worked on designing Canonical,Conceptual,Logicaland Physical data modeling using Erwin r9.6, Power Designer and ER Studio in several projects in both OLAP and OLTP applications.
- Responsible for data profiling, data analysis, data specification, data flow mappings, and business logic documentation associated with new or modified product data capture requirements.
- Used Data Stage to Extract, Transform and Load data intoNetezzaData Warehouse from various sources like Oracle12c and flat files.
- Extensively worked across various DBMS like SQL Server 2014, Oracle, Teradata 15 and IBM DB2.
- Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
- Used Informatica Power Center for extraction, loading and transformation (ETL) of data in the data warehouse.
- Worked on the development of Data Warehouse, Business Intelligence architecture that involves data integration and the conversion of data from multiple sources and platforms.
- Designed theLogicalModelinto DimensionalModelusing Star Schema and Snowflake Schema.
- Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data.
- Used SQL tools like Teradata SQL Assistant and TOAD to run SQL queries and validate thedata in warehouse.
- Extensively worked in Oracle environment using Oracle, PL/SQL, SQL developer, SQL Plus, SQL loader, Netezza, UNIX, Toad and Oracle Forms/Reports.
- Performdataprofiling anddataanalysis to enable identifydatagaps and familiarize with new source systemdata
- Worked with DB2 group to create Best-Fit PhysicalDataModelfrom theLogicalDataModelusing Forward engineering using Erwin.
- Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS
- Involved inNormalizationandDe-Normalizationof existing tables for faster query retrieval.
- Developed ETL's for Data Extraction, Data Mapping and data Conversion using SQL, PL/SQL and various shell scripts in Data stage.
- Coordinated with Data Architects and Data Modelers to create new schemas and view in Netezza for to improve reports execution time, worked on creating optimized Data-Mart reports.
- Developed DataMapping, DataGovernance, and transformation and cleansing rules for the MasterDataManagement Architecture involving OLTP, ODS.
- Created Dashboards onTableaufrom different sources using data blending from Oracle, SQL Server, MS Access and CSV at single instance.
- Designed Physical Data Model (PDM) using IBM Info sphere Data Architect data modeling tool and ORACLE PL/SQL.
- Extensively using Agile methodology as the Organization Standard to implement thedataModels.
- Worked with ETLDevelopersin creating External Batches to execute mappings, Mapplets using Informatica workflow designer to integrate Shire's data from varied sources like Oracle, DB2, flat files and SQL databases and loaded into landing tables of InformaticaMDMHub.
- Used in IT to create a PhysicalDataModeler with DDL that best suits the Requirements from the LogicalDataModel.
Environment: ERWIN r9.6, Netezza, Oracle12c, Taradata15, T-SQL, SQL Server 2014, DB2, SSIS, SSRS, R, SAS, HTML, Agile, Python, javascript, UNIX, Tableau, MySQL, Hadoop, Hive, Pig, Map Reduce, Spark, Mongodb, MDM, PL/SQL, ETL etc.
Sr. Data Architect/Data modeler
Confidential, Chicago IL
Responsibilities:
- Designed and build relational database models and defines data requirements to meet the business requirements.
- Developed strategies for data acquisitions, archive recovery, and implementation of databases.
- Responsible for developing and supporting a data model and architecture that supports and enables the overall strategy of expanded data deliverables, services, process optimization and advanced business intelligence.
- Delivered complex and ambiguous business requirements while elevating data architecture to the next level.
- Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
- Provided technical consulting to application development teams during application design and development for highly complex or critical projects.
- Designed data models for mission critical and high volume data management, real-time and distributed data process aligning with the business requirements.
- DevelopedLogicaland PhysicalDatamodelsby using Erwin r9.5 and its converted into physical database design.
- Worked with various RDBMS like Oracle 11g, SQL Server, DB2 UDB, and Teradata 14.1, Netezza.
- Designed and developed SSIS Packages to import and export data from MS Excel, SQL Server 2012 and Flat files.
- Involved in database development by creating Oracle PL/SQL Functions, Procedures and Collections.
- Strong knowledge in Data Warehousing Concepts like Ralph Kimball Methodology, Bill Inmon Methodology, OLAP, OLTP, StarSchema,SnowFlakeSchema, Fact Table and Dimension Table.
- Analyzed the web log data using the HiveQL to extract number of unique visitors per day, page views, visit duration, most purchased product on website.
- Involved in Automating and Scheduling the Teradata SQL Scripts in UNIX using Korn Shell scripting.
- Developed Initio graphs to fetch data from Oracle, Teradata, Flat Files and mainframe Files.
- Involved in Normalizationandde-normalizationOLAP and OLTP systems, process including relational database, table, constraints (Primary key, foreign key, Unique and check) and Indexes.
- Managed and reviewed Hadoop log files.
- Involved in data from different sources like Teradata, Oracle and text files usingSAS/Access,SASSQL procedures and createdSASdatasets.
- Designed Source to Target mapping from primarily Flat files, SQL Server, Oracle,Netezzausing Informatica Power Center.
- Worked with ETLDevelopersin creating External Batches to execute mappings, Mapplets using Informatica workflow designer to integrate Shire's data from varied sources like Oracle, DB2, flat files and SQL databases and loaded into landing tables of InformaticaMDMHub.
- Utilized ODBC for connectivity to Teradata&MSExcel for automating reports and graphical representation ofdatato the Business and OperationalAnalysts.
- Extracteddatafrom existingdatasource, Developing and executing departmental reports for performance and response purposes by using oracle SQL, MSExcel.
- Responsible for creating Hive tables, loading data and writing hive queries.
- Designed Physical Data Model (PDM) using ERwin data modeling tool and PL/SQL and T-SQL Managed Meta-data for data models.
- Worked with the UNIX team and installed TIDAL job scheduler on QA and Production Netezza environment.
Environment: Metadata, Netezza, Oracle11g, Taradata14.1, T-SQL, SQL Server 2012, DB2, SSIS, R, Python, Hadoop, Spark, Map Reduce, UNIX, HTML, Java, MySQL, ERWIN r9.5, MDM, PL/SQL, SPSS, ETL, Informatica Power Center etc.
Sr. Data modeler/Data Analyst
Confidential - Guilford, CT
Responsibilities:
- Created Design Fact & Dimensions Tables,Conceptual, Physical andLogicalDataModelsusing Embarcadero ER Studio.
- Transferred data from various data sources systems including MS Excel, MS Access, Oracle10g and Flat Files to SQL Server, Taradata 14 using SSIS/DTS using various features like data conversion.
- Worked on theNetezzaAdmin Console when the issues were not solved at the session/workflow level.
- Extensively used ER Studio for developingdatamodelusing star schema and Snowflake Schema methodologies.
- Designed and developed complex interfaces with external systems using Oracle PL/SQL.
- Developed LINUX Shell scripts by using NZSQL/NZLOAD utilities to load data from flat files to Netezzadatabase.
- Implemented Dimensionalmodelfor theDataMart and responsible for generating DDL scripts using ER Studio.
- UsedNormalizationas well asDe-normalizationtechniques to process the records depending on the data record structure of the source and target tables.
- PerformedDataMapping between source systems to Target systems,logicaldatamodeling, created class diagrams and ER diagrams and used SQL queries to filterdata.
- UsedNetezzaSQL to maintain MDM frameworks and methodologies in use for the company and also accessedNetezzaenvironment for implementation of ETL solutions.
- Worked extensively with RDBMS systems likeOracleand SQL Server Comfortable in SQL, PL/SQL, Triggers, Stored procedures, Functions, Sequences, and Views etc.
- Involved in debugging and Tuning the PL/SQL code, tuning queries, optimization for theOracle, Taradata, and DB2 database.
- Used the Agile Scrum methodology to build the different phases of Software development life cycle.
- Worked on SSIS and DB2 Packages, DB2 Import/Export for transferring data from Heterogeneous Database (Text format data) to SQL Server.
Environment: E R Studio, Taradata14, Data Modeler, Netezza, Oracle10g, SSAS, T-SQL, Tableau, R, UNIX, HTML, Agile, Java, MySQL Server, DB2, SSIS, MDM, PL/SQL, ETL, etc.
Sr. Data modeler/Data Analyst Data
Confidential - West Des Moines, IA
Responsibilities:
- Worked on designingConceptual,Logicaland Physicaldatamodelsand performed datadesign reviews with the Project team members Erwin r9.1.
- Worked in complex SQLs needed for ETL and analyzing data, and is proficient and worked with databases like Oracle 9i, SQL Server2008, Teradata 13.1, DB2, MS SQL, Excel sheets, Flat Files, Sybase, andNetezza.
- Extensively used Erwin for developingdatamodelusing star schema and Snowflake Schema methodologies.
- UsedNetezzaSQL to maintain ETL frameworks and methodologies in use for the company and also accessedNetezzaenvironment for implementation of ETL solutions.
- Involved in Logical modeling using the Dimensional Modeling techniques such as Star Schema and Snow Flake Schema.
- Worked on theNetezzaAdmin Console when the issues were not solved at the session/workflow level.
- Worked in importing and cleansing of data from various sources like Teradata, Oracle, Netezza flat files, SQL Server with high volume data.
- Created SSIS Reusable Packages to extract data from Multi formatted Flat files, Excel, XML files into Database and Billing Systems
- Involved in PL/SQL programming and developed Stored Procedures and Triggers and Data Stage, DB2, UNIX, Cognos, MDM, UNIX, Hadoop, Pig.
- Extensively worked on Teradata tools & utilities like Fast Load, Multi Load, T Pump, Fast Export, Teradata Parallel Transporter (TPT) and BTEQ.
- Involved inNetezzaAdministration Activities like backup/restore, performance tuning, and Security configuration.
- Utilized SDLC and Agile methodologies such as SCRUM.
- Worked in data from Flat files, SQL Server, Oracle, DB2, and Sybase and load the data into flat files, Oracle and SQL Server using Informatica Power center 9.6.
- Created various transformation procedures by usingSASETL andSASEnterprise guide.
Environment: Netezza, Agile, Oracle9i, Taradata13.1, R, SAS, T-SQL, SQL Server, DB2, SSIS, ERWIN r9.1, MDM, PL/SQL, ETL, Informatica Power center 9.6 etc.
Data Analyst
Confidential
Responsibilities:
- Designed different type of STAR schemas and Snowflake Schema using ERWIN 8.x with various Dimensions like time, services, customers and FACT tables.
- Involved in data from different databases like Oracle8.x, Neteza, Abnitio and external source systems like flat files using ETL tool.
- DesignedStarschema(identification of facts, measures and dimensions) andSnowflake schema.
- Provided technical guidance for re-engineering functions of Teradata13 warehouse operations into Netezza.
- Designed and Developed Oracle PL/SQL and Shell Scripts,DataImport/Export,Data Conversions andDataCleansing.
- Involved in data from different sources like Excel, CRM, SAP,Netezza, and others integrate these and built Dashboard.
- Involved in Integration of various data sources like DB2, Oracle, Netezza, SQL Server, SAP BW, XML Files.
- Assisted in Batch processes using Fast Load, BTEQ, MDM, UNIX Shell and LINUX, Teradata SQL to transfer cleanup and summarize data.
- Worked with requirements management, workflow analysis, source data analysis, data mapping, Metadata management, data quality, testing strategy and maintenance of the model.
- Involved in data from various sources like Oracle Database, XML, Flat Files, CSV files and loaded to target warehouse.
- Designed the ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and physical database (capacity planning, object creation and aggregation strategies) for Oracle and Teradata as per business requirements using Erwin.
- Extensively worked in SQL, PL/SQL, SQL Plus, SQL Loader, Query performance tuning, DDL scripts, database objects like Tables, Views Indexes, Synonyms and Sequences.
Environment: ERWIN8.x, Data Modeler, Netezza, Oracle8.x, SQL, PL/SQL, Taradata13, T-SQL, SQL Server, DB2, SSIS, LINUX, MDM, PL/SQL, ETL, etc.