Etl Consultant Resume
Westborough, MA
SUMMARY
Technical Skills:
ETL: IBM Infosphere Datastage 8.1/7.5.X2 (Server /PX/MVS)
IBM Qualitystage 8.1/7.5.2
IBM Information Analyzer 8.1
Informatica 8.5/8.1/7.1
OLAP: Congnos 8.1, Business Objects XIR2, SAS 9
Data Modeling : Erwin 7/4.5, MS Visio, Embarcadero
Database: Oracle 10G, Teradata, UDB/DB2, SQL Server
Languages: C, C++, JAVA, SQL, PL/SQL, VB, XML, HTML
Operating Systems: AIX 5.3, Solaris 10, Linux, Window XP Pro
Scheduling: Korn Shell, Tivoli, Autosys, Control-M
Knowledge on: Siebel 8.0(SFA), SAP CRM 5.0(Service Agrmnt)
Oracle 11.5.1 (Order Management, TCA)
SAP R/3 ABAP 4.7(IDOC, BAPI), BW 3.5
SAP Netweaver XI/PI 7.0
Professional Summery:
IT Experience: 9 Years of IT experience years of IT experience in Software Analysis, Development, Implementation of business applications for Pharmaceutical, Financial, Insurance
Database Experience: 8 years experience using Oracle 10G/9i/8i/7.x, MS SQL Server 2005/2003, Teradata V2R5/V2R4 (BTEQ, Queryman, MultiLoad, FastLoad, FastExport, Tpump ), MS Access 7.0/2000, SQL, PL/SQL, SQL*Plus
Informatica: 2 years of experience in developing ETL mappings and scripts using Informatica PowerCenter 8.5/8.1/7.1 and PowerMart 6.2/5.1 using Designer (Source Analyzer, Warehouse designer, Mapping designer, Mapplet Designer, Transformation Developer), Repository Manager, Workflow Manager & Workflow Monitor
Datastage: Over 5 Years of IBM Infosphere Data stage 8.0.1 / 7.5.2 / 6.0 (Administrator, Designer, Director and Manager) both SERVER & Parallel Extender/Orchestrate and Multi Clint Manager ( 7.5.2 to 8.0.1 ) using SCD, DTS, Change Capture, Change Apply, Modify, Dataset, DB2EE, Oracle EE, Teradata Enterprise / Multiload, Funnel, Join, Lookup, Merge, MQ Series, Pivot, CFF, Sort, Aggregator, Hashed File, Sequential and debug stages.
Data Cleansing & Standardization: 2 years of Data Cleansing experience (Deduplication, Relationships, Address Validation, Identified, Standardized, Matched, Reconciled) using IBM Datastage, QualityStage/Integrity, Trillium Software, Firstlogic, UNIX Shell Scripting and SQL coding.
Data Profiling: 1 Years of Data profiling experience on IBM Information Analyzer 8.0.1/ Profilestage (validates data values and column/table relationships, source to target field mappings, source system profiling and analysis ) join with SME & Data modeler.
SAP Knowledge: SAP ECC & R/3 ABAP 4.7 (ALV, LSMW, BAPI, IDOC, EDI), SAP BW 3.5 and SAP Netweaver XI 3.0/7.0. SAP CRM 5.0
Education:
MS Business Information Systems (e-commerce, SCM, ERP) from Fh-Wismar, Germany
Professional Experience:
Confidential,Westborough, MA July’2009 - Present
Sr. ETL Consultant (IBM Infosphere Data Stage)
Confidential,Supply Chain DataMart will contain PO info for TJMaxx, Marshalls, Winners, Homesense, TKMaxx HomeSense, TKMaxx Germany, TKMaxx, Homegoods, and AJWright. The PO info will have all Domestic and Import Purchase Orders. It will also contain freight movements and routed orders sourced from the Manugistics Application.
• Design and develop the new ETL job, modify the existing jobs.
• Follow the ETL standards, naming conventions especially for DataStage Project categories, Stage names, links and maintain best practices as per the existing EDW.
• Design Datastage PX jobs that extract, integrate, aggregate, load and transform the data into Data warehouse or data mart.
• Create and reuse metadata and job components.
• Design SCM the data mart dimensional and fact tables data coming from Manugistics.
• Design the jobs using OCI/Oracle EE stage, ODBC Enterprise stage, Lookup stage, Change Capture stage, Sort, Funnel, transformer stage, Peak, Head, Tile stages.
• Loading financial data and Oracle EBS data in Netezza server using Netezza Plug-in
• Writing the custom Netezza SQL when it requires.
• Using IBM Information Analyzer done the Table, Column level, Primary key analysis, cross table analysis.
• Find out the Impact analysis and cardinality changes.
• Created Match Specifications in Match Designer.
• Worked with Standardize stage to standardize specific fields like last name, first name, middle initial, DOB, address, zip, phone no etc…
• Used Reference Match Stage to find the matched, clerical, duplicates and no matched records from master and reference data.
• Extensively used Unduplicated Match Stage to eliminate Duplicates within the dataset/table.
• Extensively used various utilities of Teradata for Fast Load, Multi-Load, T-Pump and BTEQ to populate the Enterprise Data Warehouse.
• Extensively used External Loader to populate the target Teradata.
• Implemented the SURVIVE stage successfully.
• Using Multi Client Manager Support the projects on Datastage 7.5.2/8.0.1.
• Created tables, Index and modify the aggregator tables as per the requirement.
• Prepared the unit test case based on designed and modified jobs.
• Modify the Incremental sequencer that support to modified jobs.
• Read the MQ messages using MQ Connector Stage, design the job as per the logic.
• Implementing the SCD Type-2 using SCD stage & Change Capture stages.
• Worked with Metadata Definitions, Import and Export of Datastage jobs using
• Getting mainframe data and put at Linux box using FTP script.
• Using FTP Plug-in gets mainframe data and load into DB2 tables.
• Work with the TJX Canadian team as part of production support.
• Setup UNIX groups and defined UNIX user profiles and assigned privileges.
• Defined back up recovery process for data stage projects.
• Defined & implemented data stage jobs process monitoring.
• Created crosscheck UNIX shell scripts on interface files and audit reports on data extracted and data loaded, implemented post -Execution scripts to reconcile the data.
Environment: IBM Datastage 7.5.2/8.,DB2/UDB 9.1, Netezza 7.2, SqlServer, Sybase, Oracle EBS, Manugistics SCM, Cognos 8.0, MQ Series, AIX 5.3, Sun Solaris, Windows XP.
Confidential,Durham, NC Feb’2009 – July’2009
Sr.ETL Consultant (IBM Data Stage)
Confidential,develops technological solutions and provide ongoing technology support throughout Duke University Health System and Duke University Medical Center. Duke’s EDW supports Duke’s Health System and subsidiary Health Centers (DUH, DHRH and DRH).
Responsibilities:
• Analyze the existing EDW prepare the mapping documents.
• Design and develop the new ETL job, modify the existing jobs.
• Follow the ETL standards, naming conventions especially for DataStage Project categories, Stage names, links and maintain best practices as per the existing EDW.
• Design Datastage PX jobs that extract, integrate, aggregate, load and transform the data into Datawarehouse or data mart.
• Used PL/SQL to create Packages, Functions, and Procedure.
• PL/SQL Packages to create daily reports in CSV format and emailing these reports to business users using UTL_FILE and UTL_SMTP built in Oracle Packages.
• PL/SQL Packages for daily summarization for Sales and Customer data. This was written as a combination of UNIX and PL/SQL.
• PL/SQL packages for automating lot of manual queries used by business users.
• Create and reuse metadata and job components.
• Design the jobs using OCI/Oracle EE stage, Lookup stage, CDC stage, Sort, Funnel, transformer stage, Peak, Head, Tile stages.
• Extensively hands on experience using DataStage MVS edition stages like Relational, IMS, CFF, Delimited Flat File, Fixed with Flat Files, Multi-format flat files.
• Implementing the SCD Type-2 using SCD stage & Change Capture stages.
• Worked with Metadata Definitions, Import and Export of Datastage jobs using
• Implemented security among data stage users and projects.
• Created crosscheck UNIX shell scripts on interface files and audit reports on data extracted and data loaded, implemented post -Execution scripts to reconcile the data.
• Setup UNIX groups and defined UNIX user profiles and assigned privileges.
• Defined back up recovery process for data stage projects.
• Defined & implemented data stage jobs process monitoring.
Environment: IBM Datastage 7.5.2/7.5.3 (PX/MVS), Sun Solaris, Cognos 8.1, Oracle 10G, SqlServer 2005, TOAD, Tortoise CVS 1.8.3, Erwin.
Confidential,Sunnyvale, CA Mar’2008 – Jan’2009
Sr.ETL Consultant (IBM Infosphere Data Stage)
Confidential,is a computer storage and data management Company. As part of source system upgrade they started “Project V8”. The project must ensure that EDW canned reports and business processes continue to function after the upgrade; users who created their own ad hoc reports will be responsible to make the necessary changes to their reports. This project must ensure that the reports and businesses downstream from the EDW do not break and have correct information that reflects the systems that feed the EDW.
Responsibilities:
• Analyze the existing EDW environment and find out the gaps.
• Using IBM Information Analyzer done the Table, Column level, Primary key analysis, cross table analysis.
• Find out the Impact analysis and cardinality changes.
• Experience in Converting the Business Logic into Technical Specifications.
• Prepare a high level and low level design documents.
• Schedule the meetings with up and down streams.
• Read the JMS Queues, XML files using Datastage.
• Design and develop the new ETL job, modify the existing jobs.
• Follow the ETL standards, naming conventions especially for DataStage Project categories, Stage names, links and maintain best practices as per the existing EDW.
• Tune the Datastage jobs design level and custom SQL scripts.
• Writing and modifying PL/SQL code related to rules engine.
• Used PL/SQL to create Packages, Functions, and Procedure.
• PL/SQL Packages to create daily reports in CSV format and emailing these reports to business users using UTL_FILE and UTL_SMTP built in Oracle Packages.
• PL/SQL Packages for daily summarization for Sales and Customer data. This was written as a combination of UNIX and PL/SQL.
• PL/SQL packages for automating lot of manual queries used by business users.
• Work with different internal teams and offshore team also.
• Read data from Siebel basic tables using Siebel direct plug-in.
• Write data to Siebel basic tables through EIM table using EIM plug-in.
• Prepare mapping document for EIM Tables to Siebel Base Tables.
• Used SAP R/3 PACK (IDOC & BAPI) to extract data from SAP Systems.
• Using SAP BW_EXTRACT PACK to extract data from SAP BW.
• Using SAP BW_LOAD PACK to write data into BW.
• Knowledge on SAP BW creating metadata in BW using Datastage.
• Customized the ABAP Programs while using the ABAP Stages.
• Design the jobs using OCI/Oracle EE stage, Lookup stage, CDC stage, Sort, Funnel and transformer stage.
• Design the batch job when it required.
• Created Match Specifications in Match Designer.
• Worked with Standardize stage to standardize specific fields like last name, first name, middle initial, DOB, address, zip, phone no etc…
• Used Reference Match Stage to find the matched, clerical, duplicates and no matched records from master and reference data.
• Extensively used Unduplicated Match Stage to eliminate Duplicates within the dataset/table.
• Implemented the SURVIVE stage successfully.
• Using Multi Client Manager Support the projects on Datastage 7.5.2/8.0.1.
• Created tables, Index and modify the aggregator tables as per the requirement.
• Prepared the unit test case based on designed and modified jobs.
• Modify the Incremental sequencer that support to modified jobs.
• Maintain the defects by using HP-Quality Center, assign the defects.
• Read the SAP data apply transformation logic and write it at Siebel environment using FTP stage.
• Implemented security among data stage users and projects.
• Set up development, QA & Production environments.
• Migrated jobs from development to QA to Production environments.
• Involved in preparing FSD documentation.
• Defined production support methodologies and strategies.
• Defined back up recovery process for data stage projects.
• Extensively used AUTOSYS scheduling tool to schedule Datastage jobs.
Environment: Windows XP / Sun Solaris, IBM Datastage, Qualitystage & Information Analyzer 8.0.1, Cognos 8.1, OBIEE, Siebel 8.0, Oracle 11i, Oracle CDH, SAP ECC 6.0 my SAP CRM 5.0, SAP Net weaver PI 7.0, TIBCO, Oracle 10g, TOAD 8.0, Erwin
Confidential,Roanoke, VA Aug’2007 – Feb’2008
Sr.ETL Consultant (IBM Infosphere Data Stage)
Confidential,is the second largest automotives aftermarket retailer in USA having 3000 plus stores. AAP plans to start an enterprise data warehouse (EDW) to evaluate their past / present business and estimate their future business plans. It is a retail domain which deals with Customer, Stores, Distributed Channels (Warehouse), Product Delivery Quickly (PDQ) & Merchandise.
Responsibilities:
• Design the ETL jobs based on the DMD with required Tables in the Dev Environment
• Designed and developed Star Schema dimensional model.
• Developed various jobs using Datastage PX stages DB2API / DB2EE stages, Lookup stage, Datasets, Funnel, Duplicate stage, Change Capture stage, Change Apply stage, DRS Stage, ODBC stage.
• Provided production support and customer support to the newly developed data marts and subject areas like Replenishment Stock, Inventory Reduction.
• Applying rules set using Qualitystage to maintain customer information.
• Provide the staging solutions for Data Validation and Cleansing with Quality Stage and Datastage ETL jobs.
• Used the Qualitystage stages like Investigate, Matching stages and setup the rules.
• Used the Quality Stage to standardization and de duplication process
• Validate and standardize the address Data with USPS certified address verification (CASS) and CANDA postal databases (SERP) in Quality stage 8.0.
• Using WAVES standardize and validate the world wide customer address data.
• Using Multi Client Manager I worked on IBM Datastage 7.5.2 and IBM Datastage 8.0.1
• Implementing the SCD Type-2 using SCD stage & Change Capture stages.
• Supporting the existing jobs in Datastage 7.5.2.
• Develop the jobs new EDW environment by using IBM InfoSphere 8.0.1 & IBM Information Analyzer.
• Data extraction from iseriesDB2 database, Oracle, flat files, COBOL flat files
• Implemented Slowly Changing dimension Type- 2 concepts.
• Performance tuning of DB2 target database using explain plan (Access Plan).
• Validation testing and Unit testing using the existing AS/400 required data
• Involved in data profiling and generate reports for Column Analysis, Primary key analysis.
• Validating and compare the source flat file data using Perl script in UNIX box.
• Scheduling the Datastage batch jobs using UC4.
Environment: IBM Datastage/Quality Stage 7.5.2 & IBM Information Server & Information Analyzer 8.0, Crystal Reports, SPSS, Business Objects XIR2, PeopleSoft, JDA , Oracle 10g, Sql Server, DB2UDB 8/ 9.1, IMS DB2, Toad, AIX 5.3, Win XP Pro, UC4.
Confidential,Lexington, KY Aprl’2006 – July’2007
Sr.ETL Consultant (IBM Data Stage)
Confidential, is one of the largest manufacturers of image technology based products like printer, fax both wholesale & retail and deals with distributors.
Atlas is a scalable, worldwide system solution for Lexmark services business transactions that integrate the front-office capabilities of the Siebel application with Lexmark‘s existing back-office platform, JDE.
Responsibilities:
• Responsible for gathering business requirement from end users
• Prepare the Data Mapping Documents and pseudo code
• Design the ETL jobs based on the DMD with required Tables in the Dev Environment.
• Read data from Siebel basic tables using Siebel direct plug-in.
• Write data to Siebel basic tables through EIM table using EIM plug-in.
• Designed and developed Star Schema dimensional model.
• Design and Developed various jobs using Datastage Parallel Extender stages OCI, Hashed file, Sequential file, Aggregator, Pivot and Sort.
• Implemented Slowly Changing dimension concepts.
• Worked with Metadata Definitions, Import and Export of Datastage jobs using Data stage Manager.
• Extensively hands on experience using DataStage MVS edition stages like Relational, CFF, Delimited Flat File, Fixed with Flat Files, Multi-format flat files.
• Used PL/SQL to create Packages, Functions, and Procedure.
• PL/SQL Packages to create daily reports in CSV format and emailing these reports to business users using UTL_FILE and UTL_SMTP built in Oracle Packages.
• PL/SQL Packages for daily summarization for Sales and Customer data. This was written as a combination of UNIX and PL/SQL.
• PL/SQL packages for automating lot of manual queries used by business users.
• Setup UNIX groups and defined UNIX user profiles and assigned privileges.
• Defined back up recovery process for data stage projects.
• Defined & implemented data stage jobs process monitoring.
• Defined K-shell scripts for file watcher and file archiving process.
• Scheduled data stage jobs with TIVOLI.
• Installed packages and patch management.
• Validation testing and Unit testing using the Siebel required data
• Primary contact for business users for UAT testing.
Environment: Ascential Datastage 7.5 ( Server/PX/MVS, Crystal Reports, Micro Strategy, BO XI R2, Oracle 10g, SQL, PL/SQL, Siebel 7.3, JD Edwards, AS/400, Toad, UNIX Shell Scripts, Sun Solaris 8.0, Win XP, VM, Clear case
Confidential,SFO, CA Jun’2005 – Mar’2006
Sr. ETL Consultant (IBM Data Stage)
Confidential,is a diversified financial services company providing banking, insurance, investments, mortgage and consumer finance through almost 6,000 stores, the internet and other distribution channels across North America and internationally. It’s headquartered in San Francisco, but it was decentralized so every local Wells Fargo store is a headquarters for satisfying all our customers\' financial needs and helping them succeeds financially.
Responsibilities:
• Involved in developing Technical Design specification documents for different interfaces.
• Followed Star schema for the methodology to organize the data in the data warehouse.
• Developed several Server and Parallel jobs to improve performance by reducing runtime using different partitioning techniques.
• Performed ETL process using DataStage 7.5 Version.
• Created jobs in DataStage to import data from heterogeneous data sources like Sybase, Oracle 9i, Text files and MS SQL Server.
• Used Datastage Designer to develop jobs for extracting, cleaning, transforming and loading data into data warehouse.
• Used all stages like Sequential files, Hash files, Lookup hash, Join, Funnel, Filter, OCI9i Transformer stage and Sequence Jobs using Job Activity and Mail Notification to send mail notifying the admin about the success or failure of the jobs.
• Read the SAS dataset using SAS Dataset load in to staging environment.
• Extensively used various utilities of Teradata for Fast Load, Multi-Load, T-Pump and BTEQ to populate the Enterprise Data Warehouse.
• Extensively used External Loader to populate the target Teradata.
• Involved in developing custom stages like build ops, Wrapper\'s and Custom stages to replace complex or time consuming jobs for better performance.
• Created Ksh scripts to perform validations and run jobs on different instances (DEV, TEST and PROD).
• Involved in the Analysis of Physical Data Model for ETL Mapping and Processes Flow Diagrams to develop Data Mart Design.
• Wrote stored procedures, functions and packages using TOAD to handle complex business logic and to delete data during the Unit and Functional Testing.
• Compare the source data by implementing the delta method using Perl.
• Extensively used AUTOSYS scheduling tool to schedule Datastage jobs.
• Wrote UNIX shell script to kick off the Datastage jobs.
Environment: Ascential Datastage7.5/7.0/EE, Quality Stage, Oracle9i, PL/SQL, Toad, SQL Server, PeopleSoft, DB2UDB, Teradata, AIX.
Confidential,Omaha, NE Oct’2004 – Mar’2005
Sr.ETL Consultant (IBM Data Stage)
Confidential,is a member of the Physicians Mutual family. It was established in 1970 to provide important life insurance protection to Americans of all ages. Today, Physicians Life is one of largest in life insurance in force, and a variety of products. Physicians Life Insurance Company Annuity Division provides long term savings vehicles for customers and is committed to developing secure and innovative products. The New Data Warehouse project was to completely redesign the Warehouse including all the departments.
Responsibilities:
• Involved in migration process from DEV to Test and then to PRD.
• Obtained detailed understanding of data sources, Flat files and Complex Data Schemas.
• Used Data Stage as an ETL tool to extract data from sources systems and aggregate the data and load into the DB2.
• Used IBM Datastage as an ETL to extract data from sources like Sybase, DB2, VSAM files and flat files and loaded to target DB2.
• Read and write the data using Sybase OC stage and
• Created Re-usable repository using Data stage Manager.
• Designed XML stages for reading XML log files for capturing data stage jobs audit data.
• Installed and configured MQ Series Plug-In and captured On Line messages.
• Developed jobs in Parallel Extender using different stages like Transformation, Aggregation, Source dataset, external filter, Row generation, Column generation and vector stage.
• Created crosscheck UNIX shell scripts on interface files and audit reports on data extracted and data loaded, implemented post -Execution scripts to reconcile the data.
Environment: Ascential DataStage 6.0/7.5.1, DB2, AS/400, Sybase, AIX, Webfocus, Clear Case, Clear Quest, and Cybermation.
Confidential,PA Nov’2003 – Sep’2004
Sr. ETL Consultant (IBM Data Stage)
Merck & Co., Inc. is a leading research-driven pharmaceutical products and services company. Merck discovers, develops, manufactures and markets a broad range of innovative products to improve human and animal health, directly and through its joint ventures. DataStage was used to transform a variety of financial transaction files from different product platforms into standardized data mart. The data mart has been designed using Erwin adopting Star Schema methodology. Models created were based on the dimensions, levels and measures required for analysis.
Responsibilities:
• Created a prototype for PSL to ease the quarterly submissions.
• Used Ascential Datastage as an ETL to extract data from sources like Sybase, DB2, VSAM files and flat files and loaded to target Oracle.
• Implemented Oracle Warehouse Builder & Oracle Bulk Loader for staging area.
• Used lookup stage with reference to Oracle table for insert/update strategy and for updating slowly changing dimensions.
• Used Data Stage Parallel Extender for and then split the data into subsets and to load data, utilized the available processors to achieve job performance, configuration management of system resources in Orchestrate environment.
• Worked with Metadata Definitions, Import and Export of Datastage components using Datastage Manager.
• Extensively hands on experience using DataStage MVS edition stages like Relational, IMS, CFF, Delimited Flat File, Fixed with Flat Files, Multi-format flat files.
• Involved in extracting the data from R3, flat files, SQL Server, Oracle and Loading into Oracle & SAPBW warehouse using Ascential Datastage.
• Used ABAP_EXT_for_R3 and IDOC_EXT_for_R3 to extract the data from SAP R/3
• Used Load & Extract pack for SAP BW.
• Developed SQL scripts for data validation and testing.
• Created UNIX shell scripts using K-shell for extracting and cleaning up the data to load the data in to the target CDW and for scheduling the jobs, Email notification to capture the status of the jobs ran.
• Created Jobs in DataStage to transfer from heterogeneous data source like COBOL, fixed record flat files, CSV files, DB2, Oracle, and Text files to ORACLE 9i.
• Covert data EBCDIC to ASCI using CFF stage
Environment: Ascential Datastage 6.0/7.x(Server/PX/MVS), Web sphere, Quality Stage, Profile Stage, Meta Stage, DB2 UDB 7.0/8.0, Oracle 9.2, PL/SQL, SQL Server 2000, SAP R/3,Show Case, Erwin 4.0, Cognos, IBM AIX 4.2, Rational Clear Case and Rational Clear Quest.