Data Analyst/etl Developer Resume
Elk Grove, IL
PROFESSIONAL SUMMARY:
- 8 years of IT experience in the Data Warehousing, ETL, OLAP, Data Analysis, Design, Development, Implementation, Testing and trouble shooting of Business Intelligence, Data Warehousing, Database applications.
- Working as an ETL Developer for past Five years with strong technical expertise in Ab Initio including designing, developing, implementing Ab Initio graphs in complex, high volume Data Warehousing projects.
- Good idea of Ab Initio graph programming, writing DML, DDL and DCL commands.
- Well versed with various Ab Initio parallelism techniques and implemented Ab Initio Graphs using Data, Component, pipeline parallelism and Multi File System (MFS) techniques.
- Experience in Data Modeling, Data Extraction, Data Migration, Data Integration, Data Testing and Data Warehousing using Ab Initio.
- Hands on Experience using EME for version controls, impact analysis and dependency analysis.
- Clear understanding of various component groups such as Sort, Transform, Round Robin, Join, Rollup, Partitioning and Departitioning components(merge, interleave, gather), Dedup, Scan, Validate, FTP.
- Extensive experience with Data Warehouse and BI methodologies (Kimball - Star, Inmon-Snowflake Schemas), and strong understanding of Relational, Multi Dimensional Database concepts and OLAP Cubes.
- Proficient using in RDBMS like Oracle, SQL Server, Teradata, PL/SQL, SQL *PLUS, SQL *Loader and TOAD.
- Varied experience in writing Stored Procedures, Triggers, and Functions in PL/SQL and T-SQL for performing and supporting data warehouse processes and tasks.
- Experience in working on the Continuous Flow technologies like Database Replication and Message Queuing(Tibco and MQ series).
- Experience in Dynamic Script Generation features namely Parameter Definition Language (PDL) and Component Folding for designing Ab Initio graphs.
- Expertise in preparing code documentation in support of application development, including High level and detailed design documents, unit test specifications, interface specifications, etc.
- Able to interact effectively with other members of the Business Engineering, Quality Assurance, Users and other teams involved with the System Development Life cycle.
- Diverse programming skills: C, C++, Perl, Java, HTML, JSP, Haskell.
- Developed various UNIX shell scripts to run Ab Initio and Data base jobs with Good experience working with very large databases and Performance tuning.
- Dealt and interacted with offshore teams and accomplished the tasks as per the Business requirements.
TECHNICAL SKILLS:
ETL: AB INITIO GDE, CO >Op (1.15/3.0.1, 2.15/3.0.3)
Databases: Teradata V13/V12/V2R6, Oracle 9i,10g/11g, MS SQL Server 2005/2000/7.0/6.5, MS Access 2000
DB Tools: SQL*Plus, SQL*Loader, TOAD 7.6
Modeling Tools: ERWIN, MS Visio
Languages: C,C++,Unix Shell Scripting, Perl, SQL, PL/SQL, Java, Haskell, MATLAB
Data Warehousing: OLTP, OLAP, Normalization, Data Marts, Dimensional Modeling, Facts, Dimensions, Star Schema, Snow Flake
Job Scheduling Tools: Control-M, Autosys, UNIX Crontab
Operating systems: Windows 9x, XP, 2000, UNIX, Linux(Red Hat, Fedora, Ubuntu)
Application Server: Tomcat, Apache, JBOSS
PROFESSIONAL EXPERIENCE:
Confidential, Elk Grove, IL
Data Analyst/ETL Developer
Environment: Ab Initio (GDE 3.0.1 Co>Operating 3.0.3), Application Configuration Environment (ACE- 0.12.3), Oracle 11g, DB2, Teradata v13, XML, Control-M, Pl/Sql, Sql*Loader, Solaris
Responsibilities:
- Involved in Development, Testing, and Migration of ETL process specific to RICEFW for a business object (or functional design depending on project scope) and provide the high-level solution for extracting, Transforming and loading data between the legacy systems and SAP.
- Developed graphs to load data in enterprise wide Data Integration Hub(DIH) as middle layer between SAP and Legacy in a business common format(XML) which will hold PepsiCo enterprise wide master data.
- Worked as a data analyst to configure the mappings part of the iFramework based on the new Ab Initio framework UI product which is responsible for filtering, validating and transforming data (including the use of the cross translations services).
- Responsible for the creation and completion of the Solution Integration Design (SIDs) documents of technical aspects for Field Mapping Documents (FMDs).
- Developed Abinitio Graphs to load data from one or many legacy applications into the Conversion Database COMM table using lookups, expressions, and other transformations.
- Created components (or sub graphs) that are re-usable in AI graphs (requires custom parameters defined on component).
- Extracted data from Legacy Systems (Cobol copy books) transformed into business common format (XML) and Published Idocs to load into SAP tables.
- Involved in development and Testing of Template graphs which involve publishing and subscribing messages on Tibco Queues and Topics, Mq series Queues and implement data sequencing patterns identified for R7 Master Data Interfaces.
- Implemented XFR’s using Cross (lookup) tables from Oracle using smart mapper and Framework user Interface functions.
- Generated PSET’s for template graphs using PDL that encapsulates the transformation logic and gets embedded in the interface templates for Synchronous, Asynchronous, Batch, Event driven, point-to-point interfaces.
- Integrated developed graphs with CLE (Common Logging and Exception Handling) an enterprise wide sub-system maintained on Oracle Database to log messages processed and report exceptions with a capability of replaying the errored out messages back on the Tibco Topics.
- Involved in Extraction, Transformation and Loading of data across different platforms including COBOL copy books, SAP Idocs, XML, Serial Files, Tibco Batched files, Databases.
- Implemented First Logic to Convert Domain Data, Master Data, and Transaction Data from Legacy FLNA systems (Frito Lay United States and Frito Lay Canada) to SAP ECC to enable the PepsiCo Transformation for FLNA.
- Implemented complex SQL queries to update multiple tables in Oracle database continuously to keep the Cross Tables in sync with the conversions master data.
- Collaborated with conversion, Interface and MDM development teams to define roles and responsibilities, work breakdown and approach for effective engagement across teams.
- Involved in entire SDLC (software development life cycle) under different master data divisions of Customers, Locations, Pricing, Promotions, Authorizations of FLNA.
- Created UNIX shell scripts to automate and check pointing on the batch jobs and continuous graphs respectively.
- Worked on different cross systems which involve Tibco Queues and Topics, MQ series Queues, SAP ECC and Legacy Master Files.
- Involved in developing Watch Dog process which monitors the running of the jobs in the projects configured.
- Migrated code between different environments (development, UAT, Production) by generating Manifest Files, Savefiles and FTP service.
- Created Control-M jobs to automate and monitor the developed jobs.
- Well versed with project management tools like Star Team, Borland Caliber, Erooms, HP Quality Center.
Confidential, Atlanta, GA
Ab Initio Developer/Teradata Developer
Environment: Abinitio (GDE 1.15, Co-Operating System 2.15), UNIX shell Scripting, SQL Server, Teradata V12, UNIX, Windows XP
Responsibilities:
- Developed and supported the extraction, transformation and load process (ETL) for a Data Warehouse from their OLTP systems using Ab Initio and provide technical support and hands-on mentoring in the use of Ab Initio.
- Performed analysis, designed and prepared functional, technical design document, and code specifications.
- Responsible for all pre-ETL tasks (Staging phase) upon which the Data Warehouse depends, including managing and collection of various existing data sources.
- Involved in developing UNIX Korn Shell wrappers to run various Ab Initio Scripts.
- Developed Ab Initio XFR’s to derive new fields and solve various business requirements.
- Worked on improving the performance of Ab Initio graphs by using various Ab Initio performance technique’s like using lookups instead of Join’s, Lookup’s, lookup local, in-Memory Joins and rollup’s to speed up various Ab Initio Graphs.
- Extensively used Control-M job scheduling scripts to automate running of the graphs.
- Good understanding of new Ab Initio features like Component Folding, Parameter Definition Language (PDL), Continuous flows, Queues, publisher and subscriber components.
- Extensively used the Teradata utilities like BTEQ, Fastload, Multiload, TPump, DDL Commands and DML Commands (SQL).
- Involved in writing complex SQL queries based on the given requirements and Created series of Teradata Macros for various applications in Teradata SQL Assistant and performed tuning for Teradata SQL statements using Teradata Explain command.
- Worked with DBA team to ensure implementation of the databases for the physical data models intended for the data marts. Created proper Teradata Primary Indexes (PI) taking into consideration of both planned access of data and even distribution of data across all the available AMPS.
- Created UNIX shell scripts to automate and schedule the jobs.
- Created the migration scripts, test scripts for testing the applications, creating and supporting the Business Objects reports.
- 24*7 production support includes monitoring batch jobs, investigating and resolving the problems.
- Involved in the integration testing with ETL developers and User Acceptance Testing (UAT) with Business Analysts.
- Performed physical data modeling, performing regular refreshes of development and test database environments using the Export/Import utility.
Confidential, Atlanta, GA
Sr. Ab Initio Developer
Environment: Ab Initio (GDE1.14, Co>operating system 2.14), Oracle 10g, UNIX, IBM AIX 5.1, Control-M, Unix Shell scripts
Responsibilities:
- Served as primary lead in gathering internal requirements, designing, developing, testing, and supporting AbInitio data movement processes.
- Wrote precise, reusable ETL specifications and patterns to facilitate development of best practices and internal competencies.
- Participated in data model (Logical/Physical) discussions with Data Modelers and created both logical and physical data models.
- Ensured data quality by eliminating the complexity, risk and expense associated with manually checking data by data profiling. Created mapping document and ETL design document.
- Used various Ab Initio components like Call Web service, Read XML, Write XML, xml-to-dml utility for testing. Also did the POC with Ab Initio/Oracle Stored Procedures (PL/SQL) to evaluate the performance.
- Built Ab Initio graphs based on the business requirements that will read from load ready files (CSV files) and load them into the database tables.
- Performed detailed profiling of operational data using Ab Initio Data Profiler/SQL Tool to get better understanding of the data that can be used for analytical purpose for business analysts.
- Well versed with AB Initio parallelism techniques and implemented Ab Initio Graphs using Data parallelism and MFS techniques. Experience in using Conditional Components and Conditional DML.
- Based on business requirements, developed number of AbInitio Graphs using various Components such as Partition by Key, Partition by round robin, reformat, rollup, join, scan, gather, Broadcast, merge etc.
- Implemented Agile Iterative sessions to develop Extended Logical Models and Physical Models.
- Developed various Ab Initio Graphs for data cleansing using Ab Initio function like is valid, is defined, is error, string substring, string concat and other string * functions.
- Involved in the integration testing with ETL developers and User Acceptance Testing (UAT) with Business Analysts.
- Implemented performance tuning on the Ab Initio graphs to reduce the process time.
- Created Oracle PL/SQL functions, Procedures, Korn Shell scripts were used for staging, transformation and loading of the Data into Data mart.
- Automated the entire Data Mart process using Unix Shell scripts and scheduled the process using Autosys after dependency analysis.
Confidential, Denver, CO
Ab Initio Consultant/ETL Designer
Environment: Ab Initio (GDE1.13, Co>operating system 2.13),Teradata V2R5, Oracle 7.3, SQL, PL/SQL,SQL*Plus, SQL*Loader, Erwin.Cognos
Responsibilities:
- ETL and Data Analyst for the creation of Bank Data Mart Implementation which includes CD, MM products for the analytical and reporting use.
- Did the detailed profiling of operational data using Ab Initio Data Profiler/SQL Tool to get better understanding of the data that can be used for analytical purpose for business analysts.
- Participated in several JAD sessions with analysts from business side to come up with the better requirements.
- Created a Bank Data Mart POC at Data Mart level at user Teradata space to validate the requirements with users and also to come up with the better mapping document with right transformations.
- Participated in with the help of BT Project Manager. The iterations are CD, MM, and Integration with Card products using Customer Identification Number, Marketing Campaigns and IVR Call and agent response data.
- Implemented bank data marts in ODS, EDW, DM and ADB (Application Database). Coordinated with Enterprise Warehouse architects to follow Agile Iterative Methodology the corporate standards for the implementation. Used existing Metadata, Audit and ETL frameworks.
- Involved in creation of Logical and Physical models using Erwin for ODS, EDW, DM and ADB and created DDLs for the DBA to create structures in the Teradata environments, development, staging and production.
- Evaluated existing Teradata Industry logical data model (ILDM) related to Financial Services/Banking to be used for Banking Data Mart.
- Created mapping document for all the above 4 environments and ETL design document for the ETL developers to code.
- Created backfill strategy to have past 13 months of data in the Bank Data Warehouse. Created several Bteq, Fastload and Multiload scripts to load backfill data to Data Warehouse.
- Involved heavily in writing complex SQL queries based on the given requirements such as complex Teradata Joins, Stored Procedures, and Macros etc.
- Automated the entire Bank Data Mart process using Unix Shell scripts and scheduled the process using Autosys after dependency analysis.
- Well versed with AB Initio parallelism techniques and implemented Ab Initio Graphs using Data parallelism and MFS techniques.
- Worked extensively to create, schedule, and monitor the workflows and to send the notification messages to the concerned personnel in case of process failures.
- With the help Enterprise Metadata team, uploaded the technical and business metadata to enterprise level Metacenter. Defined audit thresholds for the Balance and control rejections during ETL process.
Confidential, Cincinnati, OH
Ab Initio Developer
Environment: Ab Initio (GDE 1.13, Co>Op Sys 2.13), UNIX, PL/SQL, Oracle 8i, IBM DB2, UNIX
Responsibilities:
- Extensively used Ab-Initio ETL tool in designing & implementing Extract Transformation & Load processes. Different Ab Initio components were used effectively to develop and maintain the database.
- Understood the business requirements with extensive interaction with Business analysts and reporting teams, and assisted in developing the low level design documents.
- Maintained locks on objects while working in the sandbox to maintain the privacy
- Used inquiry and error functions like is valid, is error, is defined and string functions like string substring, string concat and other string * functions in developing Ab Initio graphs to perform data validation and data cleansing.
- Created several packages to set up and share global variables, types and transforms which were extensively used for many Ab Initio graphs.
- Implemented a 6- way multifile system in the test environment that is composed of individual files on different nodes that are partitioned and stored in distributed directories in multifile system.
- Partition Components (Partition by Key, by Expression, by round Robin) were used to Partition the large amount of data file into multiple data files.
- Extensively used File management commands like m ls, m wc, m dump, m copy, m mkfs etc.
- Responsible for deploying Ab Initio graphs and running them through the Co-operating systems mp shell command language and responsible for automating the ETL process through scheduling.
- Involved in Comprehensive end-to-end testing.
- Worked on improving the performance of Ab Initio graphs by using Various Ab Initio performance techniques like using lookup Tables, In-Memory Joins and rollups to speed up various Ab Initio Graphs.
- Implemented phasing and checkpoint approach in ETL process to prevent data loss and to maintain uninterrupted data flow against process failures.
Confidential
Warehouse Developer
Environment: JSP, VB6.0, HTML, DHTML, VBScript, JavaScript, COM, Visual Interdev, ActiveX EXE & DLL, Visio Professional, Oracle 8i
Responsibilities:
- Involved in identifying the activities for the development of I-Shipping Module job.
- Involved in the data transformation using XML and XSL by mapping the data to the scrapping engine.
- Had the responsibility of COM / DCOM, MTS, MSMQ implementation.
- Designed and developed dynamic web pages in HTML, CSS, JSP and DHTML for end users.
- Used PL/SQL codes extensively for the creation of forms and reports and Stored procedures and Functions for application Development
- Designed the DB connectivity using Active-X Data Objects (ADO) extensively for data exchange and maintenance in the Business Objects.
- Setup Net8 connectivity between various computers in a Distributed Database Environment.
- Tuned memory structures such as shared pool buffer cache on Oracle 8i.
- Assisted developers with SQL and stored procedure development.
- Database design and client interaction for database design specifications.