Etl Developer /data Stage Resume
Charlotte, NC
SUMMARY:
- Over 8 years of IT experience in Analyzing, Designing, Testing and maintaining client/server Business systems
- Over 6 year of experience in ETL mechanisms using IBM Websphere Information Server V8.x Ascential DataStage7.5 both enterprise and standard edition using components like DataStage Designer, Data Stage Manager, DataStage Administrator, DataStage Director
- Seven years of explicit Health with end - end implementation expertise in Data migration and Data Reporting. Responsibilities included project management, workflow, requirement gathering and analysis, reengineering, development, system testing and support, which included .
- Expertise in areas like Data Cleansing, Data Acquisition, Data Integration, Data Migration and Reporting.
- Extensively involved in the development of DataStage ETL process for extracting data from different data sources, data transformation and loading the data into Data Warehouse.
- Strong SQL and Oracle database knowledge.
- Proficient in building staging environments, data warehouses, data marts.
- Experience in facilitating data requirement meetings with business and technical Stakeholders.
- Strong understanding of the principle Data modeling (Dimensional & Relational) concepts like Star-Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables with extensive implementation on Slowly Changing Dimensions - Type I & II in Dimension tables.
- Expertise in testing GUI Objects, Functional testing, System testing including Integration testing, Retesting & Regression Testing of Web based applications.
- Experienced in onshore - offshore development model.
- Excellent communication, interpersonal, analytical skills and strong ability to perform as part of a team
TECHNICAL SKILLS:
ETL Tools: IBM Info sphere DataStage 9.1, 8.5/8.1/7.5.
Reporting: Crystal Reports & Business Objects 3.1 and 4.0
Languages: PL/SQL, SQL, UNIX shell and XML
Operating Systems: Windows 7/vista/2000/2003, MS-DOS, UNIX, Sun Solaris and Linux
Databases: Oracle 10g/9i/8.x/7x, SQL Server 2000/2005, DB2/UDB & Teradata TechnologiesHTML, JavaScript
Data Modeling ER: Win & Toad
Scheduling Tools: Autosys, Zena Tool
PROFESSIONAL EXPERIENCE:
Confidential, Charlotte, NC
ETL Developer /Data Stage
Responsibilities:
- Worked as an ETLdeveloperfor migrating the code from Mainframes to ETL (Datastage) for CNE application which is to generate the alerts to customers on their day to day transactions, monthly and yearly account reports. Involved in understanding of the business process and coordinate with business analysts and SME's to get the application requirements.
- Developeddatastagejobs according to Cobol program code as part of initial migration.
- Developed thedatastagejobs that include Extract, Transformation and loading of customer data in to the databases and transmit the files to the down streams.
- Supported for ABD project, which use Clearcase tool for version controlling and migrating the code that was checked in Clearcase to the next higher environment.
- Worked on developing pearl script for ABD project to pick the code/packages that were checked in into Clearcase by thedevelopersand deploy those packages in to other environment.
- Also worked in Platform support team, which is responsible to work on any platform related issues.
- Extracted and loaded data from/to various sources like DB2 database, Oracle Database, XML, Flat Files and loaded to target warehouse.
- Used IBM Fast Track and Metadata Workbench for data mapping and data lineage purpose.
- Build a robust and high performance grid platform which can process huge volume of data in minimal time. And reduce number of processes accessing mainframe database for accessing information.
- Designed jobs using different parallel job stages such as Join, Merge, Lookup, Filter, and Dataset, Lookup file Set, Remove Duplicates, Change Data Capture, Modify, Aggregator and file stages such as sequential file, datasets and Complex Flat files.
- Involved in Unit Testing and system testing for the jobs.
- Involved in onsite and offshore coordination to resolve the issues and project tracking with monitoring the tasks to be completed within the project scope
Environment: IBM Information server (DataStage) 8.7(Parallel Extender), Delphix Application, DB2, Autosys, ClearCase, Unix Shell Scripting, Windows 7, SQL Server 2008, Oracle 10g and Putty.
Confidential, Wayne, NJ
Lead ETL Developer /Data Stage
Responsibilities:
- Developed both Server and Parallel Jobs in DataStage Designer to extract data from the Sources Oracle and Complex Flat Files, Cleanse it using Quality Stage, Transform by applying business rules, staging it in Data marts and Load (Initial/Incremental) into Target Teradata.Used the DataStage stages Oracle Enterprise, CFF, Copy, Filter, Lookup, Transformer, Sort, Funnel, Shared Containers, Join, Dataset, Aggregator, Sequential file, Remove Duplicates.
- Used Job Compare option to compare the jobs developed for efficient job design.
- Perform analysis and understand the current and future scope of the application and create process flows and high - level designs.
- Create detail designs for different module after understanding their functionalities and studying the legacy code in UNIX.
- Understand the business requirements and design processes that fulfill these requirements.
- Used the Parameter set to define the parameters at project level in 8.1.1 instead of defining the parameters for each and every job as in 7.5.2 thereby lessening the development effort and reducing the redundant task of adding new parameters at job level.
- Improved the server version jobs performance by using Hash file, In Process and Inter Process settings, Link Partitioner, Link Collector.
- Fine-tune the process to achieve better performance and efficiency.
- Guide the team in theirDataStagetasks and help them follow best practices.
- Design processes that extract data from various sources, perform change capture, transformations, rollupsand load dimension, fact, aggregate and bridge tables.
- Supported shell script creation to automate jobs.
- Review jobs / sequences developed by other team members.
- Write routines to achieve needed functionalities and for process controls. Came up with very innovative ways of designing and developing the jobs that proved highly efficient in processing high volume data.
- Facilitated migration process to move from 8.5 to 8.7.
Environment: Info Sphere DataStage 11.3, 9.1, 8.7, 8.7.2 and 7.5, DB2, Oracle, SQL server, Business Objects 3.1 and UNIX.
Confidential, Kansas City, MO
Lead DataStage Developer
Responsibilities:
- Provided effort estimates for the design and development.Developed DataStage custom jobs as per the business requirements for each module.
- Used Data Stage custom routines, shared containers and open/close run cycles to load the data as per the client standards.
- Performed system testing, integration and UAT.
- Wrote the SQL queries for extraction of data.
- Coordinated with client managers, business architects and data architects for various sign offs on data models, ETL design documents, testing documents, migrations and end user review specifications.
- The scope of the project Involved working on the data that needed to be extracted from the flat files received from Obama HealthCare and loading into the staging (IDS) database and performing ETL on the data which would be further processed before loading it into the (EDW) data warehouse.
- Worked on Facets database extensively.
- Participated in Sprint planning meetings and daily Scrum calls to determine the best architectures, requirements, and designs emerging for the team and then tuned to adjust the target dates.
- Participated in the gap analysis to resolve business and technical issues and resolve conflict, timeline, delivery plans, and other standard project artifacts and deliverables required by the project.
- Arranged meetings with BDAs, BAs and created technical design document for developing DataStage jobs.
- Performed requirements collection, analysis and design of database, documented the same and uploaded to SharePoint.
- Defined various mappings and used DataStage to access different source systems and generated XML files.
- Used Sybase Enterprise stage to extract data from Facets. Scrubbed the data using DataStage Designer Stages such as ODBC, Native plug - in, Sequential File, Remove duplicates, Filter, Aggregator, Transformer, Join, Lookup, XML input, XML output, Sort, Funnel, Dataset, Copy, Modify, Row generator and Merge Stages to generate XML files as per the Edifecs standards.
- Updated the DataStage jobs based on the user request.
- Created sequencers to sequentially execute the designed jobs. Used Execute Command Activity Stage, Notification Activity Stage, job activity, routine activity and sequencer stages to implement these sequencers.
- Involved in optimization of jobs and followed the IBM standard best practices in ETL.
- Imported Data Stage jobs from Production and modified based on updated business requirements.
- Responsible for successful management and facilitation of test plan execution, test case development, end-to-end integration testing, and UAT.
Environment: IBM DataStage 8.7, UDB/DB2, Flat files, CSV files, XML files, SQL Server 2008, Sybase, SQL, PL/SQL, Zena job scheduler, Business Objects 3.1, UNIX, AIX, Putty.
Confidential, Burbank, CA
DataStage Developer
Responsibilities:
- Developed jobs Using Parallel Extender for Extraction, Transformation and Loading (ETL) the data into the data warehouse.Designed the Jobs using the DB2 enterprise stage, Look Up Stage, Merge Stage, Join Stage, Data Sets, Sequential Stage, Row Generator Stage, Transformer Stage and Modify Stage.
- Used DataStage Manager for importing metadata from repository, new job categories and creating new data elements.
- Used the DataStage Designer to design and develop jobs for extracting, cleansing, transforming, integrating, and loading data into different Data Marts.
- Defined the data definitions and created the target tables in the database.
- Mapped the source and target databases by studying the specifications and analyzing the required transforms.
- Wrote the SQL queries & PL/SQL code tables and Data Mart Staging Tables to validate the data results by counting the number of rows in both the tables.
- Developed UNIX shell script to run jobs in multiple instances by using a parameter file.
- Performed Import and Export of DataStage jobs using DataStage Manager to take backups and maintain them in CVS.
- Interacted with Solutions analysts, system staff, ETL, BO developers and the data load staff.
- Involved in Unit, Integration, System and User Acceptance Testing (UAT).
- Using ODBC stage Extracted the data from SQL Server databases, Oracle databases and loaded in to DB2 Tables
- Created Complex SQL Queries and Used in the Jobs.
- Used the inter process stage to connect passive stages as well as active stages to enhance performance
- Used the surrogate key stage to generate surrogate keys on parallel jobs
Environment: IBM Information Server (Datastage) 8.1(Designer, Manager, Director, Administrator), EPIC, Parallel Extender, Oracle 10g, IBM DB2 UDB 8.2/8.0, SQL SERVER 2000, Windows XP, AIX 5.3/5.1, Putty.
Confidential, MD
Datastage Developer
Responsibilities:
- Gathered all the information from existing ETL team and source providers.Participated in business meetings and understood the current job flow and their dependencies.
- Exported jobs from production to development, worked on the issues raised and moved back to the production environment using VSS (Microsoft Visual Source Safe)
- Used Autosys - Scheduling tool to schedule jobs in the production environment based on the pre-conditions and job dependencies.
- Modified DataStagejobs to improve the performance of the job
- Used JIL scripts to create boxes in the autosys environment.
- UsedDataStagedirector to run the jobs manually.
- Worked closely with the mainframe developers in getting the source files, and discussed the data related issues.
- Analyzed the total architecture and documented them for the future .
- Used sequential files, Flat files, oracle database, mainframe files as source and oracle 9i as the target.
- Extensively worked with Parallel Extender for parallel processing to improve job performance while working with bulk data sources.
- Developed jobs in Parallel Extender using different stages like Transformer, Aggregation, dataset, Filter, Lookup, Surrogate key, Sort, Remove Duplicates stages etc.
- Data is loaded from different source files to staging and applied transformations then loaded into ODS.
- Worked closely with Data Analysts to make sure the data is accurate and consistent for table loads.
- Extensively worked on Error handling, cleansing of data, Creating Hash files and performing lookups for faster access of data.
- Involved in performance tuning of the ETL process and performed the data warehouse testing
- Query Optimization for improving the performance of the data warehouse
- Resolved issues occurring during the UAT (user accepting testing) and SIT (System Integration Testing)
Environment: IBM Information Server (DataStage) 8.1.1, Ascential DataStage 7.5.2 Parallel Extender, Profile Stage 7.5.1A, EPIC, Quality Stage 7.5.1A, PL/SQL, Oracle 10g, Solaris 8.9, Windows NT.
Confidential
DataStage Developer
Responsibilities:
- Was part of a team that developed Architecture for building a Data mart by using data modeling tool ErwinInvolved in creating entity relational and dimensional relational data models using Data modeling tool Erwin
- Designing the Target Schema definition and Extraction, Transformation and Loading (ETL) jobs using Data stage
- Developed Analytical jobs that can analyze large amounts of online and offline data using orchestrate environment
- Using Orchestrate, integrated third party applications with inhouse applications
- Used the Data stage Designer to develop processes for extracting, cleansing, transforming, integrating, and loading data into data mart Oracle database
- Worked on programs for scheduling Data loading and transformations using Data Stage from legacy system and Data Warehouse to Oracle 9i
- Developed DataStage Jobs to load the data from the source to various staging levels like Aggregator, Filter, Sort & Merge, Funnel, and Remove Duplicate Stage
- Worked with Data stage Manager for importing metadata from repository, new job, Schema files Categories and creating new data elements
- Involved in Designing Parallel Extender Jobs
Environment: Ascential Data Stage 7.5(Manager, Designer, Director, Administrator), Oracle 9i, DB2, DB2 UDB, SQL, UNIX, Windows NT 4.0.