Lead/senior Datastage Developer Resume
SUMMARY
- 7 years of IT experience in the System Analysis, Architecture, Design, Development, Implementation, Testing and Production support of Database, Data warehousing using Data Modeling, Data Extraction, Data Transformation, Data Loading and Data Analysis.
- Remarkable experience in ETL Development/Administration using IBM Datastage 8.5/8.1/7.5.x(DataStage Manager, DataStage Designer, DataStage Director, Parallel Extender), creating Fact Tables, Dimension Tables using Star Schema Modeling.
- Extensive knowledge in Development, Analysis and Design of ETL methodologies in all the phases of Data Warehousing life cycle.
- Experience on IBM Infosphere (DataStage, Quality Stage) / Information Analyzer 8.1.
- Experience in both Parallel Extender Jobs and Server Jobs in DataStage.
- Played an integral part in the building of a multi-server, multi-database enterprise Data warehouse using DataStage ETL (extact, transform and load) tools and SQL Server to load legacy business data.
- Knowledge of Data Warehouse Architecture and Designing Star Schema, Snow flake Schema, Fact and Dimensional Tables, Physical and Logical Data Modeling using Erwin.
- Hands on experience in Relational database (RDBMS), Oracle, SQL Server, Sybase, DB2 and MS Access.
- Strong Experience in writing PL/SQL, Stored Procedures, Functions and Database Triggers.
- Designed Mapping documents.
- Experience in usage of QualityStage for maintaining and cleansing of data.
- Solid experience in using UNIX and IBM AIX commands and writing UNIX shell scripts.
- Experienced in working with scheduling tool Autosys.
- Extensive experience in loading high volume data and performance tuning.
- Experienced in Quality Assurancecreating for Data Warehousing projects creating Test Plans, Test Objectives, Test Strategies, and Test Cases. Ensuring the Data in data warehouse meets the business requirements.
- Participated in discussions with Project Manager, Business Analysts and Team Members on any technical and/or Business Requirement issues.
- Capable of working under high stress environment with resource constraints.
- Excellent analytical, communication, and facilitation skills with the ability to gain consensus across multiple teams.
EDUCATION
- Bachelor of Technology
TECHNICAL SKILLS:
ETL Tools
IBM InfoSphere (DataStage, QualityStage) & Information Analyzer 8.1, DataStage 7.5.2EE client components (Designer, Director, Manager, Administrator), ProfileStage, Datastage 7.5.2.
BI Tools
Business Objects 6.5, Cognos 7.1 (Impromptu, PowerPlay).
Operating Systems
Windows XP/2000/2003/2007, UNIX, IBM AIX.
Languages
SQL, SQL PLUS, PL/SQL, XML, ASP.Net, C#, JavaScript, HTML 4.0, DHTML, Korn-Shell Scripting, Perl, C, C++, VBScript.
Database
Oracle 11g/10g/9i/8i, SQL Server 2000/2005/2008, Teradata V2R6/V2R5/V2R3, DB2 UDB.
Tools
SQL* Plus 9.2, SQL Loader 9.2, SQL Navigator 4.1, Autosys, Excel, TOAD, Erwin 8.0
PROFESSIONAL SUMMARY:
Confidential, AR Nov`11-Till date
Lead/Senior DataStage Developer
Confidential serves customers and members more than 200 million times per week at more than 9884 retail Units under 60 different banners in 28 Countries. Assortment Planning project is about creating the BAM (Buying Assortment Planning) tool for George Apparel Store. The existing tool Arrange Plan System involved Lot of manual effort for the planner to plan for assortment. The BAM tool is a stabilized tool that was supported by a database at the backend that reduced most of the manual effort involved.
Responsibilities:
- Involved in Business Requirements Analysis and with the stake-holders.
- Worked closely with Subject Matter Experts on the Requirement analysis, Source/Target data analysis.
- Prepared the Technical Design Documents and Lower level Design documents for the technical specs.
- Prepared the Implementation plan for the code migration to QA/Production.
- Co-ordinated and lead the offshore team.
- Migrated projects from 8.1 to 8.5.
- Designed ETL jobs with complete understanding on how GRID works and how to reduce the job wait time in the queue before getting submitted to the grid.
- Created the shell scripts for pre/post processing of the files.
- Process the files received from the business and ensured that the downstream team is able to consume the data
- Worked along with Mainframe Team,used JCL scripts for scheduling the DataStage Jobs and performed
- the End-to-End testing for the jobs.
- Provided code approvals after peer review of the DataStage Jobs, steered Committee meetings and conducted Impact analysis.
- Used CA7 Scheduler to schedule the DataStage jobs.
- Tuned DataStage jobs for better performance to bring design parallelism.
- Used CDC (Change Data Capture) Stage to Capture the New records and updated records and implemented SCD type 2.
- Utilized IS tool Manager for Import/Export.
- Extensively used DataStage Director for Job Scheduling, emailing production support for troubleshooting from LOG files. Collected full summary logs files from the Director.
- Performed end-to-end testing of the flow
- Prepared the Implementation plan for the code migration to QA/Production.
- Supported through Production for successful execution.
- Created the shell scripts for pre/post processing of the files.
- Supported DataStage specific problems on call rotation 24/7.
Environment: GRID, Datastage 8.5/8.1, DB2, Teradata,SQL SERVER 2008, Windows Server 2003,
IBM Rational Software Architect, QueryMan, CA7 Schedular
Confidential., CA May`10-Oct`11 Senior Datastage Developer
Confidential. is the one of the world`s largest Storage Company, the storage security systems deliver non-disruptive, comprehensive protection for sensitive data across your enterprise. ASUP (Auto Support) project is about integrating Auto-support data to the current eBI installed base analytics to provide very useful insight into the workings of Installed systems that NetApp has sold. The Integration of ASUP data with the current eBI analytics of Installed base would provide a much greater depth into the workings of NetApp systems, customer behavior and better identify opportunities for cross-sell up-sell/tech-refresh.
Responsibilities:
- Worked extensively on Data warehousing, extensively used DataStage an ETL tool to design mappings to move data from Source to Target database-using Stages.
- Obtained detailed understanding of data sources, Flat files and Complex Data Schema.
- Designed parallel jobs using various stages like Aggregator, Join, Transformer, Sort, Merge, Filter and Lookup, Sequence, ODBC, Hash file.
- Broadly involved in Data Extraction, Transformation and Loading (ETL process) from Source to target systems using DataStage PX.
- Extensively used DataStage Designer, Administrator, and Director for creating and implementing jobs.
- Created shared containers to use in multiple jobs.
- Involved in creating technical documentation for source to target mapping procedures to facilitate better understanding of the process and incorporate changes as and when necessary.
- Automation of ETL processes using DataStage Job Sequencer and Transform functions.
- Extensively used DataStage Director for Job Scheduling, emailing production support for troubleshooting from LOG files.
- Involved in Performance Tuning on the source and target at DataStage Level and Data Loading.
- Developed PL/SQL stored procedures for source pre load and target pre load to verify the existence of tables.
- Performed unit testing of all monitoring jobs manually and monitored the data to see whether the data is matched.
- Used Log in DataStage Director and Peek stage for debugging.
- Strictly followed the change control methodologies while deploying the code from DEV to QA and Production.
- Assisted both the DEV and Testing OFFSHORE teams.
- Involved in 24/7 ETL Production Support, maintenance, troubleshooting, problem fixing and ongoing enhancements to the Data mart.
Environment: IBM InfoSphere Information Server V8.1.1 Suite [DataStage, QualityStage], Oracle 11g,
PL/SQL, Windows Server 2003, Erwin, TOAD,UNIX Shell Scripting.
Confidential, GA Mar`09-May`10
DataStage Developer
Confidential is a privately-owned business based in Savannah, Georgia. The Company is one of the largest title lending companies in the United States with more than 500 Stores in seven states. We developed the Data Warehouse to store the customer account information, Company’s sales
And profit information. The basic objective of the project is to deliver the clean and complete data for the reporting use. The source data is completely from Oracle, which contains customer details, loan details,
Lender information etc.
Responsibilities:
- Worked extensively on Data warehousing, extensively used DataStage an ETL tool to design mappings to move data from Source to Target database-using Stages.
- Obtained detailed understanding of data sources, Flat files and Complex Data Schema.
- Designed parallel jobs using various stages like Aggregator, Join, Transformer, Sort, Merge, Filter and Lookup, Sequence, ODBC, Hash file.
- Worked extensively on Slowly Changing Dimensions using CDC stage.
- Broadly involved in Data Extraction, Transformation and Loading (ETL process) from Source to target systems using DataStage PX.
- Generating Surrogate ID’s for the dimensions in the fact table for indexed, faster data access.
- To reduce the response time, aggregated the data, data conversion and cleansed the large chunks of data in the process of transformation.
- Involved in creating technical documentation for source to target mapping procedures to facilitate better understanding of the process and incorporate changes as and when necessary.
- Successfully Integrated data across multiple and high volumes of data sources and target applications.
- Automation of ETL processes using DataStage Job Sequencer and Transform functions.
- Extensively used DataStage Director for Job Scheduling, emailing production support for troubleshooting from LOG files.
- Optimized job performance by carrying out Performance Tuning Methods.
- Used Autosys for scheduling the jobs.
- Involved in Performance Tuning on the source and target at DataStage Level and Data Loading.
- Strictly followed the change control methodologies while deploying the code from QA to Production
Environment: IBM InfoSphere Information Server V8.1.1 Suite [DataStage, QualityStage], Oracle 10g,
SQL Server 2008, SQL, PL/SQL, Autosys 4.5, Visio, UNIX Shell Scripting, Erwin,TOAD.
Confidential, NJ Mar’08-Feb’09
DataStage Developer
Confidential is the consumer banking arm of financial services giant Citigroup. This project was to develop Data warehouse for the monthly account details having the information regarding account number, month, branch, primary balance, interest paid, interest charged, and fees charged. This involves extracting data from oracle database, Flat Files, and after necessary cleansing and transformation, integrated data is finally loaded into the target data warehouse.
Responsibilities:
- Analyzed the existing ETL process and came up with an ETL design document that listed the jobs to load, the logic to load and the frequency of load of all the tables.
- Analyzed, designed, developed, implemented and maintained Parallel jobs using Enterprise Edition of DataStage.
- Developed complex jobs using various stages like Lookup, Join, Merge, Sort, Transformer, Dataset, Row Generator, Column Generator, Sequential File and Aggregator Stages.
- Extracted data from disparate sources - relational databases, oracle database, flat files and loaded into data warehouse.
- Converted complex job designs to different job segments and executed through job sequencer for better performance and easy maintenance.
- Extensively worked on Data Acquisition and Data Integration of the source data.
- Worked with Metadata Definitions, Import and Export of DataStage jobs using DataStage Manager.
- Design Mapping documents with Transformation rules.
- Defined projects and tuned parameters for fine-tuning of the projects.
- Defined & implemented DataStage jobs process monitoring.
- Implemented Quality Stage for data cleansing, data standardization and matching process.
- Used plug-in stages such as Stored Procedure, Merge and also various stages like Sequential, Hashed, ODBC, Aggregator and Inter-Process.
- Integrated data from various sources into the staging area in data warehouse for integrating and cleansing data.
- Used the ETL DataStage Director to schedule, running the solution, testing and debugging its components & monitor the resulting executable versions.
- Defined production support methodologies and strategies.
Environment: IBM InfoSphere Datastage, Qualitystage & Information Analyzer 8.0.1, Ascential DataStage 7.5.2, Oracle 9i, Erwin, TOAD, Autosys 4.0, Windows 2003, SQL*Loader.
Confidential, South Portland, ME Mar`07-Feb`08
DataStage Developer
Confidential is a leading provider of payment processing and information management services to the U.S. commercial and government vehicle fleet industry. We provide fleets using our services with detailed transaction data, analysis tools and purchase control capabilities by capturing transaction data. The Data Warehousing Project involved designing, developing, testing and implementing complex mappings and workflows using the DataStage in order to facilitate the Client and Network partner details.
Responsibilities:
- Analyzed the existing ETL process and came up with an ETL design document that listed the jobs to load, the logic to load and the frequency of load of all the tables.
- Imported data from various transactional data sources residing on SQL Server, DB2 and Flat files and loaded into Oracle database.
- Developed DataStage jobs based on business requirements using various stages like lookup file, lookup stage, join stage, merge stage and sort stage.
- Used DataStage Manager to import the metadata / schema definitions into various jobs.
- Extensively worked on Data Acquisition and Data Integration of the source data.
- Used DataStage Director to run, monitor and schedule the jobs.
- Involved in creating technical documentation for source to target mapping procedures to facilitate better understanding of the process.
- Implemented various process checks, data checks and mail notifications to ensure the quality of the data that is loaded the data warehouse.
- Used Transformer stages with stage variables to map data, create expressions and constraints to match the business transformations.
- Redesigned the DataStage jobs to parameterize the constraints for incremental load.
- Created indexes and partitions in the database to improve the performance of ETL jobs.
- Involved in writing shell scripts for reading parameters from files and invoking DataStage jobs.
- Created user defined SQL queries based on requirements.
- Created jobs in Autosys Job Management Scheduler to run the jobs in sequence without manual intervention.
- Performed unit testing of all monitoring jobs manually and monitored the data to see whether the data is matched, if matched send it to the frictionless database otherwise reject and also created migration of documents.
- Created statistical reports using Business Objects.
- Involved in 24/7 ETL Production Support, maintenance, troubleshooting, problem fixing and ongoing enhancements to the Datamart.
Environment: Ascential DataStage 7.5, QualityStage 7.5.2, UNIX, Parallel Extender/Enterprise Edition Oracle 9i, DB2, SQL Navigator, Macros in Excel, XML, SQL, PL/SQL, SQL*Loader, Erwin.
Confidential, Irving, TX Sept `06-Feb `07
DataStage Developer
Confidential implemented a system to identify potential customers in the small business market. The project is to provide billing information in a standard format for the various Verizon billing systems across all Verizon jurisdictions. The main purpose of this project is to integrate the Data from MCI systems to the existing Verizon system.
Responsibilities
- Prepared mapping documents source system (Mainframe) and the warehouse tables (Teradata) and designed the jobs based on the understanding the data model table relationship.
- Extensively worked on Teradata utility like FastLoad to load the data from file to a table while doing a history load. And also used MultiLoad utility to load the incremental data into the tables.
- Worked with DataStage Manager to import/export metadata from database, DataStage Components between DataStage projects.
- Developed jobs using different stages like Transformation, Aggregation, Source dataset, external filter, Row generation, and Column generation.
- Used different Parallel Extender (PE) stages of DataStage Designer like Lookup, Join, Merge, Funnel, Filter, Copy, Aggregator, and Sort etc.
- Used search criteria, job comparisons and used reusable parameters sets to cut down development time.
- Extensively worked on Error handling, cleansing of data, Creating Hash files and performing lookups for faster access of data.
- Used DataStage Director to schedule, monitor, cleanup resources, and run job.
- Wrote UNIX shell Scripts for file validation and scheduling DataStage jobs.
- Involved in Unit testing and provide process run time.
- Involved in 24/7 ETL Production Support, maintenance, troubleshooting, problem fixing and ongoing enhancements to the Datamart.
Environment:DataStage 7.5.1 (Manager, Designer, Director, Administrator), UNIX Shell Scripting (Korn /KSH), SQL Server 2000, DB2 8.2, Oracle 8i, UNIX, Windows Server 2003.
Confidential Apr’05-Aug ‘06 ETL Developer
This project involves the design and development of an Insurance Information System. This System comprises the essential key performance Indicators of life Insurance. This System consists of relevant functional areas namely Policy, Claims, Underwriting, Financials, expenses and Customer segmentation. Claims Data mart was developed and star schema was designed to handle the various reports and analysis.
Responsibilities:
- Interacted with Business analyst on day-to-day basis.
- Involved in the Development, Implementation of the backend systems for the Database.
- Used the DataStage Designer to develop processes for extracting, cleansing, transforming, integrating and loading into Oracle Data Warehouse.
- Used Copy, Remove Duplicates, Transformer, Sort, Aggregator, Join, Merge, Lookup, Change Capture, and Change Apply Stages in designing jobs.
- Passed values by loading parameters through Job properties.
- Used DataStage Manager to import and export metadata, job categories and data elements.
- Designed Sequencer to automate the whole process of data loading.
- Created Cognos reports, Queries with constant interaction with the end users.
- Involved in trouble shooting of various reporting errors.
Environment: Ascential Datastage 7.5.1 PE (Parallel Extender), Oracle 8i, DB2 7.0, Teradata V2R4, PL/SQL, Toad, SQL Server 2000, Kshell Script, UNIX (Solaris), Cognos 7.4.