We provide IT Staff Augmentation Services!

Datastage Developer Resume

2.00/5 (Submit Your Rating)

Kansas City, MO

SUMMARY

  • 8 plus years of professional experience in Extract, Transform and Load (ETL) experience using IBM Info sphere DataStage 11.5/11.3/ 9.1/8.5/8.1/7.5.
  • Expert Level experience in DataStage Client components - DataStage Designer, DataStage Manager, DataStage Director and DataStage Administrator
  • Extensively worked on Server Edition, Enterprise Edition (Parallel Extender) and Development of Data Warehouse/Data Mart Applications.
  • Having DataStage experience on connecting to Hadoop / BIGDATA
  • Having experience in dealing with SAP systems as source and configuring SAP related stages in DataStage
  • Extensively made use of all stages Aggregator, Sort, Merge, Join, Change Capture, Peek stages in Parallel Extender job.
  • Used Data Stage Manager to import/export Data Stage projects and jobs to define table definition in repository.
  • Used Data Stage Director to debug, validate, schedule, run and monitor Data Stage Jobs
  • Experience in designing Job Batches and Job Sequences for scheduling server and parallel jobs using DataStage Director, UNIX scripts.
  • Proficiency in data warehousing techniques for Slowly Changing Dimensions, Surrogate key assignment and Change Data Capture
  • Applied the DataStage run time column propagation (RCP) and created DataStage Parallel sync up the tables between environments
  • Extensive experience in dealing with High Volume Data, Performance Tuning, Maintaining, Multiple Job
  • Expertise in writing UNIX shell scripts and hands on experience with scheduling of shell scripts using AUTOSYS/CONTROL-M.
  • Expertise in Data Migration and Upgrading.
  • Expertise in using ERWIN for data modeling.
  • Proven track record in troubleshooting of DataStage Jobs and addressing production issues such as performance tuning and enhancement
  • Converted Complex job designs to different job segments and executed through job sequencer for better performance and easy maintenance
  • Strong Knowledge of Data Warehouse architecture in Designing Star Schema, Snow Flake Schema, FACT and Dimensional Tables. Physical and Logical modeling using Erwin.
  • Extensive experience in developing strategies for Extraction, Transformation, loading (ETL) data to various sources into data warehouse and Data Marts using Data Stage.
  • Involved in Performance Fine Tuning of ETL programs. Tuned DataStage and Stored Procedures Code.
  • Excellent experience in Relational Database (RDBMS), Oracle 11g,10g,9i, Microsoft SQL Server, Netezza, Teradata Load and MultiLoad, SQL, PL/SQL, TOAD.
  • Involved in Logical and Physical Design, Backup, Restore, Data Integration and Data Transformation Service and Creating Database Objects (Tables, Index, Triggers, Views, and Store Procedures).

TECHNICAL SKILLS

ETL tools: IBM Information Server DataStage 11.5/11.3/ 9.1/8.7/8.5/8.1

Data Modeling: E-R Modeling, Star and Snowflake Schema Modeling

Databases: Oracle 11g/10g/9i/8i, SQL Server 2008/2005, Netezza 7.x/6.x Teradata 13.x, DB2 UDB, Postgresql

Languages: C/C++, SQL, PL/SQL and Shell Scripting

Operating System: Unix, Linux, Windows XP,7,8 server

Scheduling Tools: Autosys, Control-M, Tivoli, Jenkins, Zena

PROFESSIONAL EXPERIENCE

Datastage Developer

Confidential, Kansas City, MO

Responsibilities:

  • Attended meeting with business in gathering requirements and for developing the code.
  • Analyzed if the new requirement impacts the excising system and how to fit into the system.
  • Prepared mapping documents for designing and developing the datastage jobs
  • Worked on claims, provider and membership data coming from various vendors like Bcbs Association, BcbsSC, ESI, EyeMed, MedTrack, CVS, LDI and so on.
  • With the help of datastage designer tool, extracted data from various source systems and implemented logic and loaded into target database and outbound files to send it to vendors.
  • Worked on Server and parallel jobs and created parallel jobs.
  • Extensively used ODBC connector stage, DB2 UDB stage, Complex flat file, Sequential file, dataset, Aggregator, Change capture, Copy, Filter, Link collector, hash file, FTP, Join, Lookup, Sort, transformer stages.
  • Used Data Stage Director for job monitoring, view logs, testing and debugging its components.
  • Worked on performance tuning of the SQL queries
  • Used various existing routines for date conversions and calculating amounts and age
  • Worked on deploying the code to production and there by verifying the post run data
  • Worked with different teams like code review team, deployment team, Scheduling team and production team in different situations to get the work done.
  • Prepared DDL’s and DML and provided to DBA as part of table deployment changes.
  • Created complex sequence jobs and control jobs and adding them to the zena scheduler.
  • Ran the scheduled job in request time trigger mode using zena scheduler in Development, Test, Stage environments.
  • Worked on Unix shell script based on the Provided business requirement and running them in debug mode to identify the warning and errors and fixing them

Environment: IBM InfoSphere Datastage 11.5v, AQT for DB2 6.5v, SQL Server, Unix, Winscp, TFS and GIT Version control, Zena scheduler, Visual Studio, Azure devOps, Putty, CDMA Mapping tool .

Datastage ETL Developer

Confidential, Princeton, NJ

Responsibilities:

  • Attended meetings with business teams to understand the requirement and prepared the low-level design document, technical specification documents.
  • Identified the impacts and created Data Mapping Sheets.
  • Active participation in decision making and QA meetings and regularly interacted with the Business Analysts &development team to gain a better understanding of the Business Process, Requirements& Design.
  • Used DataStage as an ETL tool to extract data from sources systems, loaded the data into the IBM DB2 database and POSTGRESQL database using ODBC connector stage.
  • Created jobs to read and write data into datasets, sequential files.
  • Designed and Developed Data stage Jobs to Extract data from heterogeneous sources, applied transform logics to extracted data and Loaded into Datamart.
  • Created Datastage 11.3 jobs using different stages like Transformer, Aggregator, Sort, Join, Merge, Lookup, Data Set, Funnel, Remove Duplicates, Copy, Modify, Filter, Column Generator, Difference, Row Generator, Sequencer, Email Communication activity, Command activity, Sequential File, Dataset, Terminator activity.
  • Used Data Stage Director for job monitoring, view logs, testing and debugging its components, and monitoring the resulting executable versions on ad hoc or scheduled basis.
  • Converted complex job designs to different job segments and executed through job sequencer for better performance and easy maintenance.
  • Created master jobs sequencers.
  • Performed performance tuning of the jobs by interpreting performance statistics of the jobs developed.
  • Prepared DDL's for table creation, table modification, index changes. Tested and executed the same in all environments.
  • Prepared the DML's for maintenance tables, reviewed, tested and executed them.

Environment: IBM InfoSphere Datastage 11.3v, Dbeaver 5.1.1 for DB2 and Postgresql, SQL Server, Unix, Jenkins, Amazon work Docs, Amazon Work Space.

Datastage ETL Developer

Confidential, Chicago, IL

Responsibilities:

  • Designed and developed Datastage jobs that load millions of records from Source to target tables
  • Established best practices for DataStage jobs to ensure optimal performance, reusability
  • Working on enhancing the sql scripts or queries to make it better performed.
  • Prepared technical design Document
  • Updated existing process flow document (Microsoft Visio) for newly added code to scheduler
  • Involved in system analysis and design of DataStage jobs
  • Design and development ETL processes for extracting data from legacy systems and loading into target tables using SQL and DataStage designer client
  • Developed jobs using various processing stages like Join, Lookup stage, Column generation and Funnel stages etc by using parallel extender partition concepts
  • Used Datastage Director to verify logs and monitoring jobs during run and after running the jobs
  • Supported Product Verification team to Identify and to fix the issues
  • Performing Unit Testing and integration testing
  • Triggering the job streams from Tivoli Scheduler and capturing the Tivoli logs
  • Worked with Tivoli scheduler team to add new jobs to schedule
  • Worked on Unix shell script for enhancement to support business requirements
  • Running the Shell scripts in debug mode to identify the cause of failure
  • Check in and check out of Datastage jobs, Shell scripts into Clear case repository
  • Provided knowledge transfer to newly on-boarded resources
  • Analyzing the production issues and fixing them by working with business analyst
  • Responsible for creating the Deployment verification guide.
  • Responsible for verifying the code after it has been deployed in production, Product Verification environments.
  • Responsible for synchronizing the code in different development environments
  • Compared the Datastage code against different environments using Datastage tool

Environment: IBM InfoSphere Datastage 11.5v, Toad for DB2 6.5v, SQL Server, Unix, Winscp, Clear case Version control, IBM Tivoli scheduler, Clear Quest.

DataStage ETL Developer

Confidential, Lebanon, NJ

Responsibilities:

  • Participated in all stages of the development life cycle including reverse Engineering, requirement analysis, preparing mapping documents, design and developing ETL jobs and Sequences.
  • Worked on Reverse engineering between existing source and target layers using Oracle SQL.
  • Developed jobs using various stages like Oracle Connector, Copy, Pivot, Funnel, Lookup, Join, Merge, Sort, Transformer, Dataset, Row Generator, Column Generator, Datasets, and Aggregator Stages.
  • Used very extensively the Peek stage for debugging purposes.
  • Developed Multi-Instance jobs to generate Pick list values.
  • Used Checksum and Change Data Capture (CDC) stages for doing incremental loads.
  • Implemented type 1, type 2 slowly change dimensions (SCD).
  • Sat with QA people helped them understanding the requirements, analysed and fixed the bugs raised by them.
  • Worked on the testing of code and fixed defects during the conversion.
  • Involve in design and code reviews and extensive documentation of standards, best practices, and ETL procedures.
  • Took the responsibility of code back up and checking it in TFS and maintain the versions of the code changes.
  • Responsible for adopting the Company Standards for Stage & Link Naming Conventions.
  • Design Technical documents on delivered functionality / code as required.
  • Reviewed, Unit tested and documented the code that was done by teammates.
  • Resolving Defects and issues in production environment and bug fixing.
  • Created the Sequences for running the jobs and logging the counts into audit tables.
  • Used Loop stage in sequence to call a Multi-Instance job which dumps data from one Virtual Database to other Virtual Database multiple times.
  • Responsible for all activities related to the development, implementation, administration and support of ETL processes for large-scale data warehouses using IBM Information Server.
  • Developed SQL queries to perform DML against the databases.
  • Performed Unit and System Testing.
  • Worked on Migrating Code from 11.5 to 9.1.
  • Worked on Data Model changes to add new required fields to the existing tables and implemented those changes in the existing code.
  • Developed test data and conducted performance testing on the developed modules and unit test plans and documents.
  • Extensively worked on user-defined SQL for overriding Auto generated SQL queries in Data Stage.
  • Worked on Agile and Scrum methodologies.
  • Identified issues, tracked, reported and resolved in a timely manner.
  • Fine Tune Jobs/Process for higher performances & debug complex jobs.
  • Have good knowledge in creating USER’S and assigning them the roles and role hierarchy through sales force front end application.

Environment: IBM InfoSphere 11.5v/9.1v, Toad for oracle 12.7, Unix, Winscp, TFS Version control, Jira, ESP scheduler, Salesforce.com

DataStage ETL Developer

Confidential, Dallas, TX

Responsibilities:

  • Creating Low Level Design documents by understanding the requirements and there by develop jobs.
  • Worked in Data Acquisition project, where we need to extract data from different sources, process the data and generate the files and transfer these files to Target Systems.
  • Worked on several change requests, which were created because of production incidents and requirement changes to the code in production environment.
  • Responsible for using different types of Stages such as ODBC Connector, Oracle Connector, DB2 Connector, Teradata Connector, Transformer, Join, Sequential File to develop different jobs.
  • Developing DataStage Parallel and Sequence Jobs.
  • Developed common Jobs, Shared containers and Server Routines, which are used across the project in most of the interfaces.
  • Used Job parameters, stage variable and created parameter files for flexible runs of Job based on changing variable values
  • Imported the required Metadata from heterogeneous sources at the process level
  • Created Batches (DS job controls) and Sequences to control set of jobs
  • Scheduled jobs using Control-M scheduler utility based on the requirements and monitored the production processes closely for any possible errors
  • Created Unix Shell Scripts that takes care of end-to-end automation. Developed UNIX shell Scripts that trigger DataStage jobs, transfer the output files, perform basic validations on file.
  • Implemented and hardcoded high performance DataStage routines
  • Supporting the testing team, Integration Team and Reporting team after ETL data Load
  • Performed the Integration and System testing on the ETL jobs
  • Responsible for generation of DDL statements which are executed for database creation
  • Deploy the developed code to SIT, Production environments and validate the code
  • Fix the defects raised by testing team and maintain the status in HP Quality Center.
  • Extensively used SQL tuning techniques to improve the performance of Data Stage Jobs.
  • Tuned DataStage transformations and jobs to enhance their performance.
  • Provide Post Implementation Support.

Environment: IBM InfoSphere 11.3v, IBM InfoSphere 9.1v, IBM InfoSphere DataStage 8.7v, UNIX shell Script, ORACLE 11g, SQL Developer, DB2,Teradata, CASSANDRA, AQT (for accessing SQL Server and DB2), Office.

Senior DataStage Developer

Confidential, Cincinnati, OH

Responsibilities:

  • Handling production issues and working on developing new design specifications into ETL coding and mapping standards.
  • Extensively used Datastage Designer to develop processes for extracting, transforming, integrating and loading data from various sources into the Data Warehouse.
  • Created ETL processes composed of multiple Datastage jobs using job sequencer and developed shell scripts to automate the process and tested the processes.
  • Extensively worked on different databases like Oracle, Db2, Netezza and SQL, Cassandra Server to extract and load the data from one another.
  • Used different types of stages like Transformer, CDC (change data capture), Remove Duplicate, Aggregator, ODBC, Join, Funnel, dataset and Merge for developing different jobs.
  • Involved in the migration of the jobs from DB2 to Netezza
  • Extensively used Parallel stage like row generator, column generator and Peek Stages for debugging purpose.
  • Used the Director and its run-time engine to schedule running the job, testing and debugging its components, and monitoring the resulting executable versions.
  • Documented data sources and transformation rules required to populate and maintain Data Warehouse content.
  • Customized UNIX scripts as required for preprocessing steps and to validate input and output data elements.
  • Developed Multi-Instance reusable Datastage jobs.
  • Used Netezza Enterprise stage for doing loads into Netezza Database
  • Effectively implemented Partitioning and Parallelism techniques to fully utilize the resources and enhance job performance.
  • Wrote SQL scripts to extract and load data from source and target databases.
  • Implemented slowly change dimensions (SCD) type 1, type 2.
  • Designed and developed various jobs for scheduling and running jobs under job sequencer and DataStage Director.
  • Extensively implemented import/export utilities for migrating code.
  • Replaced transformer stages with other stages to improve performance of job.
  • Attended daily meetings to review the status of the schedule and go through pending issues.
  • Coordinated with Release manager and DA to migrate the components from one environment to another environment. Integrated with other Projects, for sharing the table structures and data.
  • Involved in performance tuning by rewriting the queries and modifying existing Datastage jobs.
  • Involved in production support for production cycle runs and ETL related issues. Communicated data availability with users and management.
  • Documented the purpose of mapping so as to facilitate the personnel to understand the process and incorporate the changes as and when necessary.
  • Also involved in ETL test plans, test scripts and validation based on design specifications for unit testing and functional testing.

Environment: IBM Infosphere Datastage 9.1/8.7/8.1, Oracle 10g, Netezza, DB2, SQL, PL/SQL, UNIX Shell Scripting, Datastage Version Control, MS SQL server, Control M 7.5, RTC.

We'd love your feedback!