Datastage Developer - Senior Resume
Chicago, IL
SUMMARY
- Eight plus years of career reflecting experience and high performance in System Analysis, design, development and implementation of Relational Database and Data Warehousing Systems that included ETL Tools like IBM Infosphere DataStage 8.1, DataStage 7.1 and 7.5x
- Expertise in writing UNIX shell scripts and hands on experience with scheduling of shell scripts.
- Experience in Master data management (MDM), Data quality (DQ).
- Extensive experience in multi - dimensional business analytics, data architecture, data migration with multiple environments (OS/databases) & large data warehouse environments.
- Experience in data profiling and ensure data is available for cleansing and ETL process.
- Experience in System Development Life Cycle -Engagement, Analysis, Design, Development and Deployment.
- Experience in Designing ODS (operational data store) and integrated data from multiple sources for reporting and analysis purpose.
- Strong knowledge of data warehouse architecture and designing star schema, snowflake schema, FACT and Dimensional Tables, Physical and Logical Data modeling using Erwin tool.
- Experience in developing custom routines to simplify the determination of data output paths for complex business logic, prescripts and postscripts for triggering data stage.
- Experience in installing DataStage to Extract, Transform and Load data from the Source systems to the PeopleSoft EPM Data Warehouse.
- Concentration towards administeringbig data, specificallyClouderadistribution ofApache Hadoop.
- Good knowledge of Hadoop ecosystems, HDFS, Big Data, ETL, RDBMS.
- Extensive experience in dealing with high volume data, and performance tuning, maintaining and cleansing of data using QualityStage.
- Good experience in developing strategies for Extraction, Transformation, Loading data from/to various sources into data warehouse and data marts using tool like datastage
- Expertise in Data warehousing and Data migration.
- Worked on Relational database (RDBMS), Oracle, SQL Server, DB2 UDB, Informix, Sybase, Teradata, MS Access involving in Logical & Physical Design, Backup, Restore, Data Integration and DTS (Data Transformation Services) and creating Database objects (tables, indexes, triggers, views, and stored procedures).
- Excellent working knowledge on multiple platforms like windows NT/2000, UNIX (Sun Solaris, AIX, HP).
TECHNICAL SKILLS
ETL Tools: Ascential DataStage 6.x, 7.x IBM WebSphere DataStage 8.5, 8.1 and Infosphere Quality Stage 8.5.1, 8.1.1
O/S: Windows 98/2000/XP, NT 4.0, UNIX (HP, Sun Solaris 7/8, AIX, Humming Bird, PERL).
Modelling Tools: ERWIN model mart version 7.
Databases: DB2 8.2, DB2UDB, Oracle 10i/9.x/8i/8.0/7i, SQL Server 7.0/2000, Teradata, Informix and MS-Access, Master data Management (MDM).
Languages: C, C++, PL/SQL, SQL, UNIX shell programming, HTML and Visual Basic 5.0/6.0
Database Tools: SQL Plus, SQL Loader, Import, Export, Quest SQL Navigator 4.3, SQL Programmer, PL/SQL Developer
PROFESSIONAL EXPERIENCE
Confidential, Chicago, IL
Datastage Developer - Senior
Responsibilities:
- Involved in gathering business requirements along with Business Analyst from Prime and App tracker project vendors
- Involved in developing data models and new tables, worked on compatibility with existing data warehouse and existing job dependencies
- Used various parallel stages in creating parallel jobs with maximum efficiency and performance
- Developed parallel jobs, sequences, scripts to provide outbound extracts.
- Involved in enhancement for ETL code, which loads the existing EDW and Data-marts
- Involved in working with business users, in creating User related input files.
- Involved in creating parameter files for Zena (scheduling tool) and shell scripts
- Created new Zena codes for the new ETL codes and worked on dependencies and scheduling.
- Involved in working with claims, Member, product and reference related data from EDW and its source FACETS
- Was involved with code review, enhancement teams for code walk through and implementation into production.
- Worked on Agile methodology while implementing data-mart load related projects.
Environment: IBM Info Sphere Data stage 8.5 (E.E), DB2, Control Center, AIX Unix, Zena Schedule, Windows 7 client.
Confidential, Kansas City, MO
Datastage Developer - Senior/Business Support
Responsibilities:
- Involved in gathering business requirements along with BSA from vendors to work on extracts
- Involved in developing ETL architecture for the existing EDW and worked on creating Data-marts
- Involved in working with business users, in creating User related input files.
- Involved in a migration project from server to parallel.
- Developed parallel jobs, sequences, scripts to provide outbound extracts.
- Used various parallel stages in creating parallel jobs with maximum efficiency and performance
- Involved in creating parameter files for Zena (scheduling tool) and shell scripts
- Involved in working with claims, Member, product and reference related data from EDW and its source FACETS
- Was involved with code review, enhancement teams for code walk through and implementation into production.
- Worked on Agile methodology while implementing data-mart load related projects.
Environment: IBM Info Sphere Data stage 8.5 (E.E), DB2, Control Center, AIX Unix, Zena Schedule, Cognos 8 report studio, Windows 7 client.
Confidential, Detroit, MI
Datastage Developer - Lead/Business Support
Responsibilities:
- Involved in gathering business requirements for existing jobs and for the new development processes.
- Involved in gathering requirements and documenting for new project implementations with Canadian and Korean counter-parts for more effective analysis methods, which would enhance VIN level, dealer level and Regional level analysis.
- Involved in creation of separate data-mart project for the existing system and successful release.
- Developed Datastage parallel jobs, sequencer jobs, documented and deployed in the production environment.
- Worked with various parallel stages like look-up, join, transformer, modify, Datasets, sequential file, oracle etc.
- Created Reusable components Shared and Server Shared Containers, Imported and Exported Jobs, table definitions.
- Worked with Datastage Director to run, schedule, monitor and view logs for jobs and to full fill testing requirements.
- Good knowledge of Hadoop ecosystems, HDFS, Big Data, ETL, RDBMS.
- Experienced on working with Big Data and Hadoop File System (HDFS).
- Strong knowledge of Hadoop and Hive and Hive’s analytical functions.
- Implemented Proof of concepts on Hadoop stack and different bigdata analytic tools, migration from Oracle to Hadoop.
- Supervised and carried out successful monthly enhancements to the existing system.
- Created Adhoc reports and changed existing sql procedures to re-deploy in production according to the changes prescribed by business.
- Reporting involved, regular weekly and monthly reports, and reflect changes to the existing reports according to the business need.
- Involved in making changes to the existing procedure which provides VIN level information.
- Monitor the FTP process by the business users, which in-turn acts as source files for few datastage jobs.
- Leading a team of 4 developers off shore for carrying out monitoring and development works.
- Carried out emergency job processes during production failure, and made sure to perform routine weekly OLAP refresh using PLSQL-Procedures and minimize adverse effect on reports for business analysis.
- Involved in Olap process which provides the data to the cognos users, hands on experience on cognos.
- Gathered requirements and successfully implemented the SOX certification for existing Contribution Margin project.
Environment: IBM InfoSphere Datastage 8.1 (E.E), Oracle10g, TOAD 9.6, Red-hat server, Apache Hadoop, Cognos 8 report studio, Windows XP client.
Confidential, Detroit, MI
Datastage Developer
Responsibilities:
- Developed datastage parallel jobs and respective sequencer jobs with sequential files as source and Oracle as Target Database.
- Created and Modified PLSQL statements to join, update and insert in datastage jobs.
- Created design and production support documents for every job, explaining the job flow and production support in case of failures.
- Worked with various Partitioning (Round-Robin, Hash fields, Range, Random and etc.) and Collecting (Round-Robin, Ordered and Sorted Merge) techniques.
- Worked with File (Sequential, Data Set, File Set, Lookup File Set), Oracle, Processing (Transformer, Lookup, Sort, Aggregator, Funnel, Remove Duplicates, Copy, Filter, Switch and Etc.).
- Worked with Lookups, Modify stage, join, Pivot, Stored Procedure Stages.
- Used Datastage inbuilt Functions (Date, Time, Logical, Null, String, Casting) and created DS Transforms and Routines to perform the required transformations.
- Created UNIX shell scripts for Mobile App’s application to handle three different source files and modified an existing script for Insurance feeds.
- Created autosys JIL files for scheduling the datastage Sequencers.
- Modified the existing jobs, for performance tuning to maximize the performance.s
- Performed various test conditions on datastage jobs and submitted to the QA team.
- Created the deployment documents and successfully imported the developed applications on to the production BIS environment.
Environment: IBM InfoSphere Datastage 8.1 (E.E), Oracle10g, Autosys 4.1, TOAD, Linux, Windows XP.
Confidential, New York, NY
Datastage/Autosys Developer
Responsibilities:
- Involved in development, analysis and pre-production support of autosys for Confidential, for both GHI and QCARE feeds.
- The primary input sources are from mainframes and the data is transformed using Datastage 8.1 and loaded into Oracle.
- Developed few parallel job sequences for existing jobs in datastage before they implement in the autosys.
- Created Autosys JIL files for Autosys jobs and boxes of Datastage jobs.
- Monitored the autosys batch and fixed the failed datastage jobs and also made necessary changes in development and imported back into production.
- Worked with mainframe developers on data related and file related issues.
- Fixed the data load related QC’s which are raised by the end business users.
- Ensured everyday data is loaded into both Staging and ODS tables with their related batch ID’s.
- Manually loaded all the pending files with their respective batch ID’s in both staging and ODS tables.
- Created spreadsheet reports of failed and fixed autosys jobs and made sure the team worked on their autosys boxes assigned.
Environment: IBM InfoSphere Datastage 8.1 (E.E), Oracle10g, Autosys, TOAD, Linux, Windows XP, MS Word, Excel.
Confidential, Alpharetta, Georgia
Datastage Developer
Responsibilities:
- Involved in Datastage 8.1 installation on Redhat Linux with metadata tables based on Oracle platform.
- Created multi node configuration files for using in parallel jobs, for development and production environments.
- Interacted with Business Analysts to finalize the requirements and documented the technical design document for Coding.
- Developed number of parallel jobs using different types of parallel stages like sequential, Dataset, transformer, lookup, join, Merge, pivot, column generator, copy, filter, funnel, remove duplicate, switch.
- Involved in tuning of DataStage jobs for better performance by using available tuning techniques.
- Used Environment Variables, Stage Variables and Routines for developing Parameter Driven Jobs and debugging them.
- Used the Data stage Manager to import, export jobs and routines.
- Created sequential jobs and used Datastage Director for scheduling the jobs.
Environment: IBM InfoSphere Datastage 8.1 (Server & E.E), Oracle 9i/10g, TOAD, Linux, Windows XP, MS Word, Excel.
Confidential, New York
Datastage Developer
Responsibilities:
- Developed Datastage Parallel Jobs using required stages, and the obtained data from different sources like flat files, source tables were formatted, Cleansed, summarized, aggregated, transformed and loaded into data warehouse.
- Involved in design and development of parallel jobs, sequences using the Designer.
- Created, Monitored and participated in several Project deployments, which involved updating the timelines to PVC’s and performing the timeline tasks on UNIX server.
- Interacted with Business Analysts to finalize the requirements and documented the technical design document for Coding.
- Designed several parallel jobs using Sequential File, Dataset, Join, Merge, Lookup, Change Apply, Change Capture, Remove duplicates, Funnel, Filter, Copy, Column Generator, Peek, Modify, Compare, Oracle Enterprise, Surrogate Key, Aggregator, Transformer, Row Generator stages.
- Extensively worked with all the new features in IBM information Server.
- Encompasses Database designing, ETL phases of Data Warehousing. This is achieved with an emphasis on relational data modeling and dimensional data modeling for OLTP and OLAP systems.
- Tuned DataStage jobs for better performance by creating Datastage Lookup files for staging the data and lookups.
- Used the Data stage Manager to import, export jobs and routines.
- Created Shared Containers to increase Object Code Reusability and to increase throughput of the system.
- Experience developing complex transformations, surrogate keys, routines, dimension tables and fact tables.
- Used Environment Variables, Stage Variables and Routines for developing Parameter Driven Jobs and debugging them.
- Data staging of the ETL solutions from different source systems to target database.
- Performed extraction, transformation and loading of data using different types of stages and by performing derivations over the links connecting these stages.
- Enhanced the Job Performance by using proper Partitioning methods and analyzing the resources utilized using Job Monitor.
- Used Data Stage Director for running and monitoring performance statistics.
- Scheduled existing UNIX shell scripts using Crontab.
- Modified UNIX shells according to the latest business requirement.
- Created and installed public keys in various directories for FTP transfers to and from offshore teams.
- Modified UNIX shells to implement new landing path for various feeds with new FTP connections and for proper log maintenance.
- Created SQL statements used for modifying, monitoring and resetting jobs during monitoring process.
- Redesigned few jobs in Datastage Designer to meet the changes in new incoming feeds.
- Exported and imported the changed jobs in development and production servers using Datastage Manager.
- Created and modified all the jobs and shell scripts with new email Id’s and groups for abort and failure notifications.
- Actively involved in everyday job monitoring and resolved all the issues which are related to job aborts and job failures.
- Provided support for offshore teams to resolve their issues during monitoring.
Environment: IBM-InfoSphere Datastage 8.1 (Server & E.E), Erwin 4.2, Oracle 9i/10g, TOAD, UNIX- AIX, Windows XP, MS Word, Excel, Crontab, Clear Quest.