Senior Etl Developer Resume
Chicago, IL
SUMMARY
- IT professional with over 8 years of experience delivering innovative Database/Data Warehouse solutions to Telecom & Health Care verticals, utilizing ETL and Big Data (Hadoop) technologies
- Experienced in SDLC activities such as requirements definition, functional & technical design, testing, production support and implementation.
- Working Experience in Datastage, Informatica, TALEND, SQL, PL/SQL, Unix, Hadoop.
- Experienced in interacting with business in gathering requirements and developing technical design specifications for solutions involving data pipelines, data cleansing and processing of diverse data from disparate sources.
- Strong knowledge of Data Modeling (Relational, dimensional, Star and Snowflake Schema), Data analysis, implementations of Data warehousing on Windows and UNIX environments
- Strong experience in Data Migration activities, worked with Teradata utilities like BTEQ, Fast Export, Fast Load, Multi Load and SQL LOADER scripts to export and load data to/from different source systems including flat files
- Experienced in performance tuning of targets, sources, mappings, and sessions. Experience in Performance tuning of ETL processes and Datastage jobs.
- Experienced in building complex UNIX shell scripts using cron, sftp, C:D, file validation scripts, purge logic scripts.
- Worked on the XML transformation capabilities in IBM InfoSphere DataStage 8.5 Transform XML using the DataStage XML INPUT & OUTPUT stage.
- A hard working, self - motivated quick learner, possess excellent organization skills, experienced to work in large teams as well as working independently.
- Skilled in problem-solving and trouble-shooting capabilities
- Recipient of organizational recognitions such as Star Performer award for successfully leading migration from Daytona technology to ETL Data stage 8.7vx, Pat-On-The-Back award for significant contributions in projects.
TECHNICAL SKILLS
ETL Tools: IBM DataStage 11.3, 8.7, 8.5, Talend Integration ETL Tool 5 & 6, Ascential DataStage 7.5, Informatica Power Center 9 and 10
Databases: Oracle 9i/10g/11g, MS Access, CASSANDRA, MONGO DB
Languages: UNIX SHELL, HSQL, CQL (CASSANDRA QUERY), TERADATA, SQL, Confidential &T Product- Cymbel Query Language, JAVA(core), XML
Hadoop: Hbase, Pig, Hive, Ambari, Hue
Operating Systems: Windows XP/NT/2000, UNIX, Linux.
Scripting: Unix Shell Scripting. TWS Composer Files Scripts
Other Tools: MS Office, SQL*Plus, TOAD, SQL Developer, Teradata, Management Studio.
Schedulers: Datastage Internal Scheduler, Cron Tab, Zera, IBM Tivoli TWS, tidal
PROFESSIONAL EXPERIENCE
Confidential, San Francisco, CA
Senior ETL/ELT Developer
Responsibilities:
- Involved in requirement, analysis and understanding of business requirements identify the flow of information and analyzing the existing systems.
- Involved in the data analysis for source and target systems and good understanding of Data Warehousing concepts, staging tables, Dimensions,Facts and Star Schema
- Designed, developed and improvedETL structures to extract,transform and load data from data sources into Oracle data warehouse and other databases based on business requirements.
- Developed mappings in Informatica to load the data from various sources using different transformations like Source Qualifier, Lookup, Expression, Aggregate, Update Strategy, Sequence Generator, Joiner, Normalize, Filter and Router, Stored Procedure
- Responsible for testing, modifying, debugging, documenting and implementation ofInformatica mappings.
- Performed metadata validation, reconciliation and appropriate error handling inETLprocesses
- Troubleshootdata issues, validated result sets, recommended and implemented process improvements.
- Responsible for performance tuning inInformatica Power Centerat the Target Level, Source level, Mapping Level, Session Level, and System Level.
- Extensively worked withSQL queries.Created Cursors, functions, stored procedures, packages, Triggers, views, materialized views usingPL/SQL Programming.
- Extensively worked with performance tuning ofOracle
- Supported and assist the Lead Architect with project planning, analysis, design, testing, system documentation and user training.
- Managed work assignments and coordinated within the development team
Environment: s: Informatica Power Center 10, Linux, UNIX Shell Scripting, Oracle 12 c database, Netezza, Jira,Tidal.
Confidential, Phoenix, AZ
Senior ETL/ELT Developer
Responsibilities:
- Analyze business requirements and created technical specification document, source to target mapping for ETL development.
- Designed table mappings using UML’s like Use Cases, Activity Diagram, Sequence Diagram, and Class Diagram.
- Importing and exporting data from FACETS UNIX server into HDFS.
- Involved in AGILE requirement gathering and legacy system analysis.
- Based on User stories in RALLY, Estimated Development Tasks in RALLY (CENTRAL AGILE) for each Sprint User.
- Created scripts like UNIX file Validation, File Watcher Scripts and Pig Latin scripts to process the data and load data in HDFS. Extracted the data from oracle into Hive using Sqoop for data processing.
Environment: s: Horton works Hadoop, MapReduce, HDFS, Hive, Sqoop, Pig, Oozie, Linux, Shell Scripting, Oracle 12 c database, SFTP, File watcher UNIX Scripts. CA RALLY Agile central, Cron tab
Confidential
Responsibilities:
- Worked on business problem statement and provided solution, technical specification document with ETL mapping.
- Importing and exporting data form local system to HDFS.
- Composed Pig scripts to process the data and developed data pipeline using Talend Integration ETL to store data into HDFS and Hive performed the real-time analytics on the incoming data.
- Involved in creating Hive tables, loading with data and writing Hive queries using the HiveQL which will run internally in the map-reduce way.
- Extracted the data from oracle into Hive using Sqoop.
Environment: s: Hortonworks Hadoop, MapReduce, HDFS, Hive, Sqoop, HBase, Pig, Oozie, Talend Integration ETL Tool, Linux, Shell Scripting. CA RALLY Agile central
Confidential
Responsibilities:
- Analyzed business requirements and created technical specification documents, source to target mapping for the ETL development.
- Involved in preparing high level and detailed design documents and acceptable differences documents for the end users.
- Data extraction from fixed width files and transformed as per the business requirement and loaded into the staging Oracle FACETS 5.2 Tables.
- Involved in AGILE requirement gathering and legacy system analysis.
- Based on User stories in RALLY, Estimated Development Tasks in RALLY ( CA CENTRAL AGILE) for each Sprint User Story .
- Recurrent billing payment for Confidential Medicaid claim auto monthly billing payments.
- Worked extensively with Dimensional modeling, Data migration, Data cleansing, ETL Processes for data warehouses.
- Developed Slowly Changing Dimension Mappings for Type 3 SCD. Formulated UNIX File watcher and File validation Script.
- Developed Tableau workbooks from Facets Data Source for Exception and Data Balancing & Control report. Integrated a wide variety of source file layouts into the data warehouse.
Environment: RALLY, IBM datastage 11.3, FACETS 5.2, Oracle Developer, Oracle 12c database, UNIX Shell Scripting, IBM TIVOLI, Tableau
Confidential
Responsibilities:
- Responsible for building scalable distributed data solutions using Hadoop.
- Worked on Datastage 8.7 ETL migration to Talend Studio ETL process.
- Design, develop, validate and deploy the Talend ETL processes for the DWH team using HADOOP (PIG, HIVE) on Hadoop.
- Collaborate with the Data Warehouse team to design and develop required ETL processes, performance tune ETL programs/scripts.
- Handled importing of data from various data sources, performed transformation like ETL (Extract Transform and Load) and ELT (Extract Load and Transform) into HDFS.
- Extracted the data from Oracle 12c, transformed and load in HDFS using Talend Studio ETL TOOL.
- Analyzed the data extraction by performing Hive queries and running Oracle SQL to know user behavior.
- Continuous monitoring with ADMIN and managing the Hadoop cluster through Cloudera Manager.
- Developed Hive queries to process the data and generate data cubes for visualizing.
Environment: Hadoop 2.2, MapReduce, HDFS, Hbase, Hive, TALEND ETL TOOL for DWH.
Confidential
Responsibilities:
- Performed impact analysis for custom database objects, data and its dependencies in Facets core and custom tables, custom batch and front end extension etc.
- Infrastructure management performance tuning, Facets install and configuration, batch configuration etc. Data and Object migration - Custom object migration, Data migration of facets and custom tables from Sybase to Oracle 12c.
- Code Remediation - Technical design, development, unit testing for impacted components, upgrade current Data Stage Version 7.x and 8.5 to version 8.7/8.5 to be compatible with Oracle 12c RAC, Technical configuration, Batch configuration, Defect tracking and fixing during SIT and UAT.
- Closely worked on all datastage processes from Provider, Membership & Finance by formulating extraction, and transformation and load schemes for all process.
- Involved in business requirements gathering meetings and created functional, technical specification documents and source to target mapping documents.
- Development of data stage Job remediation from Sybase to Oracle stage design, execution, testing and deployment on the client server.
- Extensively worked with DataStage job for Extraction & Loading the data into the data warehouse.
- Operated on UHG possessed environment like Common ETL Framework, Data stage job run through ITG, FAST Process for Deployment.
Environment: IBM Data stage Enterprise Edition 8.7(Data stage, Quality Stage), Oracle 11g, Fixed width files, FACETS 5.2, 4.71 Windows XP, UNIX (Shell Scripting).
Confidential, Chicago, IL
Senior ETL Developer
Responsibilities:
- Analyzed business requirements and created source to target mapping for the ETL development. Involved in preparing high level and detailed design documents and acceptable differences documents for the end users.
- Worked on the XML transformation and schedule Datastage jobs using Zena scheduling tool.
- Extracted data from fixed width files, transformed and loaded into the staging Teradata Database. Created Datastage Parallel jobs using Designer and extracted data from various sources, transformed data and loaded into target databases like Teradata Database.
- Performed Data Quality Checks on the source data.
- Involved in integration testing, co-ordinate development activities, production support and maintenance of ETL Jobs.
Environment: IBM Data stage Enterprise Edition 8.7(Data stage, Quality Stage), Oracle 11g, UNIX
Confidential, Chicago, IL
ETL Developer
Responsibilities:
- Involved in business requirements gathering meetings and created functional, technical specification documents and source to target mapping documents.
- Involved in the entire life cycle from design, developing, testing using Data Stage 8.7 Designer to develop Parallel Jobs for Extracting, Cleansing, and Transforming, Developed Shell Scripts for file validation and for data loading procedures.
- Performed DQ checks as per the requirement, which include account number validations.
- Sequential file stage, transformer and Oracle connector stages to load the data as per the requirement.
- Created Job sequencers to run the data stage jobs.
- Created extracts as per the end users request from Oracle tables and delivered to end users through SFTP.
- Involved in Development activities, coordinating with testing team and production issues.
Environment: Datastage 8.7, Oracle 11g, UNIX.
Confidential
Responsibilities:
- Involved in Software Development LifecycleExperience (SDLC) from Business Analysis to Development, Testing, Deployment and Documentation.
- UsedTeradata utilities fast load, Multi Load to load data.
- Composed BTEQ, Fast export scripts to transform data
- Performance tuned and optimized various complex SQL queries.
- Worked on Joiner,Aggregator,Update Strategy, Rank, Router, Lookup, Stored Procedure,Sequence Generator, Filter, Sorter, and Source Qualifier.
- Created several UNIX scripts on BTEQ and file management.
- Gatheredsystem design requirements, design and wrote system specifications.
- Worked on data warehouses with sizes from 30-50 Terabytes.
Environment: Informatica Powercenter UNIX, SQL, Teradata BTEQ, FASTLOAD, FAST EXPORT, Shell scripting, Putty.