Senior Etl Developer Resume
NJ
SUMMARY
- Currently working as a Senior ETL Developer in a Project that aims at building an Enterprise Data Hub in Data Lake housing the NYL Insurance & Investment Data for reporting, analytical processes, legal and compliance and other needs.
- Actively involved in Data Ingestion into Hive Datawarehouse using Big Data technologies like Apache Hadoop, Sqoop, Pig and Oozie with IBM Big Integrate as the ETL Platform.
- Have extensive experience in the solution design of Insurance Data Marts using Datastage as ETL tool for Business Intelligence/Reporting.
- Recently delivered a Data Mart central to the Underwriting process.
- Looking for a contract - to-hire position which provides opportunities, growth and career advancement in Big Data Analytics/Data warehousing field while I strive to enhance my work competencies, professional skills and business efficiency.
- Experience in working in Agile ETL development projects using IBM Infosphere Datastage v 8.1/8.5/9.1/11.5 in Insurance, Banking, Healthcare and Financial Services domain.
- Experience in Data Modelling, ETL Design, Development, Scheduling and Implementation of Data warehouses, well-versed with Data Staging/Integration/Migration life cycle.
- Experience in working in Hadoop based IBM Big Integrate Environment to build ETL component for a Hive Database based Datawarehouse to support Big Data Analytics.
- Experience in using HDFS Commands through Edge Node (CLI), and Ambari (GUI) in a Hadoop cluster, and using Data wrangling tools like Trifacta and Waterline.
- Experience in using Sqoop to import RDBMS Source system data into the Data Lake(Data Ingestion), using Protegrity encryption for Data tokenization.
- Experience in admin tasks like installation, configuration and set up of environments, and migrating ETL tools and projects.
- Experience in preparing HLD/LLD Documents, Technical Design documents, Reverse Engineering documents and Test Plan Documents.
- Expert in Design and Development of Parallel (and Server) jobs and Sequences with custom routines for data prepping and parameter initialization and Shared Containers.
- Expert in Data Standardization using Qualitystage and CASS stage. (Have standardized Client Data: Address/Name/Phone).
- Efficient in all phases of the development lifecycle, coherent with Data Analysis, Data Cleansing, Staging, Data Conversion/Integration, Performance Tuning and System Testing.
- Strong knowledge in Data warehousing concepts, dimensional Star Schema and Snowflake Schema methodologies (OLAP).
- Efficient in incorporation of various data sources such as Oracle, DB2, SAP, Siebel, Teradata, XML and Flat files into ODS and DWH areas into the staging area.
- Designed and Implemented Operational Audit/Error/Exception/Reject Handling Framework to reuse across multiple Domains and Projects.
- Proven track record in troubleshooting Data Stage jobs to address failures in production and issues like performance tuning and enhancement.
- Expert in unit testing, system integration testing and well versed with tools like HP ALM -QC, JIRA and Rally for Defects Tracking and Enhancement Tasks.
- Strong knowledge in writing Unix Scripts for file manipulation, archiving data, scheduling jobs and FTP/SFTP of files.
- Experience in preparing Data Lineage Documents, Data definition for Data Governance.
- Proficient in writing SQL scripts to create/alter tables, database sequences, indexes, triggers, procedures and functions in SQL and Oracle
- Experience in Informatica (9.5/9.1.8.6 ).
- Focused with an inherent zeal to excel and succeed in new challenges, highly motivated and adaptive with the ability to grasp things quickly.
- Ability to work effectively and efficiently in a team and individually with excellent interpersonal, technical and communication/presentation skills.
TECHNICAL SKILLS
ETL Tools: Infosphere Datastage 8.x/9.1/11.5, IBM Big Integrate, Qualitystage, InfoSphere Data Replication CDC, Informatica PowerCenter 9.0
Databases: Oracle 11g/10g, MS SQL, Hive, DB2, Teradata, Sybase, Siebel
Non ETL Tools: TOAD 12.0, Microsoft SQL Developer, SQL Workbench, WinSCP, Putty, HP ALM -QC, Rally, JIRA, AQT, XML Assyst - Incident Management, Jobtrac-Mainframe Scheduler, Oracle Watchlist Screening
Operating Systems: Windows, Unix
PROFESSIONAL EXPERIENCE
Confidential, NJ
Senior ETL Developer
Responsibilities:
- Data Ingestion from 17 different sources like Policy, Marketer, Client, Beneficiery etc into the data lake using Sqoop import.
- Data tokenization using Protegrity function to protect PI data (sensitive data).
- Use data wrangling tools like Trifacta, waterline to perform data analysis to identify valid values.
- Create HLD/LLD/Source to Target Mapping Documents for moving the data from source to processed to published zones.
- Identify and define data elements required from various domains for Data Governance and Lineage.
- Set up the Big Integrate environment for the Datastage jobs to run on Hadoop Cluster.
- Create Hive tables using Beeline commands on Hadoop Edge Node and Ambari.
- Import Hive table definitions using Infosphere Metadata Asset Manager.
- Create SCD Type-2 parallel jobs to load data in hive tables in the source zone to processed and published zones.
- Extensively used Join, Merge, Lookup, Surrogate key, Change Capture, Funnel, Sort, Transformer, Copy, Remove Duplicate, Filter, Pivot and Aggregator stages for various transformations needed on data to conform to business rules.
- Develop sequence jobs proper job dependencies, job control stages, triggers and notifications.
- Extensively used Hive connectors, had file connectors in WebHDFS mode to read/write to Hive tables. Knowledge of configuring hive connectors to run in Kerberos enabled Haddop Cluster.
- Create HDFS scripts for copying files, data file manipulation, schedule jobs, data validation etc.
- Built standard Error handling framework to capture operational statistics, exceptions in data and rejects to report or reprocess.
- Document ETL test plans, test cases, test scripts, and validations based on design specifications for unit testing, system testing, functional testing.
- Participated in daily standup calls and scrum meetings and was responsible to create User Stories and tasks in Rally.
Confidential, NJ
Senior ETL Developer
Responsibilities:
- Responsible for developing data integration strategies, source-to-target mappings, data lineage, and complex data transformation logic.
- Extensively used PL/SQL tables and Bulk loader programs for processing and loading data into oracle.
- Analyze current data flows and prepare mapping and lineage documents applying reverse engineering techniques.
- Created and maintained Conceptual, Logical, and Physical models defining business requirements.
- Created FRDs, Mapping documents, Databases based on the requirements.
- Find the root causes of data quality issues, suggest solutions, and test fixes.
- Designed and developed complex Parallel jobs and Sequences for staging the data from source tables and then loading it to the Integration layer and further to a Transfer Hub layer.
- Designed and developed operational reporting jobs which read the generated job reports and extract operational statistics to load the Operations tables.
- Standardized Client Name/Address/Phone data using Qualitystage and CASS (USAC) stage.
- Participated in daily/weekly team meetings. Worked with Business Analysts/Stakeholders to develop the Functional Design Documents (FDD).
- Wrote Technical Design Documents (TDD) and Unit Test Plans (UTP) document for testing the different cases for the migration of data.
- Parsed high-level design specification of ETL coding and mapping standards to develop new SCD Type1/Type2 complex mappings, fix the old mappings into different layers and propose strategies for future growth of the data.
- Peer review of the codes to meet the accurate business requirements and project standardization. Verified the unit test results.
- Tested Datastage Jobs and Sequences to figure out the bottleneck and tune them for better performance.
- Prepared Unit, Integration and System test cases for various stages to capture the data discrepancies/ inaccuracies to ensure the successful execution of accurate data loading.
- Extensively worked on CDC to capture the data changes into sources and for delta load. Used Debugger/Director Logs to validate the Mappings and gained troubleshooting information about the data and error conditions.
- Migrated the code from Dev to Test and Test to Prod environments. Wrote the migration documentation in details for system compatibility, objects and parameter files to smooth transfer of code into different environments.
Confidential
Technical Lead
Responsibilities:
- Leading a team of fifteen members for UK Based Retail Client.
- Co-ordinating with the customer to understand the requirements.
- Interacted with business during all phases of SDLC to gather/meet the requirements.
- Requirement Analysis and Business Requirement/Functional Design Specification.
- Data Modelling/High Level Design Documentation/Source to Target mapping Documentation
- Source Data Analysis and Technical Design Documentation
- Estimating the delivery efforts, schedules and timelines. Tracking the progress thereof.
- Ensuring quality of the deliverables by internal/external quality assurance methods.
- Developing CDC Subscriptions and Datastage Jobs/Sequences, Scheduling Batch Jobs
- Preparing Unit Test Plans and Unit Test Cases for the developed jobs and sequences.
- Team Management/Ensuring Timesheet Compliance
- Change Management/Enhancement, Release Management and Deployment
- Training new team members on IBM Infosphere Datastage.
Confidential
Technical Lead
Responsibilities:
- Leading a team of four for iHub LOB, co-ordinating with the customer to understand the requirements and design and develop solutions.
- Estimating the delivery efforts, schedules and timelines. Tracking the progress thereof. Ensuring quality of the deliverables by internal/external quality assurance methods.
- Developing Datastage Jobs/Sequences for new applications started within the iHub portfolio. Release and deployment of the enhancements and new jobs.
- Preparing Unit Test Plans and Unit Test Results and the Runbook for the developed jobs and sequences.
- L2 support for the ETL part of the existing iHub Applications, investigating failures, providing fixes, suggesting improvisations.
Confidential
Technical Lead
Responsibilities:
- Analysed the existing REIMS solution in place and actively involved in the requirement gathering and analysis phase of the Project.
- Prepared the Functional/Technical Design Documents for the Mappings under the six different subject areas for Ireland and UK.
- Developed Datastage Jobs/Sequences to read in/create EDIFACT standard format compliant broker messages for Renewal Invites/Claims/Signings for Confidential UK and Ireland. (30 + Datastage Jobs)
- Prepared Unit Test Plans and Unit Test Results and the Runbook for the developed jobs and sequences.
- Have high level knowledge of the Release Packaging and Version Control using PVCS.