Informatica Developer Resume
Columbus, OH
SUMMARY
- Over 5 Plus years of progressive hands - on experience in analysis, ETL processes, design and development of enterprise level data warehouse architectures, enterprise level large databases, designing, coding, testing, integrating ETL.
- Proficient in understanding and detailing the business requirements and had experience in interacting with business users to clarify requirements and translate the requirements into technical and user stories.
- Experience in Dimensional data modelling techniques, Slow Changing Dimensions (SCD), Software Development Life Cycle (SDLC) (Requirement Analysis, Design, Development & Testing), and Data warehouse concepts - Star Schema/Snowflake Modelling, FACT & Dimensions tables, Physical & Logical Data Modelling.
- Experience in integration of various data sources like Oracle 12c,11g,10g, MS SQL Server, XML files, Teradata, Netezza, DB2, Flat files, Salesforce, Mainframe sources into staging area through ETL Process and loaded data into different target Enterprise level databases.
- Expertise in developing large integrated complex mappings with standard transformations and complex transformations like SQL transformation, Java Transformation for calling API’s, Store-procedures, Flexible Target Key transformation (AI, BI modes) also developed complex re-usable chunk of mappings called Mapplets, and Worklets and re-usable and non-reusable transformations.
- Extensively worked on ETL Automation Testing using Ruby, done unit testing and acceptance testing through Automation and captured major and minor bugs during code development and delivered highly efficient Business deliverables.
- Hands on Experience in SQL, DB2 SQL, TSQL, PL/SQL code and done Perl scripting to validate the record counts in the xml’s, getting load event id to process, updating the status of events for the load event id and building informatica parameter file and executing the informatica workflows using Perl scripts.
- Also developed UNIX shell scripts to run batch jobs and Informatica workflows from Unix server and automated the entire ETL test ruby specs through Jenkins that is integrated with the docker, so units and acceptance will run on daily basis.
- Extensively worked on InformaticaData Quality (IDQ) projects and worked on InformaticaData Quality transformations like Address validator, Parser, Labeller, Match, Exception, Association, Standardizer and other significant transformations. As a Data Quality developer initiated the process of Data profiling by profiling different formats of data from different sources.
- Expertise in maintaining the different Versions of the code in GitHub and expertise in maintaining Project and Feature branches and hands on experience in deploying the code from Development server to Test Region or to Production server using Urban Code Deployment tool (UCD).
- Designed complex Mappings and enhanced the performance by identifying the bottle necks in Performance tuning at target level, source level, mapping and session level and worked with dba to identify the bottlenecks during the transaction commits and logs.
- Maintained best practice of writing tests first and developing code next a Test-Driven Development methodology TDD which is associated in writing Good Failing Tests first, Developing Code next, then Passing already written Failed Tests and finally Refactoring the code.
- Hands on experience working in UNIX, LINUX and Windows environments.
- With both On-site and Off-shore experience have developed skills in system analysis, troubleshooting, debugging, deployment, Team management, prioritizing tasks and customer handling.
- Extensively worked in Informatica MDM Master Data Management concepts, Methodologies and ability to apply this knowledge in building MDM solutions.
- Hands on experience working with Teradata Manager and used different features of Teradata such as BTEQ Scripts, Fast Load, Multiload, SQL Assistant, DDL and DML commands also has a very good understanding of Teradata UPI and NUPI, secondary indexes and join indexes.
- Domain knowledge experience on Health, Finance, Insurance, Telecom Domains.
TECHNICAL SKILLS
ETL Tools: Informatica PowerCenter 10.2.0/10.0/9.6.1/9.5 , Informatica Data Quality (IDQ), Informatica Data Services(IDS) 9.6.1
Databases: Oracle12C/11g/10g, Teradata, IBM DB2, SQL Server, Netezza, IBM IDZ
Operating Systems: Unix, Linux, Windows
Programming: SQL, DB2 SQL, PL/SQL, TSQL, Perl Scripting, UNIX Shell Scripting, Ruby, Mainframe Cobol
Data Modelling: Conceptual Modelling, Logical Modelling, Physical Data Modelling, ER Modelling, Dimensional Modelling - Confirmed Dimension, Junk Dimension, Slowly Changing Dimension, Degenerated Dimension, Dimension Tables and Fact Tables
Tools: UCD (Urban Code Deployment), GitHub, Jenkins, Ruby Mine Automation Tool, Ruby, RTC, Jira, Autosys, SVN, CA Scheduling Tool, CA Workstation, ESP Job Scheduler.
Other Tools: DBeaver, Squirell SQL Client, SQL Navigator, SQL for DB2, Toad for Oracle, SQL Developer, MS Office, Quality centre, JIRA, SSIS, Advanced Query Tool, IBM Developer for Z Systems
Software Development Methodology: Agile, Waterfall.
Domain Expertise: Telecom, Insurance, Finance, HealthCare
Education: Bachelor of Technology in Computer Science
University: JNTU India
PROFESSIONAL EXPERIENCE
Confidential, Columbus, OH
Informatica Developer
Responsibilities:
- Worked in Agile development methodology environment and interacted with pre-amigos with the Business Analysts and Operations along with Requirements Analysts for better understanding the business requirements before the code development and interacted with test leads and other developers as a part of post-amigos after development regarding the code functionality.
- Migrated from Informatica PowerCenter 10.0.1 to Informatica PowerCenter 10.2.0.
- Worked on routing the incoming Source files from an inbound server using the SCP, FTP, SFTP to the Secure Unix server over the secure firewalls.
- Validated the Counts in the multiple XML file that we receive from our Vendors and merged them into one consolidated XML through Perl Scripting and developed the mappings that converts the consolidated XML files into multiple flat files based on entity’s.
- Used Flat files, XML’s, RDBMS as a source and loaded into RDBMS Work tables for staging and Archive tables that are used to maintain the history Information of all the transactions.
- Developed very complex mappings using advanced and reusable & non-reusable transformations like Source Qualifier, Aggregator, Router, Joiner, Union, Expression, Lookup, Filter, Update Strategy, Sequence Generator and stored Procedures etc. and calculated hundreds of Business rules and synced the data to Nationwide’s Enterprise Legacy Databases as a process of Provide phase that provides the data to the business agencies and business vendors.
- Extensively worked on Reusable sessions, Mapplets, Worklets also created several types of tasks like Assignment, Decision, Event Wait, Event Raise, Email Task and Scheduled Task based on Client requirement.
- Worked in Test Driven Development methodology TDD which is associated in writing Good Failing Automation Tests first, Developing ETL Code next, then Passing already written Failed Tests scripts and finally Refactoring the ETL code.
- Developed unit test scripts and Acceptance test scripts using JetBrains Ruby Automation tool and captured major and minor bugs during code development & developed table models, hooks, helpers, ruby specs and Acceptance features in Gherkin language that are integrated with ruby step definitions in Ruby.
- Developed Test Cases for Deployment Verification,ETLData Validation and Report testing.
- TestedInformaticaETLmappings that extracts the data from different source systems and loads in to Oracle and DB2 targets using Automation testing.
- Tested the data load process during Batch Load, Sequential Loads, Incremental Loads & Direct Dumps.
- Tested and Validated the Business tables functionality and written complex SQL queries to compare the data that got synced to the target systems with in the business validation.
- Written the Automation test scripts, experience in designing the test cases & involved in test plan with Test leads and done back-end testing and end-to-end acceptance testing, regression testing and user-acceptance testing.
- Maintained the versioning of ETL code in GitHub Repo by adding code commits to the Project branch, opening pull requests, resolving deltas in code, merging them to a project feature branch and have done forking the project branch for research purposes.
- Integrated all the Unit Automation test scripts and Acceptance features to the Jenkins Automation server so the Jenkins will kick off all the designated Automation test scripts in the desired time intervals.
- Deployed the code from the Dev Region to Test Region and to Production server using UCD Urban Code Deployment server.
- Performed Data validation for the dry runs before the production release using complex SQLs using joins, sub queries and correlated sub queries, Views, Case Statements to perform the Cross Verification of data to make sure all the modules are integrated correctly, and the code is functioning as per the Business scenarios.
- Extensively used Teradata Utilities like Fast-Load, Multi-Load, Fast-Export and UPSERT while loading data into the target tables in Teradata Database.
- Extensively worked with Teradata Fast load utility to load huge tables for initial loads and truncate and load also worked with Teradata multi load utility for insert, update or incremental loads.
- Used Teradata for writing scripts in Teradata database and Worked on creating and managing the partitions and performed database health checks and tuned the databases using Teradata Manager.
- Wrote complex TSQL Queries to perform data validation and graph validation to make sure test results matched back to expected results.
- Hands on experience in Teradata Macros and Error Handling and Performance tuning of Teradata SQL statements and analysing the data distribution among AMPs and index usage, collect statistics, definition of indexes, revision of correlated sub queries, usage of Hash functions, etc
- Involved in resolving Production issue tickets related to Job failures, Data issues, Code Change, and Performance related issues, etc.
- Created UNIX shell scripts for Informatica post and pre-session operations, database administration and day-to-day activities like, monitor network connections and database ping utilities.
- Worked on Unix shell scripting to create job for polling and loading files, control load validation and scripts to schedule and run jobs using PMCMD Command.
- Created various UNIX Shell Scripts for scheduling various data cleansing scripts and loading process. Maintained the batch processes using Unix Shell Scripts.
- Worked on Performance Tuning to optimize the Session performance by utilizing, Partitioning, Push down optimization, Index Cache, Data Cache and incremental aggregation.
- Scheduled Informatica workflows using CA Workstation ESP scheduler.
- Involved in Production Support by performing Normal Loads, Bulk Loads, Initial Loads, Incremental Loads, Daily loads and Monthly loads and extracted reports based on business issues.
Environment: Informatica PowerCenter 10.2.0, Oracle12C, IBM DB2, Toad for Oracle 12.8, Teradata SQL Assistant 15.0, Unix, Putty, WinSCP 5.9.4, JetBrains Ruby Mine Automation 2017.3.1, Ruby, Git Hub, Jenkins, Docker, Toad Data modeler 5.5, Mainframe, IBM Developer for Z Systems, Advanced Query Tool v10.0.4, DBeaver, DB Visualizer 10.0.7, CA Workstation ESP
Confidential, Sunnyvale, CA
Informatica Developer/ IDQ Developer
Responsibilities:
- Worked in Agile development methodology environment and Interacted with the users, Analysts for collecting, understanding the business requirements.
- Worked on building the ETL architecture and Source to Target mapping to load data into Data warehouse.
- Worked on variety of data sources like SQL, XML, Text, CSV files, Excelfiles, Oracle, MySQL, DB2 sources
- Developed various mappings using Mapping Designer in Informatica PowerCenter and worked with Aggregator, Lookup, Filter, Router, Joiner, Source Qualifier, Expression, Stored Procedure, Sorter and Sequence Generator, Transaction Control, SQL transformations.
- Worked with Informatica Data Quality 9.6.1 designer, Analysis and worked on data cleansing, data matching, data conversion, exception handling, and reporting and monitoring capabilities using IDQ 9.6.1.
- Used Informatica Data Quality for addresses and names clean-ups and developed error handling & data quality checksto pull out the right data.
- Created Profiles, Rules, Score Cards for Data Profiling, Classifier models, Probabilistic models and Exception record management as the process in Informatica Data Quality.
- Worked on differentInformatica Data Qualitytransformations in thedeveloper like Address Validator, Standardizer, Match, Association, Parser, Comparison, Expression transformations.
- Configured match properties match paths, fuzzy match key, fuzzy and exact match columns using IDQ.
- Worked on Data cleansing & Standardizing by configuring Address doctor component in IDQ tool.
- UsedIDQto cleanse the duplicate records or redundant records and to check the accuracy of project data.
- Worked on Address Validator transformation to compare input address data with address reference data to determine the accuracy of input addresses and fix errors in those addresses.
- Worked on complex quality rule designing, development and implementation patterns with cleanse, parse, standardization, global Address Validation, Contact Verification and format, Search Match, Match Merge.
- Written PL/SQL scripts, created stored procedures and functions and debugged them.
- Worked on Mapplets, reusable and Non-reusable transformations, also developed Complex SCD Types mappings to update slowly Changing Dimension Tables.
- Created document withMDMdata model, source system definition, data mapping and cleansing requirements, trust score and matching rule definitions.
- Implemented in MDM to ensure that the master data is validated as correct, consistent, and complete and to ensure that Master Data is circulated in context for consumption by internal or external business processes, applications, or users. Configured cleanse server and address doctor.
- Created mappings using cleanse functions to standardize data, cleansed data and validated IDD fields to display custom error message.
- ImplementedInformaticaMDMworkflow including data profiling configuration specification and coding match rules tuning migration.
- Also created IDD application by creating subject area group, subject area child for an access to view MDM hub data.
- Worked on Change Data Capture (CDC) using CHKSUM to handle any change in the data if there is no flag or date column present to represent the changed row.
- Working with PowerCenter Versioning (Check-in, Check-out), Querying to retrieve specific objects, maintaining the history of objects.
- Worked on Debugging and Troubleshooting to fix the bugs identified in System Testing and identified the bottlenecks in all levels to tune the performance and resolved the production support tickets.
- Validated the Data after the dry runs before the production release using complex SQLs to make sure the code is functioning as per the Business requirements.
- Created Unix shell scripts for parsing and moving the data files and for creating parameter files and finding and removing duplicates in the data feed using AWK commands.
- Scheduled the Informatica workflows using Control-M scheduler tool.
- Involved in unit testing, acceptance testing and User acceptance testing in the multiple environments.
Environment: Informatica PowerCenter 9.6.1, Oracle11g, Informatica Data Quality 9.6.1, Informatica MDM 9.6.1, Toad for Oracle, Advanced Query tool, IBM DB2, Control-M, Netezza, UNIX, PL/SQL, Jira, WinSCP, Putty, Squirrel SQL Client, SQL Developer
Confidential
MicroStrategy Developer/ETL Developer
Responsibilities:
- Involved in requirement gathering, user meetings, discussing the issues to be resolved and translated the user inputs into ETL design documents and understood the Business requirements and prepared a functional spec.
- Worked on Informatica PowerCenter for (ETL) extraction, transformation and loading data from heterogeneous source systems and flat files, XML files which includes fixed-length as well as delimited files.
- Developed Informatica mappings, sessions and workflows to load data from various sources into respective database as per the requirement.
- Extensively used various active and passive transformations like Filter Transformation, Router Transformation, Expression Transformation, Source Qualifier Transformation, Joiner Transformation, and Look up Transformation, Update Strategy Transformation, Sequence Generator Transformation, and Aggregator Transformation.
- Developed mapping logic using various Passive transformations like Expression, Lookups (Connected and Unconnected), Joiner, Filter, Sorter, Update strategy and Sequence generator.
- Documented Informatica mapping logics, design structures and validation rules.
- Created reusable and non-reusable Sessions and Workflows in Informatica Workflow manager
- Actively implemented Informatica performance tuning by identifying and removing the bottlenecks and optimized session performance by tuning complex mappings
- Extensively involved in unit testing, integration testing and system testing of the mappings and writing Unit and involved in System Test Plan.
- Used the debugger to test the mapping and fixed the bugs.
- Performed Tuned performance of Informatica mappings and session for large data files by increasing block size, data cache size, sequence buffer length and target commit interval.
- Migrated the code from DEV repository to QA and then to PROD Repositories while in our monthly releases.
- Created and modified Multi Loads and batch loads for Informatica using UNIX.
- Developed Batch scripts for doing automated file transfer to customer via FTP, SCP and SFTP.
- Created a UNIX shell scripts script to pass database connection values in multiple environments, and the parameter entries for sources and targets.
- Developed Data Validation and Polling processes for some of the ETL designs.
- Developed ABC (Audit, Balance and Control) processes in ETL.
- Created complex metrics with different level dimensionality, condition and transformations for user and business specific reporting needs.
- Worked extensively with filters, custom groups, prompts, consolidations and transformations to generate reports meeting business needs.