Lead Data Engineer Through Tcs Resume
Bloomfield, CT
SUMMARY:
- Seasoned Big Data and Data warehousing Developer with experience of various tools including Hadoop ecosystem with EC2 AWS cloud computing, Informatica, Teradata, ODI and additional experience in Visualization tools like OBIEE, Tableau, Podium and Looker. Proficient in scratch implementations of Data Analysis, Extraction, Transformation, Loading into Enterprise Data Lakes / Warehouse / Marts and building multiple analytical Dashboards and Reports. Data Integration and Migration experience includes Scoping, Impact analysis, Design, Developing mappings and automating process. Solid background in Dimensional Modeling such as Star, Snowflake, Galaxy and worked with various databases including Teradata, Oracle and SQL Server. Expertise in handling various production incidents, issues through service manager tools like BMC ITSM and HPSM tools. Has excellent communication skills and ability to work on a team as well as independently as necessary.
TECHNICAL SKILLS:
Big Data: Cloudera Hadoop CDH4&5, HDFS, HIVE, Sqoop, Oozie, Spark, Kafka
Cloud Computing: Amazon EC2, S3, RDS, DynamoDB
ETL Tools: Informatica (Tools, MDM), ODI, Teradata, SSIS
Visualization Tools: Podium, tableau, Looker, Oracle BI, SSRS, Cognos
Databases: Oracle, Teradata, SQL Server, DB2
Programming Languages: C, SQL, PL/SQL, JAVA, Shell Scripting, Groovy, Python
Scheduling Tools: DAC, AUTOSYS, Control - M, JAMS, CA Automation (ESP)
Operating Systems: LINUX, AIX, UNIX and Windows
DevOps: Jenkins, Ansible, Maven, Kubernetes
PROFESSIONAL EXPERIENCE:
Confidential - Bloomfield, CT
Lead Data Engineer through TCS
- Prepared the Estimations and Design specifications for the requirements by having proactive discussions with the Business Analysts, Data Analysts and End users.
- Worked on Apache Hadoop ecosystem components like MapReduce, HDFS, HBase, Hive, Sqoop, Pig, Oozie, Zookeeper, Flume, Spark, Python with CDH4&5 distributions and EC2 cloud computing with AWS.
- Analyzed data which need to be loaded into Hadoop and contacted with respective source teams to get the table information and connection details.
- Used IBIS (in-house) Sqoop to import data from different RDBMS systems like Oracle, DB2 and Netezza and loaded into HDFS.
- Developed Oozie workflow schedulers to run multiple Hive and Pig jobs that run independently with time and data availability.
- Developed Data warehousing & Reporting solutions using Informatica, Oracle Data Integrator (ODI), SSIS, Tableau, OBIEE, SSRS technologies.
- Performed administration activities like creating users, roles, responsibilities and backups for repository, topology. Purging the session logs and clean-up the stale sessions.
- Hands-on experience in visualizing the metrics data using Podium Data.
- Experience setting up and configuring Jenkins pipeline for Continuous testing and Continuous Deployments for all the non-prod and prod environments.
- Prepared detail migrations tasks for the code migration from Lower to Higher instances
- Troubleshooting the load performance, data issues and then providing detail root cause analysis
- Scheduling and monitoring of the Full and Incremental load cycles.
- Effective coordination with Source, DBAs and other Up/down steam teams during Implementation and support phases.
- Implemented the Proof of Concept on the new solutions proposed for business requirements.
Environment: Cloudera Hadoop CDH5, AWS EC2&S3, Informatica, Oracle Data Integrator (ODI), Oracle BI, Podium Data, Looker, ERWIN, Python, Jenkins, Docker, Ansible, HDFS, Oracle, SQL Server, Flat Files, XMLs, SQL Server, Unix Scripting, Groovy, Python, Putty, SQL, PL/SQL, SAFe Agile Framework.
Confidential - Portland, ME
Sr. ETL Developer through CSC
- Closely worked with Business Analyst and Architects to clarify Requirements and then translate them into technical specifications.
- Implemented e mappings using Normalizer and other transformations to process data from COBOL VSAM files by preparing respective Informatica Copybook for each and every file and then propagating data into Teradata warehouse.
- Developed CDC (Change Data Capture) mechanism using Informatica Power Exchange for Real time analytics by capturing live updates in OLTP / Source systems.
- Experienced in using various transformations like Aggregator, Filter, Expression, Sequence Generator, Update Strategy, Joiner, Lookups, Union and XMLs to develop robust mappings.
- Implemented Push down optimization (PDO) technique and used to Teradata utility adapters (BTEQ, MLOAD, FLOAD, FEXTRACT and TPT) to minimize time for loading process.
- Used Workflow Manager and Workflow Monitor for scheduling ETL jobs, repository creation/Backup/Restore using PMREP, Creating repository users/roles/groups.
- Extensively worked on SQL tuning process to increase source qualifier throughput by analyzing queries with explain plan, creating new indexes, partitions and Materialized views.
- Troubleshooting memory related issues like DTM, Line Sequential buffer size, Stack over flow in Informatica and Database related spaces issues.
- Worked with Informatica support team to resolve Environment and Product bugs.
Environment: Informatica 9.5, Informatica Power Exchange (CDC), Informatica Data Quality (IDQ) 9.5, Informatica Master Data Management (MDM) 9.5, Teradata 14.00, Erwin, Control-M, SQL Server, DB2, Qlikview, ERWIN, TOAD, Teradata SQL Assistant, Oracle 11g, Flat Files, XMLs, COBOL Files, Mainframe JCLs, Hadoop, HIVE, PIG, Map Reduce, HBASE, AUTOSYS, Unix Scripting, Erwin, LINUX, PL/SQL.
Confidential - Los Angeles, CA
ETL & BI Developer through CSC
- Requirement gathering and preparing low level and high-level design documents after doing proper impact analysis for new and enhancement changes.
- Build efficient ETL Process for processing data warehouse facts &dimension tables through complex transformations and SCD type 1 and SCD type 2 implementations.
- Customized OOB Out Of the Box mappings and also developed new mappings, sessions, workflows based on business requirements.
- Hands on experience on various transformations like Qualifier, Filter, Expression, SQ Generator, Update Strategy, Joiner, Lookups, Unions and Aggregator to develop robust mappings.
- Built new DAC execution plans for scheduling by customizing Out of Box Subject areas or creating new Subject areas with required tables for triggering different ETL Loads.
- Performance tuning of Informatica mappings using various components like Parameter files, Variables, Dynamic Cache and usage of Transformations.
- Customized and administered Physical layer, Business Model Mapping layer (BMM) and Presentation layer of OBIEE repository.
- Experienced in running daily production loads, experienced in reconciling issues on spot by analyzing bad / reject records and log files.
- Publishing the load reports to the business stakeholders on a weekly basis.
- Conducted trainings to end users on new functionality before go live.
Environment: Informatica Power center 8.1.1, Informatica Power Exchange, Informatica Quality Center (IDQ), Erwin, Informatica Metadata Manager (MDM), ODI, TOAD, PL/SQL, SAP BO, Data Stage, Flat Files, BI Publisher, OBIA 7.9.6.X, OBIEE 11g, Oracle EBS (Forms, Reports), PeopleSoft CRM, Unix Scripting, Shell Scripting, DAC, Oracle 11g, HP Quality center, PVCs (Poly version control system).
Confidential - O’FALLON, MO
ETL/BI Application Support
- Actively participated during design discussions with customer and prepared both High level and Low-level design HLD&LLD documents.
- Implemented few complex mappings with all kinds of transformations and customized Out of Box mappings to match requirements.
- Extensively used Informatica Transformations like Union, Look Up, Aggregator, Filter, Router, Joiner, Sequence Generator, Expression, Update Strategy, Stored procedure and Rank.
- Built new DAC execution plans using Out of Box Subject areas and also fixed issues occurred, while triggering different initial ETL Loads.
- Customized and administered Physical layer, Business Model Mapping layer and Presentation layer with OBI Apps repository. Worked on creating Segments used for Campaign Management.
- Enhanced performance of Reports, Segments and Segment trees by cache management.
- Created Security settings in Oracle Administration tool to set up groups, access privileges and also managing security for groups in BI Answers.
- Responsible in Managing, scheduling/triggering different DAC loads in production box.
- Experienced in Automation of loads using Shell script programming (PMCMD).
- Performed Peer to Peer code reviews for code quality and standardization.
Environment: Informatica Power Center 8.6.1, Informatica MDM, Informatica Power Exchange, OBIA 7.9.6.X, OBIEE 10.1.3.X, Siebel 7.8 CRM, Siebel Tools, TOAD, Flat Files, Oracle, UNIX Shell Scripting, File Transfer Protocol (FTP), WebLogic, SQL, PLSQL, DAC, Sun Solaris, HP Quality Center.
Confidential - New York, NY
ETL & BI Developer/Support
- Configured and Implemented the OBI sales analytics by sourcing the data from Siebel CRM
- Created new Mappings, Mapplets, Transformations, Sessions and Workflows in Informatica.
- Defined Target Load Order Plan and Constraint based loading for loading data into Target.
- Scheduled and monitored loads by using DAC (Data Administration Console)
- Configured measures in Business Model Layer and mapped to multiple logical source tables.
- Experienced in both Object and Data level security as per business requirement.
- Developed reports with different kinds of pivot, chart, compound and column selector views.
- Worked on Analytics iBots to send messages conditionally to Groups / Selective Recipients.
- Analyzed and resolved defects logged during ST, SIT and UAT test cycles.
- Conducted KT sessions post production on new changes to support team.
- Monitoring and scheduling of the production jobs on a weekly basis and reconciling the issues by coordinating with upstream and downstream teams.
- Handled the Reporting, Data and access related user issues during the Hyper-care period.
Environment: Informatica 7.1.5, OBIEE 10.1.3.3, Oracle 10g, Siebel Sales CRM, Siebel Tools, HP Quality Center, BMC Remedy ITSM 7.0, Windows, UNIX, TOAD, SQL, PL/SQL, DAC, Flat Files, MS Visio.
Confidential, Lexington, KY
Application Support Analyst
- Design and Develop Mappings and session as per technical Design specifications.
- Used various Transformations like Expression, filter, Joiner, Router and Lookups for better data Propagation, migrate clean and consistent data without any redundancy.
- Created and executed Unit test case, DB Scripts for new maps and executed loads.
- Used existing connection pools to import new physical tables, defined joins and designed Physical, BMM layers of repository.
- Developed different kinds of Reports (pivots, charts, compound layout) using global, Local Filters and assigned them to application specific dashboards.
- Worked on setting up Users, groups, roles, privileges and cache management.
- Handled and fixed various support issues in both Siebel Analytics and Siebel CRM.
- Promptly acknowledged to all owned Service Requests in SLA.
Environment: Informatica 7.1, Siebel Analytics 7.8.x, Oracle 9i, Siebel CRM 7.8, AIX 5.1, BMC Remedy ITSM tool 6.0, Buxfer quality tool and Waterfall software development model.