We provide IT Staff Augmentation Services!

It Analyst & Etl Lead Resume

0/5 (Submit Your Rating)

CA

SUMMARY

  • Has over 16 years of experience in Architecture, Design, Development, Testing and Implementation of large scale Data warehousing projects on Teradata, Oracle & Big Data Platforms.
  • As Sr Data Engineer for Data and Business Analytics, responsible for the Solution design, Data architecture, Data modeling, Data integration and Business Analytics design and development.
  • Extensive experience in managing cross functional teams, leading the development, gathering Business Requirements, Analysis, Design, and Implementations of various functionalities of Data Warehouse, Business Intelligence, ERP applications and Big Data - Hadoop Solutions.
  • Developed a scalable architecture to support data analytics requirements for operational and analytical reporting
  • Proficient in understanding the business processes/requirements and translating them into technical requirements
  • Expert in handling large data volumes and solving complex data challenges through innovative solutions & Implementations
  • Extensively involved in Data modeling activities such as creating conceptual, logical & physical models. Well versed with both 3NF & Dimensional modeling Techniques
  • Worked directly with business teams to understand the data needs and reporting SLA’s for business.
  • Performed activities such as Source System Study and Analysis (SSA), Data Collection, Data Cleansing, Transformation, Source-to-target mappings, Data Profiling, verifying data quality & GAP Analysis.
  • Defined new ETL standards and Templates to simplify and reduce Teradata resource consumption.
  • Designed and Implemented complex NCR Logic (Net Change Reporting) for Bookings & Overlay Bookings.
  • Defined standards for Sandbox usage for business to provide quick response time.
  • Extensive experience in Big Data - Hadoop solution implementation for Operational Reporting and Analytics using HIVE, Sqoop. Good Knowledge in SPARK & SPARK SQL.
  • Used Pandas Library to process the data from various file formats like CSV, XML & JSON Structures.
  • Used Python API with Spark to process the data in Hadoop Cluster
  • Used Spark Data Frames to process the structured data in HDFS
  • Used HIVE Context to read HIVE Tables in Spark
  • Worked on offloading data processing from Teradata to Hadoop framework using HIVE, Spark & Sqoop
  • Hands on experience in ETL/ELT Solution Design and Development using Informatica 9.x, Informatica MDM, Data Stage 7.x, Talend
  • Expertise in Anaplan cloud-based business modeling and planning platform for sales, operations and finance.
  • Good Functional & Technical Knowledge in Bookings, Sales Order, Sales Credits, Overlay Sales Credits, Shipments, NCB, CTMP, Customer Validated Bookings, Sales Account View, Customers & IAM Subject Areas
  • Worked as an Subject Matter Expert (SME) to provide solution to Business teams for their reporting needs in various functional areas
  • Worked closely with cross functional teams to effectively co-ordinate and manage Business user expectations
  • Expert in Agile software development and release management using Scrum Process. Hands on Experience with Agile software development Tools such as JIRA
  • Well experienced in executing multiple concurrent projects.
  • Performed Lead role in development and deployment of Enterprise Data warehousing projects in various domains
  • Expertise in implementing Data warehouse solutions using Teradata .Well Versed with Bteq Scripts & Teradata utilities- Fast load, Multi load, TPUMP, Fast Export & TPT
  • Hands on experience with the Performance tuning of complex SQL Queries, Informatica ETL Jobs.
  • Extensively worked on Data warehouse and Business Intelligence tools/technologies - Informatica, Data Stage Teradata, Oracle, Tableau. Business objects, OBIEE, Unix Shell Scripts, Python, UC4, $U and Tidal Scheduling Tools. ER/Studio 7.6.0, Erwin4.x Data Modeling Tools
  • Designed ETL process for Exception handling, control tables, parameterization and Reconciliation
  • Expertise in Tableau Reporting Tool.
  • Involved in project management activities such as work allocation to the offshore team, schedule tasks and coordinating and administering the offshore and onsite team and tracking the progress on daily basis, setting up the data modeling & design review sessions and work closely with SME’s and Architects for resolving the design related issues, Work with QA & Business teams for unit, integration & UAT testing Schedules, Follow up with infrastructure team, DBA & informatica support team for setting up the Dev, QA & Production Environments and resolve any environment issues which affects project timelines.
  • Familiar with Data Modeling Tools Erwin, ER/Studio and version controlling Tools: Git, Microsoft Visual Source Safe and PVCS, HP Project and Portfolio Management Work bench Tool: Kintana, Artifactory, U Deploy, Test Management & defect Tracking Tool Mercury Quality Center.
  • Demonstrated interpersonal, communication, team building, issue resolution, project coordination & technical leadership skills
  • Expertise with Unix shell scripting, Python & Java Programming.

TECHNICAL SKILLS

ETL Tools: Informatica 9.x/8.x/, Informatica MDM, DataStage and Talend

Reporting Tools: Tableau, Business Objects and OBIEE

Data Modeling Tools: ER/Studio 7.6.0, Erwin and Oracle Designer

Business modeling Tools: Anaplan

Languages: SQL, PL/SQL, Java, Python, Shell Scripting

Hadoop Framework: HIVE, HBASE, SPARK, Sqoop, SPARK SQL, PIG & MAP Reduce

Databases: Teradata V15.0/12.0/V6.0, Oracle, SQL Server

Tools: & Utilities: GIT,U Deploy, Artifactory, Microsoft visual source safe (VSS), PVCS, Kintana7.1, Mercury Quality Center 9.2, Rational Rose, Teradata Fast Load, TPUMP, Multi Load, TPT

Scheduling Tools: $Universe and Tidal Enterprise Scheduler

PROFESSIONAL EXPERIENCE

Confidential, Sunnyvale, CA

Environment: Teradata, Hadoop frame work, Spark,Python,HIVE, UC4 scheduling Tool, JIRA Software Development Tool, GIT Version Controlling Tool, Tableau Visualization Tool, Talend ETL Tool.

Sr Data Engineer

Responsibilities:

  • Analyzing the Business requirements thoroughly by working closely with multiple cross functional teams / business users.
  • Assess business problems, and translated them into specific system requirements to find technology/automation solutions that were cost effective and met business needs.
  • Create a scalable architecture to support data analytics requirements for operational and analytical reporting
  • Worked as Lead in Data Modeling, Dimensional Modeling and Physical Design of Data warehouse projects.
  • Worked on multiple concurrent projects by coordinating with global teams on providing end to end solutions delivery
  • Participate in code reviews and mentor the team for Technical as well as Domain related issues.
  • Worked as subject matter expert (SME) to provide solution to Business teams for their reporting needs in various functional areas
  • Document best practices, Continuous process improvement with bi-Monthly sessions with teams.
  • Experience in conducting analysis, designing, implementing, testing, and configuring system settings, and implementing business solutions using information technology in a large organization.
  • Involved Design, Development and implementation of Orders Functional Domain in the Global e-commerce(GeC) data warehouse for GIS team
  • Used Pandas Library to process the data from various file formats like CSV, XML & JSON Structures.
  • Used Python API with Spark to process the data in Hadoop Cluster
  • Used Spark Data Frames to process the structured data in HDFS
  • Used HIVE Context to read HIVE Tables in Spark
  • Troubleshoots business and production issues.
  • Performance Tuning of Teradata Queries to meet SLAs.
  • Develop the HIVE Scripts and load data to HIVE Tables using HDFS Frame Work.
  • Schedule HIVE Jobs through UC4 Scheduling Tool
  • Design, Develop & Test complex ETL Jobs using Tera Data.
  • Design and Develop UC4 (Automic ) jobs and workflows to schedule the batch jobs
  • Involved in Performance Tuning of Oracle SQL Queries.
  • Worked on JIRA Agile Software Development Tool.
  • Expertise in GIT Version Controlling Software
  • Created and Developed Reports and Dash boards using Tableau Visualization Tool

Confidential, Milpitas, CA

Environment: Informatica 9.5/8.x, Informatica MDM, Talend, Anaplan, SQL, Bteq, ER/Studio 7.6.0, Tidal Scheduling Tool, Tableau Reporting Tool, OBIEE, Teradata V15.0,Oracle, PVCS,GIT, UCS UNIX Server, Hadoop, HIVE,PIG, SPARK,SPARK SQL,PYTHON, FAST LOAD, TPUMP, MULTILOAD,FAST EXPORT & TPT

Sr Data Engineer

Responsibilities:

  • Analyzing the Business requirements thoroughly by working closely with multiple cross functional teams / business users
  • Create a scalable architecture to support data analytics requirements for operational and analytical reporting
  • Worked as Lead in Data Modeling, Dimensional Modeling and Physical Design of Data warehouse projects.
  • Worked on multiple concurrent projects by coordinating with global teams on providing end to end solutions delivery
  • Participate in code reviews and mentor the team for Technical as well as Domain related issues.
  • Worked as subject matter expert (SME) to provide solution to Business teams for their reporting needs in various functional areas
  • Design, Architect and implement highly efficient & scalable ETL/ELT Processes using Informatica.
  • Used Pandas Library to process the data from various file formats like CSV, XML & JSON Structures.
  • Used Python API with Spark to process the data in Hadoop Cluster
  • Used Spark Data Frames to process the structured data in HDFS
  • Used HIVE Context to read HIVE Tables in Spark
  • Design & Developed the source-to-target mappings for Enterprise Data warehouse for different source systems.
  • Created and maintained Standards and Best Practices documents for Data warehouse Design and Development
  • Worked as Lead in Data Modeling, Dimensional Modeling and Physical Design of Data warehouse projects.
  • Designed Rochard Tool to capture Metadata for Data Elements and Informatica mappings. Developed clear definitions for data elements and documented data lineage where applicable.
  • Performance Tuning of SQL queries in Teradata & Informatica workflows to meet SLAs.
  • Worked on Customer, Product, Sales Hierarchy, SAV master and reference data as well as Transaction Data Such as Bookings, Sales Order, Sales Credits, NCB & Overlay Bookings
  • Reconcile and logically categorized data collected from multiple source systems;
  • Designed Scalable Solutions in migrating large (Terabytes of Data) volume of Data to EDW Teradata Warehouse.
  • Extensively used Teradata utilities like fastload, fastexport, multiload, TPump. & TPT
  • Designed and developed Reports & Dashboards using Tableau Data visualization Reporting Tool
  • Investigating and resolving Issues and Change Requests from customer
  • Work with QA & Business teams for unit, integration & UAT testing
  • Designed the process to load data from Teradata to Anaplan Planning Tool for Dashboard Creation
  • Designed and Developed Job Groups and Jobs and schedule informatica workflows using Tidal Enterprise (TES) Scheduling Tool.
  • Actively involved in supporting UAT (User acceptance testing), Production Deployment & Normalization.
  • Follow up with infrastructure team, DBA & informatica support team for setting up the Dev, QA & Production Environments and resolve any environment issues which affects project timelines
  • Document best practices, Continuous process improvement with bi-Monthly sessions with teams.

Confidential, Milpitas, CA

Environment: Informatica 8.x, SQL, Bteq, ERWin 7.6.0, $Universe Scheduling Tool, OBIEE & Business Objects Reporting Tool, Teradata, Oracle, PVCS,UCS UNIX Server, FAST LOAD, TPUMP, MULTILOAD,FAST EXPORT.

IT Analyst & ETL Lead

Responsibilities:

  • Analyzing the Business requirements thoroughly by working closely with multiple cross functional teams / business users
  • Worked on multiple concurrent projects by coordinating with global teams on providing end to end solutions delivery
  • Participate in code reviews and mentor the team for Technical as well as Domain related issues.
  • Worked as subject matter expert (SME) to provide solution to Business teams for their reporting needs in various functional areas
  • Worked as Lead in Data Modeling, Dimensional Modeling and Physical Design of Data warehouse projects.
  • Design, Architect and implement highly efficient & scalable ETL/ELT Processes using Informatica.
  • Created and maintained Standards and Best Practices documents for Data warehouse Design and Development
  • Performance Tuning of SQL queries in Teradata & Informatica workflows to meet SLAs.
  • Designed Scalable Solutions in migrating large (Terabytes of Data) volume of Data to EDW Teradata Warehouse.
  • Extensively used Teradata utilities like fastload, fastexport, multiload, TPump.
  • Investigating and resolving Issues and Change Requests from customer
  • Work with QA & Business teams for unit, integration & UAT testing
  • Developed $U Uprocs, Sessions, Tasks, Management Units(MU’s) &Rules and schedule informatica workflows using $U Scheduling Tool.
  • Actively involved in supporting UAT (User acceptance testing), Production Deployment & Normalization.
  • Follow up with infrastructure team, DBA & informatica support team for setting up the Dev, QA & Production Environments and resolve any environment issues which affects project timelines
  • Document best practices, Continuous process improvement with bi-Monthly sessions with teams.

Confidential, Mountain View, CA

Environment: Informatica 7.1, PL/SQL, Shell Script, Business Objects, Hyperion, Erwin 4.x, Tidal Scheduling tool, Oracle 9i, Solaris-OS.

Tech Lead

Responsibilities:

  • Customer interaction and understanding the requirements
  • Responsible for Star Schema Design.
  • Analyzing, designing & developing ETL Strategies and Processes.
  • Writing ETL Specifications for developers
  • Creating & Designing of mappings using informatica7.1
  • Used the transformations like Lookup, Router, Filter, Aggregator, Stored Procedure, UPDATE strategy transformations in the development of mappings
  • Developed reusable components by using mapplets.
  • Scheduling of informatica workflows by using Tidal Scheduling tool
  • Developed Tidal Jobs and Job Groups
  • Involved in the performance tuning of SQL Queries by using optimizer hints, creating proper indexes etc.
  • Responsible for IDW Production Support
  • Involved in fixing the Production bugs
  • Working on the new enhancements for the existing Data warehouse
  • Unit Testing & Integration Testing

We'd love your feedback!