Etl/talend Developer Resume
Minneapolis, MN
SUMMARY
- Around 8 years of IT experience in all aspects of Analysis, Design, Testing, Development, Implementation and Support of Relational Database, Data Warehousing Systems and Data Marts in various domains.
- Expertise in Data Warehouse/Data mart, ODS, OLTP and OLAP implementations teamed wif project scope, Analysis, requirements gathering, data modeling, Effort Estimation, ETL Design, development, System testing, Implementation, and production support.
- 7+ years of experience using Talend Integration Suite (7.2/6.3/6.1) / Talend Open Studio (7.2/6.1/5.x) and experience wif Talend Admin Console (TAC) and TMC.
- Extensively created mappings in Talend using tMap, tJoin, tReplicate, tparallelize, tConvertType, tFlowToIterate, tAggregate, tSortRow, tFlowMeter, tLogCatcher, tRowGenerator, tNormalize, tDenormalize, tHashInput, tHashOutput, tJava, tJavarow, tAggregateRow, tWarn, tLogCatcher, tMysqlScd, tFilter, tGlobalmap, tDie,tFTP components,tRESTClient, tRESTRequest, tRESTResponse etc.
- Extensive experience in using Talend features such as context variables, triggers, and connectors for Database and flat files like tMySqlInput, tMySqlConnection, tOracle, tMSSQLInput, TMsSqlOutput, tMysqlRow, tFileCopy, tfileInputDelimited, tFileExist.
- Experienced in working wif Informatica Cloud and Informatica’s DE product suite like DEI, DES, DEQ.
- Experience in various Hadoop distributions like Cloudera, Hortonworks, Databricks and EMR.
- Solid knowledge in integrating ETL tools like Informatica/Talend wif Hadoop and Spark platforms to build complex ETL pipelines
- Efficient in performance tuning and debugging complex ETL jobs - Spark tuning, Hive query tuning, data portioning etc.
- Experience working wif Data Warehousing Concepts like OLAP, OLTP, Star Schema, Snowflake Schema, Logical/Physical/ Dimensional Data Modeling.
- Extracted data from multiple operational sources for loading staging area, Data warehouse, Data Marts using SCDs (Type 1/Type 2/ Type 3) loads.
- Expertise in Data modeling techniques like Data Modeling- Dimensional/ Star Schema and Snowflake modeling, Slowly Changing Dimensions (SCD Type 2)
- Experience in using AWS cloud components and connectors to access AWS s3, Redshift, Parameter store.
- Experience in parsing complex X12 EDI files using Talend Data Mapper
- Created mappings using Lookup, Aggregator, Joiner, Expression, Filter, Router, Update strategy and Normalizer Transformations. Developed reusable Transformation and Mapplets
- Strong Experience wif shell scripting, understanding of approaches for business intelligence, data warehouse.
- Experience in submittingTalendjobs for scheduling usingTalendscheduler which is available in teh Admin Console
- Good knowledge in Talend Integration experience wif AWS/Azure
- Experience building and optimizing big-data data pipelines and data sets including Postgres, AWS Relational Data Service
- Experience in data migration wif data from different applications into a single application
- Responsible for Data migration from MySQL Server to Oracle Databases
- Prepared migration document to move teh mappings from development to testing and then to production repositories
- Experienced in writing SQL Queries and used Joins to access data from Oracle, and MySQL
- Implemented Error Logging, Error Recovery, and Performance Enhancement’s reated Audit Process (generic) for various Application teams
- Experienced in Waterfall, Agile methodologies
- Hands-on experience across all stages of Software Development Life Cycle (SDLC) including business requirement analysis, data mapping, build, unit testing, system integration and user acceptance testing
TECHNICAL SKILLS
Databases: Netezza, Teradata & utilities (FASTLOAD, MULTILOAD), Oracle12c/11x/10g, DB2, Microsoft SQL Server
Programming: T-SQL, PL/SQL
Scripting: UNIX shell scripting
Languages: SQL, Core Java, Python
Other Tools: Autosys, Control-M, Apache Airflow, Jenkins, Maven
Big Data: Cloudera 5.x, 6.x, Hortonworks 2.x, 3.x, Databricks 6.x, EMR 5.x
ETL Tools: Talend Big Data 6.x., 7.x, Talend Data Mapper, Informatica Data Engineering 10.x, IICS(Informatica Cloud)
Defect /Project Management Tools: JIRA
Version Control Tool: Bit Bucket, GitHub
PROFESSIONAL EXPERIENCE
ETL/Talend Developer
Confidential, Minneapolis, MN
Responsibilities:
- Designed ETL process using Talend Tool to load from Sources to Targets through data Transformations.
- Monitored and supported teh Talend jobs scheduled through Talend Admin Center (TAC)
- Developed teh Talend mappings using various transformations, Sessions and Workflows.
- Used Talend to Extract, Transform and Load data into Snowflake Data Warehouse from various sources like Oracle and flat files.
- Designed and developed Talend Jobs using big data Hadoop Talend ETL components, AWS S3 Buckets and AWS Services to migrate data from Teradata and Vertica to Redshift.
- Created complex mappings by using different transformations like Filter, Router, lookups,
- Created complex mappings by using different transformations like Filter, Router, lookups, Stored procedure, Joiner, Update Strategy, Expressions and Aggregator transformations.
- Implemented (CDC)Change Data Capture technology in Talend to load deltas to a Data Warehouse/Redshift.
- Created multiple Job lets (reusable code) & Java routines in Talend
- Experienced in writing SQL Queries and used Joins to access data from Oracle, and MySQL.
- Created mapping documents to outline data flow from sources to targets
- Involved in Dimensional modeling (Star Schema) of teh Data warehouse, used Erwin to design teh business process, dimensions, and measured facts.
- Designed teh Talend ETL flow to load teh data into hive tables and create teh Talend jobs to load teh data into HDFS files and Hive tables.
- Developed Talend Bigdata jobs to load heavy volume of data into S3 data lake and then into Redshift using bulk load components.
- Used tRest components to call GET, PUT and POST services on RestAPI.
- Worked on a module where Informatica is teh ETL tool - built various DEI mappings, dynamic mappings.
- Used ISD to parse complex file patterns and also leveraged advanced transformations like mid-stream parsing on Hadoop using Informatica DEI/DES.
- Developed Talend job to run PySpark scripts which pushes to Big Data where data is stored in Hive tables.
- Unit testing, code reviewing, moving in UAT and PROD.
- Working wif high volume of data and tracking teh performance analysis on Talend job runs and session.
- Prepare teh Talend job level LLD documents and working wif teh modeling team to understand teh Big Data Hive table structure and physical design.
- Conducted code reviews developed by my teammates before moving teh code into QA.
- Used Talend reusable components like routines, context variable and global Map variables.
- Modified existing mappings for enhancements of new business requirements.
- Prepared migration document to move teh mappings from development to testing and then to production repositories.
- Works as a fully contributing team member, under broad guidance wif independent planning & execution responsibilities.
- Scheduling teh ETL mappings on daily, weekly, monthly and yearly basis.
- Worked on teh project documentation and prepared teh Source Target mapping specs wif teh business logic and involved indatamodeling
Environment: Talend for Big Data 7.1, Talend Data Mapper, Talend Administrator Console, Informatica DEI(S) 10.2.x, AWS EMR, MS SQL server, PostgreSQL, SQL Workbench, WinSCP
Talend Developer
Confidential, Arlington, VA
Responsibilities:
- Participated in all phases of development life-cycle wif involvement in teh definition and design meetings, functional and technical walkthroughs.
- Created Talend jobs to copy teh files from one server to another and utilized Talend FTP components
- Created and managed Source to Target mapping documents for all Facts and Dimension tables
- Used ETL methodologies and best practices to create Talend ETL jobs. Followed and enhanced programming and naming standards.
- Created and deployed physical objects including custom tables, custom views, stored procedures, and Indexes to SQL Server for Staging and Data-Mart environment.
- Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.
- Created Implicit, local and global Context variables in teh job. Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
- Developed stored procedure to automate teh testing process to ease QA efforts and also reduced teh test timelines for data comparison on tables.
- Created Talend jobs to copy teh files from one server to another and utilized Talend FTP components.
- Experience in CI/CD using Jenkins to deploy jobs and rancher to monitor logs.
- Used Talend as integration and migration tool to move salesforce data into Azure ADLS Gen2 storage and later move data to Synapse for BI teams to build dashboards.
- Worked on migration of talend jobs from older versions to teh latest.
- Involved in production and deployment activities, creation of teh deployment guide for migration of teh code to production.
Environment: Talend Open Studio 6.3, Talend Open Studio 7.1, Rancher, Jenkins, Oracle 11g, Microsoft Azure, Databricks 5.x, Synapse, SQL Navigator, Salesforce, Toad, Putty, WinSCP.
Talend Developer
Confidential
Responsibilities:
- Worked closely wif Business Analysts to review teh business specifications of teh project and to gather teh ETL requirements.
- Closely worked wif Data Architects in designing of tables and even involved in modifying technical Specifications.
- Involved in Extraction, Transformation and Loading of data from multiple source systems to Aws S3.
- Involved in teh Development of copying Data from AWS S3 to Redshift using teh Talend Process.
- Involved in writing custom copy command query’s and used teh context extensively and implemented in tRedshiftRow Component.
- Extensively used tsystem component to push large sets of data to s3.
- Developed a Talend Code for S3 Tagging in teh Process of Moving data from source to S3
- Utilized Talend components like tS3Put, tS3Get, tS3File List, tRedshiftRow, tRedshiftUnload, tRedshiftBulkExec, tflowtoiterate, tSetGlobalVar, tsnowflakeinput, tsnowflakeoutout, tsnowflakerow.
- Experienced in creating external staging tables in snowflake to move data between snowflake and Aws S3.
- Experienced in writing complex snow SQL queries in tsnowflakerow.
- Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures
- Experienced in creating standard jobs. Involved in development of Big data batch Jobs.
- Extensively used Talend components tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tOracleInput, tOracleOutput, tfileList etc, ts3put, ts3get, treplicate, tsortrow, tdenormalize, tnormalize, tredshiftrow
- Utilized Big Data components like tHDFSInput, tHDFSOutput, tHiveLoad, tHiveInput, tHiveOutput, tHiveRow, tHiveConnection .
- Experienced in executing teh jobs in parallel using tparallelize component.
- Used debugger and breakpoints to view transformations output and debug mappings.
- Load and transform data into HDFS from large set of structured data /Oracle/SQL server using Talend Big data studio.
- Worked on Global variables, Context variables, and extensively used tcontextload in teh jobs.
- Experienced in creating reusable jobs for error handling.
- Experienced in tuning teh ETL Jobs for better performance.
- Extensively worked wif TAC (Talend Administrator Console) for scheduling jobs using teh execution plan.
Environment: Talend Data Integration 6.1/5.5.1, Talend Enterprise Big Data Edition 5.5.1, Talend Administrator Console, Oracle 11g, Cloudera 5.x, SQL Navigator, Toad, Putty, WinSCP.
Talend Developer
Confidential
Responsibilities:
- Participated in all phases of development life-cycle wif extensive involvement in teh definition and design meetings, functional and technical walkthroughs.
- Created Talend jobs to copy teh files from one server to another and utilized Talend FTP components
- Created and managed Source to Target mapping documents for all Facts and Dimension tables
- Used ETL methodologies and best practices to create Talend ETL jobs. Followed and enhanced programming and naming standards.
- Worked on Talend Administration Console (TAC) for scheduling jobs and adding users.
- Created and deployed physical objects including custom tables, custom views, stored procedures, and Indexes to SQL Server for Staging and Data-Mart environment.
- Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.
- Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput.
- Extensively used tMap component which does lookup & Joiner Functions, tjava, tOracle, txml, tdelimtedfiles, tlogrow, tlogback components etc. in many of my Jobs Created and worked on over 100+components to use in my jobs.
- Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more)
- Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
- Created Implicit, local and global Context variables in teh job. Worked on Talend Administration Console (TAC) for scheduling jobs and adding users.
- Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
- Developed stored procedure to automate teh testing process to ease QA efforts and also reduced teh test timelines for data comparison on tables.
- Involved in production n deployment activities, creation of teh deployment guide for migration of teh code to production, also prepared production run books.
Environment: Talend Data Integration 6.4, Talend Enterprise Big Data Edition 6.4 Talend Administrator Console, Oracle 11g, Hive, HDFS, Netezza, SQL Navigator, Toad, Control M, Putty, WinSCP.
ETL Developer
Confidential
Responsibilities:
- Worked wif Data mapping team to understand teh source to target mapping rules.
- Analyzed teh requirements, framed teh business logic, and implemented it using Talend Involved in ETL design and documentation.
- Analyzed and performed data integration using Talend open integration suite. Worked on teh design, development and testing of Talend mappings.
- Created ETL job infrastructure using Talend Open Studio.
- Worked on Talend components like tReplace, tmap, tsort and tFilterColumn, tFilterRow, tJava, Tjavarow, tConvertType etc.
- Used Database components like tMSSQLInput, tMsSqlRow, tMsSqlOutput, tOracleOutput, tOracleInput etc.
- Worked wif various File components like tFileCopy, tFileCompare, tFileExist, TFileDelete, tFileRename.
- Worked on improving teh performance of Talend jobs.
- Created triggers for a Talend job to run automatically on server. Worked on Exporting and Importing of Talend jobs.
- Created jobs to pass parameters from child job to parent job. Exported jobs to Nexus and SVN repository.
- Implemented update strategy on tables and used tJava, tJavarow components to read data from tables to pull only newly inserted data from source tables.
- Observed statistics of Talend jobs in AMC to improve teh performance and in what scenarios errors are causing Created Generic and Repository schemas.
- Developed project specific 'Deployment' job responsible to deploy Talend jar files on to teh windows environment as a zip file, later, this zip file is unzipped and teh files are again deployed to teh UNIX box.
- Also, this deployment job is responsible to maintain versioning of teh Talend jobs dat are deployed in teh UNIX environment.
- Developed shell scripts in UNIX environment to support scheduling of teh Talend jobs.
- Monitored teh daily runs, weekly runs and adhoc runs to load data into teh target systems
Environment: Talend 6.1/5.5.2, UNIX, Shell script, SQL Server, Oracle, Business Objects, Erwin, SVN, Red gate, Capterra.