Etl Talend Developer Resume
PROFESSIONAL SUMMARY
- ETL TalendDeveloper 8+ years of experience as a Technical analyst, safety Developer and administrator.
- Extensive experience in IBM Info Sphere Data Stage - ETL integration with SAP bank analyzer functional module performing data extract, loading and transformation process for financial general ledger accounts data.
- Experience in designing and implementing Data Warehouse applications mainly using ETL tool Talend Data Fabric for Big data integration and data ingestion.
- Worked with different components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more).
- Worked with different tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
- Well versed with XML technology and DOM and SAX parsers
- Experience in UNIX file/dataset management to keep the load ready data for all financial transactional data.
- Experience on waterfall and Agile methodology project implementation.
- Involved in complete Software Delivery Life Cycle (SDLC) in large data warehouse environment for financial data system.
- Thorough knowledge in data warehousing, dimensional modeling, data integration, data virtualization, data synchronization, star schema, waterfall schema, ETL development & performance tuning, BI data analysis, SAP integrating, DFS & HDFS cluster segregation.
- Worked with senior Bio-statisticians and clinical data managers to provide ETL programs in analyzing the data, generating safety and efficacy loads, summary tables.
- Excellent oral and written communication skills. A quick learner with an eye for detail and excellent analytical and quantitative skills. Ability to work independently and adept at managing multiple competing tasks.
PROFESSIONAL EXPERIENCE:
Confidential
ETL Talend Developer
Responsibilities:
- Architect, design, construct, test, tune, and deploy ETL infrastructure based on the Hadoop ecosystem based technologies.
- Work closely with administrators, architects, and application teams to insure applications are performing well and within agreed upon SLAs.
- Work closely with Management and Data Scientist teams to achieve company business objectives.
- Collaborate with other technology teams and architects to define and develop solutions.
- Deploy ETL code that aligns with the ETL target state architecture standards and development standards.
- Research and experiment with emerging ETL technologies and tools related to Big Data.
- Contribute to the Big Data open source ecosystem.
- Work with the team to establish and reinforce disciplined software development, processes, standards, and error recovery procedures are deployed; ensuring a high degree of data quality.
- Maintain, tune, and support the ETL platform on a day-to-day basis to insure high availability.
Confidential
ETL Talend Developer
Responsibilities:
- Developed complex ETL jobs from various sources such as SQL server, Postgressql and other files and loaded into target databases using Talend ODS ETL tool.
- Created Talend jobs using the dynamic schema feature.
- Created the Talend jobs for Store Support Center, Coupons jobs and Promotions etc.
- Interact with business community and gathered requirements based on changing needs.
- Incorporated identified factors into Talend jobs to build the Data Mart.
- Performance tuning - Using the tmap cache properties, Multi-threading and Parallelize components for better performance in case of huge source data.
- Tuning the SQL source queries to restrict unwanted data in ETL process.
- Involved in Preparing Detailed design and technical documents from the functional specifications.
- Prepared low level design documentation for implementing new data elements to EDW.
- Provided Knowledge Transfer to the end users and created extensive documentation on the design, development, implementation, daily loads and process flow of the mappings.
- Have used AWS components (Amazon Web Services) - Downloading and uploading data files (with ETL) to AWS system using S3 talend components.
- Used more components in Talend and Few to be mentioned: tMap, tFilterRow, tjava, toracle, txmlMap, tdelimited files, tlogrow, tlogback, components etc. in many of my Jobs Design
- Worked on Joblets (reusable code) & Java routines in Talend.
- Design, Develop and Test ETL processes in order to meet project requirements
- Created Projects in TAC and Assign appropriate roles to Developers and integrated SVN (Subversion).
- Used to be On call Support if the Project is deployed to further Phases
- Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis (Cron Trigger)
Environment: Talend Open studio V (5.6), UNIX, AWS-S3, Microsoft SQL Server management Studio, Postgres SQL, Netezza Database, Oracle Database, XML Processing.
Confidential
ETL Talend Developer/Admin
Responsibilities:
- Participated in all phases of development life-cycle with extensive involvement in the definition and design meetings, functional and technical walkthroughs.
- Created Talend jobs to copy the files from one server to another and utilized Talend FTP components.
- Created and managed Source to Target mapping documents for all Facts and Dimension tables.
- Design, develop, validate and deploy the Talend ETL processes for the DWH team using HADOOP (PIG, HIVE) on Hadoop.
- Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
- Extensively used tMap component which does lookup & Joiner Functions, tjava, tOracle, txml, tdelimtedfiles, tlogrow, tlogback components etc. in many of my Jobs Created and worked on over 100+components to use in my jobs.
- Developed multiple tracks (Modules) for the data migration between multiple different systems like MRA, TIPP to NRSC, LIM to JAS, EPM to MR, ONW to TIPP.
- Used ETL methodologies and best practices to create Talend ETL jobs. Followed and enhanced programming and naming standards.
- Involved in Data Extraction from Flat files and XML files using Talend by using Java as Backend Language.
- Created and deployed physical objects including custom tables, custom views, stored procedures, and Indexes to SQL Server for Staging and Data-Mart environment.
- Develop ETL mappings for various Sources (.TXT, .CSV, XML) and also load the data from these sources into relational tables with Talend Enterprise Edition.
- Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, Spark and loaded data into HDFS.
- Experience in using DOM4J parsing and XML Bean to process, validate, parse and extract data from XML files.
- Implemented installation and configuration of multi-node cluster on cloud using AWS on EC2.
- Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more).
- Copy data to AWS S3 for storage and use COPY command to transfer data to Redshift. Used Talend connectors integrated to Redshift.
- Developed jobs to expose HDFS files to Hive tables and Views depending up on the schema versions.
- Imported data from RDBMS (MySQL, Oracle) to HDFS and vice versa using Sqoop. (Big Data ETL tool) for Business Intelligence, visualization and report generation.
- Worked on Talend Administration Console (TAC) for scheduling jobs and adding users.
- Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
- Automated SFTP process by exchanging SSH keys between UNIX servers. Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.
Environment: Talend Data Integration 6.4.0, Talend Enterprise Big Data Edition, TalendAdministrator Console, XML, Oracle 11g, Hive, HDFS, Sqoop, SQL Navigator, Toad, Control M, Putty, Winscp.
Confidential
ETL Talend Developer
Responsibilities:
- Interacted with business team to understand business needs and to gather requirements.
- Designed target tables as per the requirement from the reporting team and also designed Extraction,
- Transformation and Loading (ETL) using Talend.Worked on Data Integration from different source systems.
- Created Technical Design Document from Source to stage and Stage to target mapping. Worked with Talend Studio (Development area) & Admin Console (Admin area).
- Created Java Routines, Reusable transformations, Joblets using Talend as an ETL Tool.
- Created Complex Jobs and used transformations like tMap, tOracle (Components), tLogCatcher, tStatCatcher, tFlowmeterCatcher,
- File Delimited components and Error handling components (tWarn, tDie).
- Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
- Identified performance issues in existing sources, targets and Jobs by analyzing the data flow, evaluating transformations and tuned accordingly for better performance.
- Manage all technical aspects of the ETL Jobs process with other team members.
- Developed mappings to load Fact and Dimension tables, SCD Type 1 and SCD Type 2 dimensions and Incremental loading.
- Created contexts to use the values throughout the process to pass from parent child to child jobs and child to parent jobs.Worked on Joblets (reusable code) & Java routines in Talend.
- Expertise in Service Oriented Architecture (SOA) and Involved in publishing Web Services that include components like WSDL, SOAP, UDDI.
- Performed Unit testing and created Unix Shell Scripts and provided on call support. Schedule Talend Jobs using Job Conductor (Scheduling Tool in Talend) - available in TAC.
- Used XML for ORM mapping relations with the java classes and the database.
- Retrieved data from Oracle and loaded into SQL Server data Warehouse.
- Created many complex ETL jobs for data exchange and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structure.
- Monitoring the Data Quality, generating weekly/monthly/yearly statistics reports on production processes - success / failure rates for causal analysis as maintenance part and Enhancing existing production ETL process.
Environment : Talend Platform 6.2, Big Data, UNIX, Oracle, XML, Oracle, TAC.
Confidential
ETL Talend Developer
Responsibilities:
- Worked closely with Business Analysts to review the business specifications of the project and also to gather the ETL requirements.
- Created Talend jobs to copy the files from one server to another and utilized Talend FTP components.
- Created and managed Source to Target mapping documents for all Facts and Dimension tables
- Analyzing the source data to know the quality of data by using Talend Data Quality.
- Involved in writing SQL Queries and used Joins to access data from Oracle, and MySQL.
- Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project from development to testing environment and then to production environment.
- Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.
- Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
- Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more)
- Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
- Experienced in using debug mode of talend to debug a job to fix errors.
- Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend Integration Suite.
- Conducted JAD sessions with business users and SME's for better understanding of the reporting requirements.
- Developed Talend jobs to populate the claims data to data warehouse - star schema.
- Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis.
- Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
- Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.
Environment: Talend Data Integration 5.5.1, Talend Enterprise Big Data Edition 5.1, XML, Talend Administrator Console, MS SQL Server 2012/2008, Oracle 11g, Hive, HDFS, Sqoop, TOAD, UNIX.
Confidential
Software Engineer - Production Support
Responsibilities:
- Interacted with Data Modellers and Business Analysts to understand the requirements and the impact of the ETL on the business.
- Designed ETL specification documents for all the projects. Created Tables, Keys (Unique and Primary) and Indexes in the DB2 server.
- Extracted data from Flat files, DB2, SQL and Oracle to build an Operation Data Source. Applied business logic to load the data into Global Data Warehouse.
- Extensively worked on Facts and Slowly Changing Dimension (SCD) tables.
- Extensively used the Add Currently Processed Flat File Name port to load the flat file name and to load contract number coming from flat file name into Target.
- Worked on complex Source Qualifier queries, Pre-and Post SQL queries in the Target.
- Extensively used workflow variables, mapping parameters and mapping variables.
- Created sessions, batches for incremental load into staging tables and scheduled them to run daily.
- Prepared migration document to move the mappings from development to testing and then to production repositories.
Environment: Informatica Power Center 8.6.1/9.1.0 , Oracle 11g, SQLServer2008, IBM (DB2), MS Access, Windows XP, Toad, Tidal, SQL developer.