We provide IT Staff Augmentation Services!

Talend Bigdata Developer Resume

4.00/5 (Submit Your Rating)

Oak Brook, IL

SUMMARY:

  • 8+ years of experience in full life cycle of software project development in various areas like design, Applications development of Enterprise Data Warehouse on large scale development efforts leveraging industry standard using Talend and Informatica Power Centre.
  • 3+ years of experience using Talend Data Integration/Big Data Integration (6.1/5.x) / Talend Data Quality.
  • Extensive knowledge of business process and functioning of Financial, Retail, Manufacturing, Insurance, HealthCare and Technology Solutions industries.
  • Experience in XML related technologies such as XML, DTD, XSD, XSLT, DOM and SAX.
  • Expertise in reading and writing data from multiple source systems like oracle, Microsoft SQL, MYSQL,
  • Access, DB2 db and source files like delimited files, Excel, Positional and CSV files.
  • Experience using Talend Integration Suite and using components like tXmlInput, tXmlOutput, tXmlAdvanced Output, tFilterRow, tMap, tJoin, tReplicate, parallelize, tJava, tJavaFlex, tAggregateRow, tDie, tWarn, tLogCatcher, tStatsCatcher, tDie, tLogRow, Talend Data Integration
  • Hands on experience on Hadoop technology stack (HDFS, Map - Reduce, Hive, HBase, Pig, Cassandra, Flume, Kafka and Spark)
  • Created Talend custom components for the various use cases and worked on XML components, Data quality, Processing and Log & Error components.
  • Strong Knowledge on Data Warehousing Basics and relational database management Systems and
  • Dimensional modeling (Star schema and Snowflake schema).
  • Implemented Slowly Changing Dimension phenomenon while building Data Warehouses.
  • Implemented Error Logging, Error Recovery, and Performance Enhancement's & created Audit
  • Well versed with Talend Big Data, Hadoop, Hive and used Talend Big data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
  • Created Talend ETL jobs to receive attachment files from pop e-mail using tPop, tFileList and tFileInputMail and then loaded data from attachments into database and archived the files.
  • Expertise in processing data from HDFS, HBase using Hive, Sqoop and Pig components.
  • Used Spark and MapReduce frame works to configure the big data batch jobs.
  • Strong understanding of NoSQL databases like HBase, MongoDB.
  • Expertise in using context variables, Routines and metadata.
  • Experience on working various distributions of Hadoop like CloudEra, HortonWorks and MapR.
  • Good Knowledge in Spark and MapReduce frameworks.
  • Good Knowledge on Big data Hadoop architecture.
  • Expertise in Data modeling techniques like Data Modeling- Dimensional/ Star Schema and Snowflake modeling, Slowly Changing Dimensions (SCD Type 1, Type 2, and Type 3).
  • Excellent working experience in Waterfall, Agile methodologies.
  • Proficient in performance analysis, monitoring and SQL query tuning using EXPLAINPLAN, Collect Statistics, Hints and SQL Trace both in Teradata as well as Oracle.
  • Expertise in deploying from DEV to QA, UAT and PROD with both Deployment group and Import/Exports method.
  • Experience in gathering and writing detailed business requirement and translating them into technical specifications and design.
  • Expertise in understanding and supporting the client with requirements definition, analysis, design, testing, system documentation and user training.
  • Extensive experience in gathering requirements and documenting the same as per the industry best practices.

TECHNICAL SKILLS:

ETL/BI Tools: Talend Data Integration/ Big Data Integration/ Data Quality/ESB 6.1/5.5/5.0, Talend Administrator Console, Informatica Power Center 9.0/8.6

Databases: Oracle 11g/10g, DB2 9.x/8.x, Teradata, Netezza, SQL Server, Hive, HBase, MongoDB.

Programming Languages : XML, PL/SQL, SQL, Hive SQL, T-SQL, Unix/LINUX shell programming and scripting

Development Tools: Teradata SQL Assistant, SQL developer, SQL Server Management Studio, Toad, iSeries navigator, QMF for windows, SQL*Loader, Agnity Workbench for Netezza

Operating Platforms: Windows 8/ 7/XP/2000, Unix, Linux, AIX, Mac OS

Other Tools: AutoSys, CA7 Job Scheduler, Maximo, Tivoli Ticketing System, Altova XML Spy, SQL *Plus, Putty

PROFESSIONAL EXPERIENCE:

Confidential, Oak Brook, IL

Talend BigData Developer

Responsibilities:

  • Closely worked with Data Architects in designing of tables and even involved in modifying technical Specifications.
  • Since this is Migration project we are doing migrating from data stage to Talend Using Big Data Components.
  • We are working sprint wise. Using Talend big data components like Hadoop and S3 Buckets and AWS Services for redshift.
  • Involved in Extraction, Transformation and Loading of data.
  • Work with the offshore team for the day to day work and review the tasks done by them get the status updates in the daily meetings.
  • Data ingestion with different data sources and load into redshift.
  • Developed jobs to send and read data from AWS S3 buckets using components like tS3Connection, tS3BucketExist, tS3Get, tS3Put.
  • Designed and Implemented the ETL process using Talend Enterprise Big Data Edition to load the data from Source to Target Database.
  • Involved in Data Extraction from Flat files and XML files using Talend by using Java as Backend Language.
  • Using Talend to load the data into our warehouse systems
  • Used over 20+ Components in Talend Like (tMap, Tfilelist, Tjava, Tlogrow, ToracleInput, ToracleOutput, tsendEmail etc)
  • Used debugger and breakpoints to view transformations output and debug mappings.
  • Automation Testing of web based applications and services with proficiency in Java, BULK API.
  • Develop ETL mappings for various Sources (.TXT, .CSV,.XML) and also load the data from these sources into relational tables with Talend Enterprise Edition.
  • Worked on Global Context variables, Context variables, and extensively used over 30+components in Talend to create jobs.
  • Extracting transformed data from Hadoop to destination systems, as a one-off job, batch process, or Hadoop streaming process.
  • Worked on Error handling technique's and tuning the ETL flow for better performance.
  • Worked Extensively TAC (Admin Console), where we Schedule Jobs in Job Conductor
  • Extensively Used Talend components tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tOracleInput, tOracleOutput, tfileList, tDelimited etc
  • Migrated the code and release documents from DEV to QA (UAT) and to Production.

Environment: Talend Enterprise Big Data 6.3.1, Redshift, Data stage, Putty, GitHub, Teradata SQL Assistant, AWS Redshift.

Confidential, Wilton, CT

Sr. ETL/ Talend Developer

Responsibilities:

  • Implemented File Transfer Protocol operations using Talend Studio to transfer files in between network folders.
  • Experienced in fixing errors by using debug mode of Talend.
  • Created complex mappings using tHashOutput, tMap, tHashInput, tDenormalize, tUniqueRow. tPivotToColumnsDelimited, tNormalize etc.
  • Schedule the Talend jobs with Talend Admin Console, setting up best practices and migration strategy.
  • Implemented indexing for logs to Elastic Search and analysis on integrating Kibana with Elastic Search
  • Created Context Variables and Groups to run Talend jobs against different environments like Dev, Test and prod.
  • Design and developed end-to-end ETL process from various source systems to Staging area, from staging to Data Marts.
  • Experienced in creating a generic joblet by using tDie, tLogRow, tStatsCatcher to store processing stats into a Database table to the record job history.
  • Experienced in integrating java code inside Talend studio by using tJavaRow, tJavaRow, tJavaFlex and Routines.
  • Developed mappings to load Fact and Dimension tables, SCD Type 1 and SCD Type 2 dimensions and Incremental loading and unit tested the mappings.
  • Experienced in writing SQL Queries and used Joins to access data from Oracle, and MySQL.
  • Broad design, development and testing experience with Talend Integration Suite and knowledge in Performance Tuning of mappings.
  • Developed jobs in Talend Enterprise edition from stage to source, intermediate, conversion and target.
  • To run sub jobs in parallel, used tParalleize component and multi thread execution. These components are useful to increase job performance.
  • Troubleshooting, debugging & altering Talend particular issues, while maintaining the health and performance of the ETL environment.

Environment: Talend Big Data 6.0.1/5.5, Oracle 11i, Mongo DB 3.0.12 XML files, Flat files, HL7 files, JSON, 2.4.1, HDFS, Map Reduce, Sqoop, Hive 0.13, FileZilla 3.18.0, Talend Administrator Console, IMS, Agile Methodology.

Confidential, Milwaukee, WI

Sr.Talend Big Data Developer

Responsibilities:

  • Designed the Data model and Load strategy to get data from different systems and create a Data Lake for TCS Aviation.
  • Data ingestion to and from HDFS and HAWQ/Teradata for storage and analytics with end user reporting.
  • Created Talend jobs to populate the data into dimensions and fact tables.
  • Created Talend jobs to load data into various Oracle tables. Utilized Oracle stored procedures and wrote few
  • Created Hive tables, partitions and implemented incremental imports to perform ad-hoc queries on structured data.
  • Java code to capture global map variables and used them in the job.
  • Created Talend jobs to copy the files from one server to another and utilized Talend FTP components.
  • Designed, developed and improved complex ETL structures to extract transform and load data from multiple data sources into data warehouse and other databases based on business requirements.
  • Used tStatsCatcher, tDie, tLogRow to create a generic joblet to store processing stats and responsible for Pre and Post migration planning for optimizing Data load performance, capacity planning and user support.
  • Troubleshooting, debugging & fixing Talend specific issues, while maintaining the health and performance of the ETL environment.
  • Performed migration of mappings and workflows from Development to Test and to Production Servers.
  • Supported team using Talend as ETL tool to transform and load the data from different databases.
  • Data ingestion in HAWQ using gpload utility for more than 200 tables daily.
  • Created complex jobs to reach the business logic using Orchestration components with gpload to get best f Talend and HAWQ utilities. We are now processing more than 1 TB of data daily.
  • Managing messages on Kafka topics using Talend Jobs.
  • Utilized Big Data components such as tSqoopExport, tSqoopImport, tHDFSInput, tHDFSOutput, tHiveLoad, tHiveInput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult,, tHbaseInput, tHbaseOutput along with executing the jobs in Debug mode and also utilizing the tlogrow component to view the sample output.
  • Developed MapReduce programs in Java for parsing the raw data and populating staging Tables
  • Schedule Big Data Job execution from the Talend Administration Center (TAC) and enriching logs.
  • Developed mappings to load Fact and Dimension tables, SCD Type 1 and SCD Type 2 dimensions and Incremental loading and unit tested the mappings.
  • Used tStatsCatcher, tDie, tLogRow to create a generic joblet to store processing stats into a Database table to record job history.
  • Experienced in using debug mode of talend to debug a job to fix errors.
  • Created complex mappings using tHashOutput, tHashInput, tNormalize, tDenormalize, tMap, tUniqueRow. tPivotToColumnsDelimited, etc.
  • Implemented FTP operations using Talend Studio to transfer files in between network folders as well as to FTP server using components like tFileCopy, TFileAcrchive, tFileDelete, tCreateTemporaryFile, tFTPDelete, tFTPCopy, tFTPRename, tFTPut, tFTPGet etc.

Environment: Talend 5.5/5.0, Talend Bigdata 5.1, Oracle 11g, My Sql, Teradata SQL Assistant, HDFS, Sqoop, Hive, MS SQL Server 2012/2008, PL/SQL, Kafka, Agile Methodology, Cloudera, TOAD, Java ERwin, AIX, Shell Scripts, AutoSys, SVN.

Confidential, Cincinnati, OH

ETL/Talend Developer

Responsibilities:

  • Worked on evaluation and analysis of Hadoop cluster and different big data analytic tools including Pig, H- Base database and Sqoop.
  • Continuous delivery/Continuous Integration (CD/CI) using Jenkins/Cloud Bees and hostage of Quality Control
  • App.
  • Ingesting streaming data into Hadoop using Spark Framework and Scala.
  • Creating Spark SQL queries for faster requests.
  • Used Spark streaming for the real-time processing of the data from HDFS.
  • Imported and exported data into HDFS using Sqoop.
  • Used Qlikview to create visual interface of the real-time data processing.
  • Designed a data warehouse using Hive.
  • Implemented partitioning, dynamic partitioning and bucketing in Hive.
  • Developed Hive queries for the analysts.
  • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
  • Imported, exported data from various databases ORACLE, and MYSQL into HDFS using Talend.
  • Created analysis batch job prototypes using Hadoop.
  • Designed, documented operational problems by following standards and procedures using JIRA.
  • Worked closely with Business Analysts to review the business specifications of the project and also to gather the ETL requirements.
  • Developed jobs, components and Joblets in Talend
  • Designed ETL Jobs/Packages using Talend Integration Suite (TIS)
  • Created complex mappings in Talend using tHash, tDenormalize, tMap, tUniqueRow. tPivotToColumnsDelimited as well as custom component such as tUnpivotRow.
  • Created Talend Mappings to populate the data into dimensions and fact tables.
  • Frequently used Talend Administrative Console (TAC)
  • Implemented new users, projects, tasks within multiple different environments of TAC (Dev, Test, Prod, DR).
  • Developed complex Talend ETL jobs to migrate the data from flat files to database.
  • Implemented custom error handling in Talend jobs and also worked on different methods of logging.
  • Created ETL/Talend jobs both design and code to process data to target databases.
  • Created Talend jobs to load data into various Oracle tables. Utilized Oracle stored procedures and wrote few Java code to capture global map variables and use them in the job.
  • Successfully Loaded Data into different targets from various source systems like Oracle Database, DB2, Flat files, XML files etc. into the Staging table and then to the target database.
  • Troubleshot long running jobs and fixing the issues.
  • Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project from development to testing environment and then to production environment.
  • Performed Unit testing and System testing to validate data loads in the target.

Environment: Talend Open Studio 5.0.1, UNIX, Oracle, Java, TOAD, MS SQL Server 2012/2008, PL/SQL, DB2, Flat files, XML, Talend Integration Suite, Talend Administrative Console, AutoSys

Confidential, Scottsdale, AZ

Senior Informatica Specialist

Responsibilities:

  • Developed high level technical design specification and low-level specifications based on the business requirements.
  • Extensively used Informatica client tools (Source Analyzer, Warehouse Designer, Mapping Designer and Workflow Manager).
  • Used Informatica Designer for developing mappings, using transformations, which includes aggregate, Update, lookup, Expression, Filter, Sequence Generator, Router, and Joiner etc.
  • Created reusable transformations and mapplets and used them in mappings to reduce redundancy in coding.
  • Extensively used Informatica Power Exchange Change Data Capture (CDC) for creation of Data Maps using Mainframe Tables.
  • Coded number of batch and online programs using COBOL-DB2-JCL.
  • Designed Complex mappings, Used Lookup (connected and unconnected), Update strategy and filter transformations for loading historical data.
  • Extensively used SQL commands in jobs prior to extracting the data in the ETL tool.
  • Implemented different tasks in workflows which included Sessions, Command Task, Decision Task, Timer, Assignment, Event-Wait, Event-Raise Control, E-Mail etc.
  • Used Debugger to test the data flow and fix the mappings.
  • Involved in Performance tuning of the mappings to improve the performance.
  • Performed Unit Testing and prepared unit testing documentation. Developed the Test Cases and Test Procedures.
  • Extensive use of IDQ for data profiling and quality.
  • Built a Unix Script which checks the Mapping, Session and Workflow names by identifying the Power Center folders, builds an XML and then it zips (tar) all documents based on the names given and deploy them across the environments.
  • Scheduled Jobs and box jobs in AutoSys and analyzed the Run status of both jobs and box jobs in DB2 Environment.

Environment: Informatica Power Center 9.0/8.6, Informatica Power Exchange CDC, DB2 Mainframe, AutoSys, Toad, Windows XP, UNIX.

Confidential, Cincinnati, OH

Informatica Developer

Responsibilities:

  • Worked on Informatica - Repository Manager, Designer, Workflow Manager & Workflow Monitor.
  • Integrated data into CDW by sourcing it from different sources like SQL, Flat Files and Mainframes (DB2) using Power Exchange.
  • Extensively worked on integrating data from Mainframes to Informatica Power Exchange.
  • Extensively worked on Informatica tools such as Source Analyzer, Data Warehouse Designer, Transformation Designer, Mapplet Designer and Mapping Designer to designed, developed and tested complex mappings and Mapplets to load data from external flat files and RDBMS.
  • Used output xml files, to remove empty delta files and to FTP the output xml files to different server.
  • Worked with the Business Analyst team during the functional design and technical design phases. Designed the mappings between sources (external files and databases) to operational staging targets.
  • Extensively used various transformations like Source Qualifier, Joiner, Aggregators, Connected and Unconnected lookups, Filters, Router, Expression, Rank Union, Normalizer, XML Transformations and Update Strategy & Sequence Generator.
  • Used XML transformation to load the data XML file.
  • Worked on Informatica Schedulers to schedule the workflows.
  • Extensively worked with Target XSD's in order to generate the output xml files.
  • Created mappings to read parameterized data from tables to create parameter files.
  • Good Experience in Co-Coordinating with Offshore.

Environment: Informatica Power Center 8.6.1, Power Exchange 8.6.1, Windows, IBM DB2 8.x, Mainframes, SQL Server 2008, ERwin.

Confidential

SQL Developer

Responsibilities:

  • Analyzed reports and fixed bugs in stored procedures using SSRS.
  • Used complex expressions to group data, filter and parameterize reports.
  • Created linked reports and managed snapshots using SSRS.
  • Performed various calculations using complex expressions in the reports and created report models.
  • Generated Complex SSRS reports like reports using Cascading parameters, Snapshot reports Drill-down Reports, Drill-Through Reports, Parameterized Reports and Report Models and ad hoc reports using SSRS based on Business Requirement Document.
  • Conducted performance tuning of complex SQL queries and stored procedures by using SQL Profiler and index tuning wizard.
  • Provided Production support to analyze and fix the problems and errors on daily basis by modifying SSIS Packages and Stored Procedure if necessary.
  • Designed and developed Tables, Stored procedures, Triggers and SQL scripts using TSQL, Perl and Shell scripting for enhancements and maintenance of various database modules

Environment: MS SQL Server 2005/2008, SSRS, SSIS, SSAS, T-SQL, Erwin, SQL Explorer.

We'd love your feedback!