Talend Developer Resume
Detroit, MichigaN
SUMMARY
- Experience in development, design, implementation and support with major focus on Data Warehousing, and Database Applications.
- Good experience on Talend Open Studio and Enterprise Version
- Strong Understanding of Data Warehousing concepts like Star Schema, Snowflake Schema, and Dimension Data Modeling.
- Worked with multiple operational information sources like Flat Files/Excel, Oracle, SQL Server, Redshift and MySQL
- Used output xml files, to remove empty delta files and to FTP the output xml files to different server.
- Expertise in extracting information from various sources, data cleaning and approval in view of business prerequisites.
- Implemented Parallelism and ELT utilizing Talend in Vast Database Frameworks Condition.
- Worked with an assortment of planning apparatuses including Autosys and Job Conductor.
- Experienced in scheduling Talend jobs using Talend Administration Console (TAC)
- Working experience with Agile Software development method.
- Worked extensively with slowly changing dimensions.
- Involved in building the ETL design for job and Source to Target mapping to load data into Data Warehouse.
- Experience in Application Development and Maintenance using RDBMS - Oracle (PL-SQL), Business Intelligence, ETL, UNIX/LINUS.
- Skilled at UNIX/LINUX for writing shell scripts and Batch process.
- Created Audit framework/Error Logging in Talend.
- Created mapping documents to outline data flow and metadata information from sources to targets.
- Extracted the data from multiple databases into staging area and populated onto Data warehouse.
- Used various transformations like tMap, tAggregator, tFilterRow, tJoin, tPreJob, tPostJob, tSystem, tSortRow, tRunjob, Stored Procedure, and Union to develop robust mappings in the Jobs.
- Performance tuning - Using the tMap cache properties, Multi-threading and tParallelize components for better performance in case of huge source data and Tuning the SQL source queries to restrict unwanted data in ETL process.
- Implemented custom exceptional handling in Talend jobs and worked on different methods of logging as discussed above.
- Developed mappings parameters to load into staging tables and then to Dimensions and Facts and Used Type 2 SCD mappings to update slowly Changing Dimension Tables. Prepared document for Unit test cases from each Flow of Data and designed ETL Job for Testing Involved in Performance tuning at source, target, mappings, sessions, and system levels.
TECHNICAL SKILLS
BI Tools: SSIS, Talend, SSDT (BIDS), Visual Studio, Teradata, Performance Point Server, ProClarity,SQL*Plus,SQL*Loader, Jasper Soft Studio., AWS S3
Databases: SQLServer 2008/2012/2014 , MySQL, Oracle 10g, 11g, Amazon Redshift, Snowflake
Application package: MS Office Suite, Visio.
Operating Systems: Windows Server 2003, 2007, 2010, 2013, Windows 8/NT/XP/Professional
Languages: T-SQL, PL/SQL, C/C++, HTML, XML, C#.Net, ActiveX, XML, MS Access, HTML, DHTML, Shell Scripting
Tools: SQLServer Enterprise Manager,SQLServer Management Studio, Team Foundation Server (TFS)
Protocol: FTP, TCP/IP, Named Pipes
TECHNICAL SKILLS
Confidential, Detroit, Michigan
Talend Developer
Environment: Talend Enterprise Big Data cloud Edition 7.0.2, Talend Administrator Center, Talend Remote engine, MS SQL Server 2012/2008, HDFS, Snowflake, Spark, Amazon Redshift, AWS, UNIX, Putty, WinSCP, ServiceNow.
Responsibilities:
- Converted Existing DataStage jobs to Talend jobs for fact loads and dimension loads.
- Designed and Implemented ETL for data load from SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2 to capture the changes to snowflake cloud database.
- Performance tuning - Using the tmap cache properties, Multi-threading and tParallelize components for better performance in case of huge source data. Tuning the SQL source queries to restrict unwanted data in ETL process.
- Created the Talend jobs to load the data into Snowflake and SQL server tables. Creating new batches that utilized parallelism better (Multi Instances concept in DataStage).
- Experience in using Talend features such as context variables, triggers, connectors for Database and flat files.
- Hands on Experience on many components which are there in the palette to design Jobs & used Context Variables to Parameterize Talend Jobs.
- Extracted data from databases like Oracle, SQL server and DB2 using Talend to load it into a single repository for data analysis.
- Debug Talend jobs issues in production environment and Updated the job Accordingly for redeployed it.
- Experience in Debugging, Error Handling and Performance Tuning of sources, targets, Jobs etc.
- Experience in using the Talend Troubleshooting and Datastage to understand the errors in Jobs and used the tMap/expression editor to evaluate complex expressions and look at the transformed data to solve mapping issues
- Created SQL Store processors for transforming the data and optimized them.
- Created complex mappings in Talend using components like: tMap, tJoin, tReplicate, tParallelize, tAggregateRow, tDie, tUnique, tFlowToIterate, tSort, tFilterRow, tcontextload.
Confidential, Lawrenceville, NJ
Talend Developer
Environment: Talend Enterprise Big Data Edition 7.0.1, Talend Administrator Center, MS SQL Server 2012/2008, HDFS, Spark, SnowFlake, Amazon Redshift, AWS, UNIX, Putty, WinSCP, ServiceNow.
Responsibilities:
- Moving files to job server, required transformation to be done and load them in REDSHIFT or any delimited files.
- Data ingestion with different data sources and load into redshift.
- Designed and Implemented the ETL process using Talend Enterprise Big Data Edition to load the data from Source to Target Database.
- Excellent experience working on Talend ETL and used features such as Context variables, Database components like tMSSQLInput, tOracleOutput, tmap, tFileCopy, tFileExist, tIterateToFlow, tFlowToIterate, tfiledelimited, tRunjob, tJava etc.
- Involved in Data Extraction from Flat files, Json and XML files using Talend by using Java as Backend Language.
- Experienced in transferring the data from relational data base to Cloud such as Amazon S3, Redshift by using Talend Big data Spark Jobs
- Created External Schemas and External Tables in redshift and loaded nested data into it.
- Worked on AWS Athena and Glue consoles and created Athena tables on top of it according to the requirements.
- Created context variables and groups to run Talend jobs against different environments and created different types of triggers for a Talend job to run automatically on server.
- Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project from development to testing environment and then to production environment.
- Used SQL queries and other data analysis methods, as well as Talend Enterprise Data Quality Platform for profiling and comparison of data, which will be used to make decisions regarding how to measure business rules and quality of the data.
- Collaborating with Data Integration Team to perform data and application integration with a goal of moving more data more effectively, efficiently and with high performance to assist in business-critical projects coming up with huge data extraction.
- Used debugger and breakpoints to view transformations output and debug mappings.
Confidential, Durham, NC
ETL/ Talend Developer
Environment: Talend Data integration, Snowflake, JIRA, TAC, AWS S3, MS Excel and MSSQL Server.
Responsibilities:
- Extensively worked on migrating all the client data from AWS S3 bucket to Snowflake DB.
- Modifying existing Talend mappings to load to Snowflake DB.
- Recreating existing AWS objects in Snowflake.
- Involved in design, development of Talend mappings and Maintaining proper code throughout the whole ETL process.
- Tuned Talend mappings for better performance on request.
- Designing the ETL mappings and workflows to fetch data from multiple sources (.xml, .csv, .txt) and also loading the data from these sources into relational tables or destination sources using Talend Open Studio.
- Extensively used database connections, file component, tmap, tAggregate etc.
- Exception Handling in TalendtDie, tlogcatcher
- Working knowledge on the reusable components like contexts, Global variables in Talend.
- Prepare metadata in Talend Integration studio repository.
- Used Orchestration components liketpre and tpostjob, tfilelistetc
- Processing components such as tmap, tjoin, tReplicate, tUniteetc. Have created reusable joblets.
- Complete validation of data from SOURCE to TARGET.
- Performed data manipulations using various Talend components like tMap, tJavarow, tjava, tSqlRow, tMSSQLInput, tJDBCInput, tJDBCOutput and many more.
- Used tStatsCatcher,tDie,tLogrow to create a generic joblet to store processing stats into a database table to record job history.
- Thorough understanding of the SDLC with proficiency in mapping business requirements, technical documentation, application design, development.
Confidential, Boston, MA
ETL/ Talend Developer
Responsibilities:
- Experienced in fixing errors by using debug mode of Talend.
- Created complex mappings using tHashOutput, tMap, tHashInput, tDenormalize, tUniqueRow. tPivot To Columns Delimited, tNormalize etc.
- Schedule the Talend jobs with Talend Admin Console, setting up best practices and migration strategy.
- Used components like tJoin, tMap, tFilterRow, tAggregateRow, tSortRow, Target Connections and Source Connections.
- Mapping source files and generating Target files in multiple formats like XML, Excel, CSV etc.
- Transform the data and reports retrieved from various sources and generating derived fields.
- Reviewed the design and requirements documents with architects and business analysts to finalize the design.
- Created WSDL data services using Talend ESB.
- Created Rest Services using tRESTRequest and tRESTResponse components.
- Used tESBConsumer component to call a method from invoked Web Service.
- Implemented few java functionalities using tJava and tJavaFlex components.
- Developed shell scripts, PL/SQL procedures for creating/dropping of table and indexes of performance.
- Attending the technical review meetings.
- Implemented Star Schema for De-normalizing data for faster data retrieval for Online Systems.
- Involved in unit testing and system testing and preparing Unit Test Plan (UTP) and System Test Plan (STP) documents.
- Responsible for monitoring all the jobs that are scheduled, running completed and failed. Involved in debugging the jobs that failed using debugger to validate the jobs and gain troubleshooting information about data and error conditions.
- Performed metadata validation, reconciliation and appropriate error handling in ETL processes.
- Developed various reusable jobs and used as sub-jobs in other jobs.
Environment: Talend 5.1, Oracle 11g, DB2, Sybase, MS Excel, MS Access, SQL, UNIX.
Confidential
ETL Developer
Responsibilities:
- Reverse engineered existing data models to fit in new changing requirements.
- Coded extensively in writing and debugging complex stored procedures, views, triggers and user - defined functions.
- Optimized the queries by creating various clustered, non-clustered indexes and indexed views Performance tuned stored procedures and issues related to parameter sniffing with SQL queries utilizing SQL Profiler.
- Created SQL server jobs and scheduled them to load data periodically utilizing SQL server agent.
- Identified the dimensions, fact tables and designed the data mart utilizing star schema.
- Created multitude of SSIS packages imported and exported data from various data sources such as SQL server databases, flat files, MS access, MS excel and other OLE DB providers.
- Designed packages utilizing fuzzy lookup, derived columns, condition split, aggregate, execute SQL task, data flow task, and execute package task.
- Incorporated Event Handlers, Loggings, Breakpoints and Data Viewers for effective debugging of packages including watch windows.
- Deployed SSIS packages into various environments (development, test) utilizing package/project deployment utility.
- Generated multi parameterized reports in SSRS 2012 allowing users the ability to make selections before executing reports; thus, making them user friendly.
- Provided daily support to the user team as well as technical team for system wide tasks including monitoring, alerting and problem resolution. Gathered business requirements from business users to perform technical design, development and testing of prototype.
- Created gateways to communicate between the data on premise and AWS cloud platform and migrated data onto AWS for universal access. Developed Stored Procedures and used them in Stored Procedure transformation for data processing and have used data migration tools.