Lead Datastage Developer Resume
Atlanta, GA
SUMMARY
- Lead with 12 years of professional experience in designing and developing ETL / Data warehouse solutions.
- Experience in all Phases of System/Software Development Life Cycle (Process Consulting, Architecture, Design, Development, Testing, Deployment and Support).
- Extensive experience in design and architecture of ETL interfaces and data marts.
- Familiar with aspects of technology projects like Business Requirements, Technical Architecture, Design Specification, Development and Deployment.
- Over 2+ years of experience on Hadoop Ecosystems (Hive, SQOOP and Pig)
- Handled the importing of data from various data sources like TERADATA, ORACLE, DB2, SQL SERVER and GREENPLUM using SQOOP and TPT, performed transformation - using Hive and loaded the data into HDFS.
- Experience standardization the address, name, and Area using standardization stage through IBM InfoSphere QualityStage.
- Experience in Word Investigations, Character Discrete using Investigate stage
- Experience in Investigate stage, Standardize stage, Match stage
- Information Analyzer administrative tasks such as managing logs, schedules, active sessions and security roles etc
- Experience in Data Analysis and modeling
- Experienced in creating Data Lineage report, Impact analysis report through IBM Data Governance Catalog & IBM Meta Data Asset Manager
- Experience in Data modelling, Design and architecture review.
- Import/export projects along with rules and bindings successfully from one environment to another using Information Analyzer
- Experience in estimation, planning, forecasting
- Excellent knowledge in Business Intelligence, Analytics and Optimization.
- Proficiency in Data Warehouse Architecture and Designing Star Schema, Snow flake Schema, FACT and Dimensional Tables, Physical and Logical Data Modeling using Erwin and Designer 2000
- Familiarity with Business Intelligence Tools like Business Objects.
- Extensive experience in loading high volume data, and performance tuning.
- Hands on experience in writing, testing and implementation of the triggers, Procedures, functions at Database level using PL/SQL.
TECHNICAL SKILLS
ETL TOOLS: IBMInformationAnalyzer/IBMWebsphereDataStage11.x//7.5x2/7.x,Qualitytage, SSIS/SSRS/SSAS
HADOOP TOOLS: HIVE,PIG,SQOOP
REPORTING TOOLS: Business Objects, Tableau
LANGUAGES: PLSQL,C,C++
PROFESSIONAL EXPERIENCE
Confidential
Lead Datastage Developer
Responsibilities:
- Played a Major role in gathering business requirements from Users for the migration of the legacy system tables to Data warehouse & developing the DataStage jobs by implementing the best practice in DataStage Design and Development.
- Involved and Responsible in entire project lifecycle - requirements gathering through implementation including: preparing the functional, scope, mapping documents and high level and low level design specification documents, Dimensional Data Modeling (Logical and Physical).
- Used different types of Stages like Join, Lookup, Sequential File, Dataset, Transformer, Sort, Aggregator, Merge, Horizontal Pivot, Vertical Pivot, Funnel, Filter, Copy, Modify, Remove Duplicate, Change Capture, etc. for designing of the Jobs.
- Developing Scripts and Batch Job to schedule various Hadoop Program.
- Written Hive queries for data analysis to meet the business requirements.
- Creating Hive tables and working on them using Hive QL.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Experienced in defining job flows.
- Involved in creating Hive tables, loading data and writing hive queries.
- Developed a custom File system plug in for Hadoop so that it can access files on Data Platform
- Extensively worked on DataStage jobs for splitting bulk data into subsets dynamically distribute to all available processors to achieve best job performance.
- Responsible for DataStage Code Deployment in 11.3/11.5 and Export/Import of codes from server to server
Environment: IBM InfoSphereDataSatge 11.5/11.3, SQL Server, Oracle, Hadoop, AIX UNIX Shell Scripting, MS VISO, Putty
Confidential, Atlanta, GA
Lead Datastage Developer
Responsibilities:
- Used DataStage as an ETL tool to extract data from sources systems, loaded the data into the IBM DB2 database.
- Created jobs to read and write data into complex flat files, EBCDIC, COBOL files, sequential files.
- Worked closely with the project team in understanding end user requirements and provided technical feasibility.
- Analyzed Data Models with Business and prepare the ETL Technical Design Document.
- Used Fast load, Multi load scripts to load data into Oracle from flat files and COBOL copybooks.
- Expertise in designing and implementingDataStageArchitecture in data warehousing and Business Intelligence projects.
- Worked with Functional team and Data Modelers/Architects to identify and understand the data from different source systems.
- Involved in the Analysis of the functional side of the project by interacting with functional experts to design and write technical specifications.
- Prepared ETL Technical Design documents for the different integration points between the Enterprise Data Warehouse (EDW) and legacy applications
- Experience in writing rule sets using QualityStage for data profiling.
- Identify duplicate Records using Matching Stage through IBM InfoSphere QualityStage.
- Experience Select the best record of a match group using Survivorship stage through IBM InfoSphere QualityStage.
- Experience standardization the address, name, and Area using standardization stage through IBM InfoSphere QualityStage.
- Experience in Word Investigations, Character Discrete using Investigate stage
- Experience in writing rule sets using QualityStage.
- Experience in writingSurvivorship rules using QualityStage
Environment: Datastage 11.5, QualityStage, Teradata, DB2, UNIX Shell Scripting, Zeke scheduler, HPQC, Visio.
Confidential, Dearborn, MI
Lead Datastage Developer / Datastage Architect
Responsibilities:
- Performed Installation and migration of IBM Websphere Datastage Enterprise Edition/PX on UNIX platforms and testing of latest version of Datastage.
- Reduced manual workload 75% for 12-person operations team by automating core ETL process
- Developed and maintained ETL (Extract, Transformation and Loading) mappings to extract the data from multiple source systems like Oracle, SQL server and Flat files and loaded into Teradata.
- Experienced in managing and reviewing Hadoop log files.
- Involved in loading data from UNIX file system to HDFS.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Responsible to manage data coming from different sources.
- Supported Map Reduce Programs those are running on the cluster.
- Involved in creating new table structures and modifying existing tables and fit into the existing Data Model.
- Resolved Datastage configuration and performance issues along with development team
- Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Responsible to manage data coming from different sources.
- Supported Map Reduce Programs those are running on the cluster.
- Involved in creating new table structures and modifying existing tables and fit into the existing Data Model.
- Resolved Datastage configuration and performance issues along with development team
- Performed administrative functions such as monitoring and troubleshooting on daily basis
- Created Teradata SQL Queries, creating Tables, and Views by following Teradata best practices. Documented the purpose of the jobs so as to facilitate the personal to understand the process and in corporate the changes as and when necessary
- Also involved in assigning roles to the users created in SAP using the different user defined BAPI.
- Used ABAP and BAPI stages to extract and load the data into SAP system
- Involved in technical design phase, gathering the business requirements from business users, working with business analysts and helping them translate the requirements into technical specifications.
Environment: Datastage 11.3Hartonworks Hadoop, Hive, Pig, Oozie, Sqoop, Datastage, Teradata, Oracle 11g, Hbase, Toad, SQL Server. UNIX Shell Scripting, Infosphere MDM.
Confidential, St.louis, MO
Sr. DataStage Analyst
Responsibilities:
- Involved in Data Warehouse design, Data cleansing and maintenance.
- Worked in the development and implementation and production support for all the teams those were involved in the Gateway application.
- Designed and developed jobs using Data Stage Designer to extract data from different resources, cleansing, applied various Business rules and logic at transformation stage and loading data into data warehouse.
- Used Datastage Manager for importing metadata from repository, created new job categories and new data elements.
- Extensively used Parallel Extender and Sever edition to load data into the data warehouse with different techniques like pipeline and partition in MPP environment.
- Used Datastage Designer to develop server and parallel jobs for extracting, transforming, integrating and loading data to targets. Improved the performance of the jobs by running the jobs in parallel depending on the size of the files using shared containers.
- Worked on Data profiling to do the column analysis, Table analysis, and primary Key analysis.
- Worked on the design of jobs that use the data coming from in the xml format and loaded into the queue with the help of MQ Series.
- Experience in full life cycle of Data Warehouse design and implementation.
- Worked with XML stages for handling XML data and used Folder stage to read the XML data.
- Obtaining detailed Performance Statistics on jobs for performance tuning purposes
- Create and used Datastage Shared Containers, Local Containers for DS jobs and retrieving Error log information.
- Extensively worked on the scheduling and maintenance of Data Stage jobs using Autosys and job sequencers.
Environment: Ascential DataStage SE, ProfileStage 7.0.1/7.5, Oracle9i,DB2 UDB, Teradata, MQ Series, TOAD, Crystal Reports, SQL,PL/SQL, SQl Loader, Unix Shell Scripting, Autosys, Rational Clear Case, Rational Clear Quest, Windows NT.