Ab Initio Developer Resume
Foster City, CA
SUMMARY
- 8 years of work experience in teh area of Data warehousing and Business Intelligence, ETL development and testing, migration using Ab Initio, UNIX shell script, PL/SQL, Teradata, Core Java, Hadoop, Map reduce, Hive, PIG, SQOOP, Python
- Venturing into Machine learning
- Through grasp and maturity to design dimensions and facts. Robustly design slowly changing dimensions, ETL batch/online applications which integrate teh star, snowflake or galactic schema
- Good understanding of various pattern matching algorithms, name and address matching/cleansing logic using Trillium
- Extensively worked on Banking/Payment Domain Confidential, Confidential Inc. and Retail Domain Williams - Sonoma, Walgreens, Landmark Group etc.
- Strong Migration project handling capabilities.
- Experience working wif various Heterogeneous Source/Target Systemslike Oracle, Teradata, SQL Server, flat files and MF (Ab Initio Parallelism)
- Worked extensively wif Ab Initio GDE in configuring, developing graphs and validating teh data against teh business requirements using Ab Initio GDE 3.02/1.15/1.14, Microsoft DTS, ETL Functions and related OLAP tools
- Worked wif using Ab initio EME for Version control, Impact analysis, Dependency analysis for Common projects, higher environment migrations
- Experience in using various Ab Initio components such as Join, Reformat, Scan, Rollup, Normalize, De-normalize, Partitioning and De-partitioning component
- Experience in debugging and implementing various techniques in tuning teh graphs to improve teh performance
- Experience in Streaming teh Data to HDFS using Flume
- Experience in working wif Hadoop clusters using Cloudera (CDH3) distributions
- Hands on experience in using Map reduce programming model for Batch processing of data stored in HDFS
- Expertise in writing ETL Jobs for analyzing data using Pig
- Experience in NoSQL Column-Oriented Databases like HBase and its Integration wif Hadoop cluster
- Strong in analysis of complex technical situations, critical thinking ability
- Creative technical problem-solver. Ability to Multitask and meet development milestones ensuring high quality & on-time delivery
- Excellent Communication skills in interacting wif various people of different levels on all projects
- Clear and thorough understanding of business processes and workflow
- Effective Team Player wif excellent quantitative, analytical, programming, written and verbal communication skills
- Leadership roles in designing and development of various projects and support tasks
- Strong experience on conducting Review Meetings and Process Documentations
- Good communication skills, analytical skills and documentation skills
TECHNICAL SKILLS
Domain Knowledge: Retail data warehousing, Anti-money Laundering (Banking DW), Electronic payments ( Confidential Inc. )
Tools: Co>op Ab Initio 3.1.2.5, GDE 3.1.7.3, EME, Trillium 14
Big Data Tools: SQOOP, ZOO Keeper, Apache OOZIE, HDFS and its components
Operating Systems: AIX, Solaris, Linux, Windows, DOS
Languages: Pl/SQL, Unix shell script, HIVE QL, PIG Latin, Map Reduce (core Java), Python
Database: Oracle 10g, Teradata 12, IBMDB2
Frameworks: Hadoop HDFS (Apache, Cloudera)
Scheduler: Control - M 7.0.00, Ab Initio Operational Console
PROFESSIONAL EXPERIENCE
Confidential, Foster City, CA
Ab Initio Developer
Responsibilities:
- Involved in teh Functional, detailed design and Quality Analysis of teh project.
- Design and review of data driven data model Data volume and capacity analysis of per day 60 millions of transaction data and 5.5 billion of raw transaction data to handle in a year
- Design and develop Ab initio graphs wif several performance tuning to make teh system scalable to handle this big volume of data
- Go live preparation wif back loading of 13 months of data
- Converted Raw Real time XMLS to readable format to use it as Source for Transformation and loading and process delta feeds of customer data, which comes in daily
- Teh Graphs developed, were tested wif Large Scale redacted data at various levels for better performance tuning
- Worked on Quality analysis using Data profiling by maintaining various Transform components, Dataset Components to pull, cleanse and load data into target Data warehouse
- Extensive usage of Multifile system where data is partitioned into four and eight partitions for parallel processing.
- Used GDE in Ab-Initio to deploy components onto a canvas, configure and connect them into executable flowcharts
- Performed transformations of source data wif Transform Components like Join, Dedup Sorted, Denormalize, Normalize, Reformat, Filter-by-Expression, and Rollup.
- Successfully Implemented Stored Procedures for teh DB in Ab initio graphs using Join wif DB component
- Wide usage of Lookup Files while getting data from multiple sources and size of teh data is limited.
- Wrote several Shell scripts for Project maintenance (to remove old/unused files and move raw logs to teh archives).
- Used several air utility commands in teh Project Development and Data profiling.
- Used sandbox parameters to check in and checkout of graphs from repository wif Hands on experience in working wif EME and Project management.
- Queried DB regularly using SQL queries to keep a constant check on teh data loaded in teh respective tables
- Developed dynamic graphs to load data from data sources into tables and to parse records.
- Accessed teh Role of Production support Lead from ETL team and created numerous Instant One time Ab initio graphs to solve any error’s in teh Live environment on spot
- Analyzed teh data using Ab Initio Data Profiler giving Statics of teh data like null values, maximum, minimum and average values
- Used Ab Initio for Quality Analysis, Error Handling by attaching error and rejecting files on each transformation and making provision for capturing and analyzing teh message and data separately.
- Java and Hadoop changes to include new fields and its related business requirements.
- Implemented a logic to filter bad records on Java and Hadoop areas of Golden Record identification
- Supported Map Reduce Programs those are running on teh cluster Jobsmanagement using Fair scheduler
- Involved in loading data from UNIX file system to HDFS.
- Installed and configured Hive and also written Hive UDFs.
- Loaded large sets of structured, semi structured data coming from UNIX, NoSQL and variety of portfolios into HBASE by creating teh HBASE table
- Created reports for teh BI team usingSqoop to export data into HDFS and Hive.
- Developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
- Assisted wif data capacity planning and node forecasting
- Involved in creating Hive tables, loading wif data and writing hive queries which will run internally in map way
- Automated all teh jobs, for pulling data from FTP server to load data into Hive tables, using Oozie workflows
- Working on ad-hoc changes(Hadoop and Java code changes, unit testing) based on teh change request
- Experienced Incremental and full back up methodologies using Sqoop.
- Developed Moduls in Python for Web crawling
Environment: Co>op Ab>Initio 3.1.6.2 GDE 3.1.7.3, EME, AIX, UNIX shell script, IBM DB2, Trillium, Hadoop, HIVE, PIG, OOZIE, Python 2.7.10
Confidential, San Francisco, CA
Ab Initio Developer
Responsibilities:
- Involved in analyzing business needs and document functional and technical specifications based upon user requirements wifextensive interactions wif business users.
- Involved as designer and developer for Enterprise data warehouse.
- Lead teh data provisioning team for all teh tracks (flat files, DB & mainframe data)
- Utilize variousAbInitiocomponents likeLookup, Join,RollupandReformatto process data
- Developed various Ab-Initio graphs to validate using data profiler, comparing teh current data wif previous month data, applying teh standards.
- Engaging wif teh onsite coordinators to identify, consult and clarify requirements
- Detailed analysis of existing RDW Ab>initio graphs and shell scripts
- Acquired detail level understanding of dimension and fact data loading processes by analyzing teh existing graphs
- Designing of teh data migration process from prod to dev env.
- Preparation of technical design for teh required change
- Coding teh generic graphs using Ab>initio GDE which interact wif Teradata for extracting and loading of data.
- Coding teh Ab initio plan
- Has used Teradata utilities,BTEQ,FastLoad,MultiLoad,TPump, FastExport, set, multiset tables for scripting, extraction and data loading
- Task allocation and coordination between team members are performed
- Wide usage of lookup files while getting data from multiple sources and size of teh data islimited.
- Involved in project promotion from development to UAT and UAT to promotion.
- Implemented Data Parallelism unitizing MFS in teh graphs, which deals wif data, divided into segments and operates on each segment simultaneously through teh Ab Initio partition components to segment data.
- Used phases and checkpoints to avoid deadlocks in AB Initio graphs to facilitate recovery after failures
- Using modification of teh Ab intio EME to house teh required redaction Meta data.
- Used Ab Initio Data Profiler to analyze & validate source data and Data types. And also to determine join datasets by analyzing cross functional relations.
- TEMPHas used Control M for scheduling jobs in Development Environment.
- Used different EME air commands in project promotion like air tag create, air save, air load, air project export etc
- Identify data issue and perform root cause analysis and take corrective actions
Environment: Ab Initio 2.14, GDE 2.14, EME, Oracle 11g, Control M, UNIX, windows, Teradata SQL Assistant, Teradata, Teradata warehouse miner
Confidential
Ab Initio Developer
Responsibilities:
- Understanding of Requirements docs & preparation of queries.
- Create high level and detail level design documents
- Engaging wif teh Business teams to identify, consult and clarify requirements
- Coding teh graph using Ab>initio GDE
- Coding graphs to implement Acquire and publish, cluster, DMU process of ETL design.
- Coding A>me graphs to implement change data capture (CDC), account reusability (using surrogate key) concepts.
- Walkthrough of teh developed Ab>initio graphs to customers.
- Preparation of Unit test cases and testing of teh developed code
- Configuration control of developed code through EME and assist change team during deployment of code in various environments me.e. Testing, UAT and Pre-prod.
- Create Unix wrapper script to call graphs
- Worked in performance tuning of teh Ab>initio graphs
- Prepared impact analysis document and given presentation of teh same to client
- Also worked as Test Lead for system integration testing
- Preparation test data using Ab>initio GDE
- Preparation of graphs through Ab>initio GDE to implement ETL requirements of teh project and to facilitate independent testing
- Preparation of several Unix shell script to achieve reusability and automation for testing
- Preparation of Test strategy and Test Plan.
- Preparation and review of Test Scenarios, Test cases
- Execution of test scripts and defect management.
- Estimation and testing done for CR.
- Organizing and attending defect calls, client meetings and presentations to Business.
- Task allocation and coordination between team members are performed.
- Identification of resource for project and resource allocation forecasting is performed
- Estimation of timeline in accordance to effort for project is performed and communicated to customer
- Performed Quality gate process
- Daily and weekly Status reporting to program management.
- Providing KT, Sharing knowledge and providing guidance to teh team members
Environment: Ab Initio (Co> Operating system 2.13, GDE 1.14/1.13), Solaris UNIX, Oracle 8i/9i, EME, SQL * Loader and Windows NT
Confidential
AB Initio Developer
Responsibilities:
- Involved in all phases of ETL cycle and prepared detailed level design that depicts teh transformation logic used in Ab Initio.
- Responsible for collecting and documenting teh business requirements.
- Responsible for designing teh source to target mapping diagrams
- Developed generic graph to extract data from various heterogeneous source such as Oracle databases and from flat files.
- Responsible for creating Ab Initio graphs for landing, teh validated source data received from various divisions, in multifiles and creating lookups for cross reference.
- Used Ab Initio EME for repository of Graph object and performed check in/ check out.
- Extensively used database, dataset, partition / departition, transform, sort and partition components for extracting, transforming and loading.
- Developed number of Unix Korn Shell wrapper scripts to control various Ab Initio processes and complex functionalities such automated ftp, remote shell execution, and remote copy etc
- Wrote UNIX shell scripts to automate teh some of teh data extraction, and data loading process.
- Involved in Creation of Sandboxes and managing Enterprise Metadata using EME.
- Involved in creation of teh database and developed teh backend procedures for teh system
- Developed a number of scripts and procedures to address teh changes to system based on change requests of teh user needs
- Involved in creating interfaces from and to teh manufacturing, shipping and fulfillment systems
- Involved in doing migration of legacy system data to new systems. Initially teh migration was done to teh beta test systems to establish data validation and stability of migrations and subsequently ported to teh production
Environment: Ab Initio (Co >Operating system 2.12 GDE 1.12), Oracle 8i, SQL plus, UNIX and Window NT.