We provide IT Staff Augmentation Services!

Etl Data Integration Architect / Hadoop Developer Resume

0/5 (Submit Your Rating)

Warren, NJ

SUMMARY:

  • ·18 years of IT experience working in all phases of the system development life cycle including business requirements gathering, business process modeling, interacting with business users, functional specifications documentation, software architecture, data analysis, JAVA, data warehousing, data modeling, ETL (Informatica) and deployment.
  • ·Experience in Master Data Management concepts, Methodologies and ability to apply this knowledge in building MDM solutions.
  • ·Expert knowledge in data migration / integration technologies (Informatica) successfully completing complex multi terabyte EDW systems and data marts (Star Schema, Snowflake Schema) from OLTP and EDB systems.
  • ·Experience in Relational and Dimensional Data modeling using ERWIN.
  • ·Sound understanding of Financial Industry including Prime Brokerage Business.
  • ·17 years of experience working with Relational Databases (Oracle, Informix, Informix XPS, Sybase, DB2, Teradata).

TECHNICAL SKILLS:

Technology and Tools: Data Warehousing Informatica (9.1 / 8.6 / 5.1 / 4.7 ), Informatica MDM Java, C, C++ Teradata Oracle DB2 Modeling (Erwin) Unix / Linux Shell Scripting SQL Python Business Objects Trillium Greenplum 4.2 Hadoop / Pig / Sqoop / Python

PROFESSIONAL EXPERIENCE:

Confidential, Warren, NJ

ETL Data integration Architect / Hadoop Developer

Responsibilities:

  • Designed and developed ETL components using Talend, Informatica and PL/SQL.
  • Guided and assisted data modeling efforts to accommodate the data from various sources in citiKYC database.
  • Customer Metadata Using Informatica MDM: Produced customer master data using Informatica MDM modules of data quality procedures such as profiling, standardization, elimination of duplicates and consolidation are applied to the data
  • Hadoop / Greenplum: Successfully completed Hadoop POC project extending company Data Integration layer to include Hadoop as source and Greenplum as target as well as a framework for distributed processing of large data sets. The framework is built as a data layer where all company data is integrated (thru Java, Python, Sqoop, and Pig code) for consuming by business users and application teams.

Environment: Oracle 11g, Linux, Talend 5.1, Informatica MDM, Hadoop, Pig, Python, Sqoop, Pivotal DCA / Greenplum 4.2

Confidential, Iselin, NJ

Data Integration Architect / ETL Lead

Responsibilities:

  • Participated in requirements gathering and documentation, and created functional specification documentation, interacted with end business analysts to fulfill the real - time business data requirements.
  • Real - Time ODS: Architected a process to receive messages from MQ system from outside vendor (LPS) and process using Informatica real time web services hub via in-house messaging system. Led the team of developers to program the Informatica components to bring the real-time data into ODS system 24/7.
  • Integration of Batch / Real - Time Data: Designed a complex system of integrating real-time data with batch data while maintaining the latest near real-time data with in 30m sec interval in ODS system. Led the team of developers to program stored procedures for the integration of data.

Environment: Oracle 11g, Linux, Informatica 9.1.

Confidential, Bridgewater, NJ

ETL Architect / Solution Architect

Responsibilities:

  • Daily routine includes project initiation, requirements gathering, and understanding business requirements, interacting with business users, writing functional specification documentation, data analysis, planning, data modeling, process architecture, writing detailed design documents, ETL development, system integration, and deployment.
  • Leading a team of 11 (Onshore/Offshore) Programmers / Analysts / Data Modelers.

Environment: UDB 7.2/8.1 (Data warehouse (EDW), Data marts), Teradata 8.1(EDW), Informatica PowerCenter 8.1, Windows2000/NT, UNIX (Sun Solaris 2.6), Java, J2EE, Trillium 6.0.

Confidential, Parsippany, New Jersey

ETL Specialist

Responsibilities:

  • Installation, Configuration and Administration of Informatica PowerCenter 5.1.2 on development, QA, production environments. Mentoring Informatica to project members, and enforcing standards and best practices.
  • Prepared Technical, Detailed Design Documents for Mega Loyalty project.
  • Modeled Staging database for Tibco Feed, Mega Loyalty Data mart (Star Schema), Operations (FIS) Data mart (Snowflake) using Erwin 4.0 while helping the architect design Operations (CRS) Data mart.
  • Designed and Developed complex mappings/mapplets to load Operational, Marketing, and Trip Rewards Data marts from OLAP and EDW (terabyte warehousing system) while implementing logging statistics functionality in the mappings. Designed, Implemented Enforced a common methodology for Initial and Incremental loads for team members to follow.
  • Advising team members to follow development methodologies to achieve performance needed to load from million rows tables from EDW.
  • Created custom reports using PowerAnalyzer and helped users comparing the results of Data marts and existing OLAP systems.

Environment: Oracle 9.2 (Data marts), Informix XPS 8.32 (EDW), Informix 7.3, Oracle 7.1, Flat Files (Sources), Informatica Power Center 5.1.2, Windows2000/NT, J2EE, UNIX (Sun Solaris 2.8)

Confidential, Macon, Georgia

Sr. Programmer/Analyst

Responsibilities:

  • Developed new modules, reports, bar code generation programs for SQL*LIMS users with new methods, and operations to enter and track the tasks and results. Support the system from various testing instruments (Gas Chromatographs, Mass Spectrometers).
  • Prepared Functional, Detailed Design Documents for Rods.
  • Day to Day interaction with R&D scientists for their reporting needs and design/reporting changes of RODs.
  • Programming in ESQL/C to load complex reporting structures.
  • Designed part of RODs EDW data warehouse as per the new requirements and prepared technical documents for acceptance by Senior Managers (R&D).
  • Modeled staging database for fast loading from OLAP systems using Erwin 3.5 for FabOp.
  • Modeled FabOp (Star Schema) Maintenance Data mart using Erwin 3.5.
  • Designed and Developed complex mappings/mapplets, procedures (PL/SQL) to FabOp from OLTP systems.
  • Created custom reports using Brio 5/6.6 for FabOp users.

Environment: SQL*LIMS, Oracle 8i (Data marts), Oracle 8i (Staging), Informix 7.3 (EDW), Informix 7.3, Oracle 7.3.2, Flat Files, Mainframe Data Files(Sources), Informatica PowerCenter 5.1, Brio 5/6.6, Windows2000/NT, UNIX (HP/UNIX (10.2)), UNIX(AIX).

Confidential, Dallas, Texas

Sr. Programmer/Analyst

Responsibilities:

  • Developed mappings, mapplets, reusable transformations, custom procedures, and packages (PL/SQL) to load into EDW and Data marts from various source systems.
  • Participated in designing standards for error handling.
  • Interacting with QA group to fix bugs and changes in the design.
  • Assisted the architect design staging database.

Environment: Oracle 8i (Data marts, staging), Oracle 7.3 (EDW), Sybase, Oracle 7.3, Informatica PowerMart 4.7, Windows 98, UNIX (Sun Solaris 2.4)

Confidential, Memphis, Tennessee

Programmer/Analyst

Responsibilities:

  • Developed complex custom GUI Applications using JAVA2, designed/programmed broker servers (back end (JAVA2)).
  • Participated in requirements gathering from user group for extra features and modifications.
  • Testing the system for User Acceptance and end users on the system.
  • Developed and deployed a small web project (Apache, HTML, and JavaScript) for internal users to generate reports from existing systems.

Environment: Sybase, J2EE, RMI, Apache, Windows 98, UNIX (Sun Solaris 2.6)

Confidential, Dallas, Texas

Programmer/Analyst

Responsibilities:

  • Programmed ESQL/C application programs to achieve fast loading (12000 rows/sec). Programming involved data extraction, cleaning and fast loading using the INFORMIX parallel loader in staging dimensional (Star Schema) data warehouse.
  • Build highly complex ESQL/C common libraries for usage by application programs. These libraries manage memory for aggregation, lookups (binary search of lookup data on HEAP), calculations on HEAP for the data from flat files and dimensional tables. The high volume of data is handled on Sun box with 25 GIG of RAM.
  • Unit testing, Load testing, and Deployment into production.

Environment: Informix XPS 8.11, Flat Files, UNIX (Sun Solaris 2.6)

Confidential, St. Louis, Missouri

Programmer/Analyst

Responsibilities:

  • Participated in the design process of the project including user requirement gathering.
  • Designed the database to store letter templates. programming in JAVA objects to manipulate data in the database using JDBC 1.22 as API and Oracle’s JDBC OC17 as driver. Programmed branch interface used by GUI to communicate with RPC server.

Environment: Oracle, J2SE, JDBC 1.22, UNIX (Sun Solaris 2.6(SUN OS 5.5.1))

Confidential

Responsibilities:

  • Duties include programming in PRO*C (ORACLE 7.3.23) which will be invoked by DIA, and programming in C++ using RougeWave libraries.

Environment: Oracle 7.3, UNIX (Sun Solaris 2.6(SUN OS 5.5.1))

We'd love your feedback!