Data Warehouse Architect Resume
Rochester, MN
SUMMARY
- Over 7 years of IT experience in Analysis, Design, Modelling, Development, Testing and Implementation of Data Warehousing Business Applications.
- Excellent experience in Data Warehouse applications, directly responsible for the Extraction, Transformation and Loading of data from multiple sources into Data Warehouse using various versions of Informatica PowerCenter.
- Strong experience in entire life cycle of Data warehouse/Data mart using Informatica Power Centre (Designer, Workflow manager, Workflow monitor, Repository manager).
- Extensive experience in creating mappings, mapplets and complex transformations using Informatica Designer and scheduling workflows, worklets and tasks using Informatica Workflow Manager.
- Thorough knowledge of Ralph Kimball's Data Modeling concepts including Dimensional Modelling, Star and Snowflake schema, Slowly changing Dimensions (Type1, Type2, and Type3) and Surrogate Keys.
- Good experience in UNIX and writing shell scripts for Informatica pre & post session operations and database administration activities.
- Experienced in Installation, Configuration, and Administration of Informatica Power Center/Power Mart on Client/Server environment.
- Good Experience in analysis and resolution of bottlenecks for tuning Informatica mappings to improve performance.
- In depth knowledge of RDBMS concepts, with hands on exposure in the development of relational database environment using SQL, PL/SQL using database tools like Toad.
- Worked on various databases like Oracle 11g/10g/9i, SQL SERVER, and on various operating systems like UNIX and Windows.
- Highly motivated and adaptive with the ability to grasp things quickly with excellent interpersonal, technical and communication skills. Ability to work effectively and efficiently in a team and individually.
TECHNICAL SKILLS
Data Warehousing: PowerCenter 9.1/8.6/8.x/7.x, PowerMart 7.x, PowerConnect, Power Exchange.
Data Modelling: Dimensional Data Modelling, Star Schema Modelling, Snow - Flake Modelling, FACT and Dimensions Tables, Physical and Logical Data Modelling, ERwin.
Front-End Tools: TOAD, SQL Navigator.
Databases: Oracle 10g/9i/8i, SQL Server, DB2.
BI & Reporting Tools: Cognos EP 7 Series, Cognos ReportNet 1.1, Cognos 8.1 BI.
Programming: C, C++, Java, SQL, PL/SQL, UNIX Shell Scripting.
Operating Systems: Windows XP/2003/2000/NT/98/95, Unix.
PROFESSIONAL EXPERIENCE
Confidential, Rochester, MN
Data Warehouse ArchitectEnvironment: Informatica Power Centre 9.1/ 8.6, Oracle 11g, SQL, SQL Server, UNIX Shell Scripting, Toad
Responsibilities:
- Analysed the Functional Specs provided by the Architect and created Technical Specs documents for all the mappings.
- Worked as part of the development team, along with the Systems Analysts/Business Analysts.
- Design and development of mappings and mapplets using Informatica PowerCenter Designer to populate staging tables.
- Analysed the source data coming from fixed-width flat files and working with business users and developers to develop the Model.
- Extensively used transformations like Source Qualifier, Filter, Aggregator, Connected and Unconnected Lookups, Joiner, Expression, Sequence Generator, Update Strategy and Router transformations to extract and transform data in accordance with the business logic.
- Involved and create complex mappings and populate the relational tables in an efficient manner.
- Used Workflow Manager for managing workflows and Session, database connections, memory properties and perform Pre- and Post-Session tasks.
- Performance tuned the workflow by identifying the bottlenecks in sources, targets, mappings and sessions. Identifying read and writes errors using Workflow and Session logs.
- Used Parameter files to initialize workflow variables, Mapping parameters and mapping variables and used system variables in mappings for filtering records in mappings.
- Documentation to describe logic, coding, changes and corrections.
- Performance Tuning of theInformaticaMappings and database by adopting Explain plans, cutting down query costs using Oracle hints, changing the mapping designs
- Created Unix shell scripts for flat file handling, creating list of file which is input forinformaticaread, pre post session operations on source and target databases ( i.e dropped and re-created indexes, collect stats),executed these shell scripts through pre post session command tasks in workflow.
- Created oracle Tables (partition tables), Indexes, constraints, Views, Synonyms, functions and procedures, developed the SQL queries for data validation and tuned the SQL override queries by providing optimizer hints.
- Performed unit testing which validates the data and system testing which provides qualitative check the overall data flow in to target tables, and volume testing with full production load.\
- Supported for Quality Assurance Testing (QAT), User Acceptance Testing and production deployment.
- Migrated theInformaticacode from Development to Test and Test to production, Created ETL mapping documents, Unit Test documents and run books for new jobs.
Confidential, Culver City, CA
Environment: Informatica Power Centre 8.6, Oracle 11g, SQL, SQL Server, UNIX Shell Scripting, Toad
Responsibilities:
- Analysed business requirements and coordinated with the business analyst to develop ETL procedures that confirm to the enterprise standards while maintaining consistency across all applications and systems.
- Parsed high-level design specs to simple ETL coding and mapping standards.
- Analysis, Design and Development, test and implementation of Informatica transformations and workflows for extracting the data from the multiple legacy systems.
- Responsible for the Data Modeling and populating the business rules using mappings into the Repository for Meta Data management.
- Used most of the transformations such as the Source Qualifier, Expression, Aggregator, Connected & unconnected lookups, Filter, Router, Sequence Generator, Sorter, Joiner, and Update Strategy.
- Attended meetings to finalize strategies for handling various concepts like Error Handling and Slowly changing dimension
- Troubleshoot problems by checking sessions/error logs in the Workflow Monitor and also used debugger in Mapping Designer to debug complex mappings.
- Performance tuned the workflows by identifying the bottlenecks in targets, sources, mappings, sessions and workflows and eliminated them.
- Enhanced performance for Informatica session using large data files by using partitions, increasing block size, data cache size, sequence buffer length and target based commit interval.
- Developed UNIX Shell scripts for data extraction and file transfers.
- Assist operation support team for transactional data loads.
- Performed Unit and System Testing of developed mappings.
- Documented program development, logic, coding, testing, changes and corrections.
- Worked with pmcmd to interact with Informatica Server from command mode and execute the shells scripts.
- UsedInformaticaRepository Manager to maintain the Metadata, Security, and Reporting.
- Involved in writing project documentation of the ETL Process flow for better maintenance and analyzing the process flow.
- Profiled the data usingInformaticaData Explorer (IDE) and performed Proof of Concept forInformaticaData Quality (IDQ).
- Involved in production and deployment phase to make sure the job schedules as dependencies are developed in such a way that we are not missing the SLA on a day to day basis.
Confidential, MINNETONKA, MN
Environment: Informatica Power Centre 8.6, Oracle 11g, SQL, UNIX Shell Scripting, Toad, SQL Server 2008, DB2, IMS data, Power ConnectPL/SQL
Responsibilities:
- Involved in all phases of SDLC from gathering requirements, design, development, testing, support for UAT and production deployment. Involved in design of logical physical Dimensional data models and implementing Star Schema with Fact and Dimension tables
- Used ER-win to identify Facts and Dimensions.
- Designed and developed Mappings using Mapping Designer to load the data from various sources using different transformations like Expression, Aggregator, Lookup (connected and unconnected), Filter, Router, Joiner, Stored Procedure, Sorter, Sequence Generator and Update Strategy transformations.
- Created Sources, Targets in shared folder and developed re-usable transformations, mapplets and user defined function (UDF) to re-use these objects in mappings to save the development time.
- Designed and Developed complex mapping with multiple target load plans with mapplets reusable transformations and mapping parameters variables to process initial incremental data.
- Decreased load time by implementing Performance Tuning techniques by using lookup caches, providing sql override in lookups, using tables with fewer rows as the master table in joiner transformations, passing the sorted data into aggregate transformation, dropped indexes and re-created them after loading data to targets.
- Using workflow manager, created workflows, worklets, sessions, command and email tasks.
- Created sequential concurrent data flows with worklets sessions tasks with appropriate link conditions as per data flow requirements.
- Configured the session properties i.e. high value for commit intervals, external loaders (SQL* Loader) to increase the performance.
- Extensively used almost all the transformations, which includes (Sequence Generator, Expression, Filter, Router, Sorter, Rank, Aggregator, LOOK UP (Static and Dynamic), Update Strategy, Source Qualifier, Stored Procedure, Joiner, Normalizer, and XML Source Qualifier).
- Scheduled Sessions and Batches on the Informatica Server using Informatica Server Manager/Workflow Manager.
- Performed Adhoc queries using PL/SQL developer to check for consistency of the data in the tables and to update the tables as per the requirements.
- Wrote stored procedures for dropping and re-creating indexes for efficient Data Load.
- Tuning of SQL queries for better performance.
- Optimizing the mappings by changing the logic to reduce run time
- Developed the mappings to extract the data on incremental load basis
- Developed reusable transformations for Date Conversion.
- Responsible for UNIT, System and Integration testing. Developed Test scripts, Test plan and Test Data.
- Extensively involved in Fine tuning the Informatica (mappings and Sessions), Stored Procedures, SQL to obtain optimal performance and throughput.
- Writing documentation to describe program development, logic, coding, testing, changes and corrections.
Confidential, Wilmington, DE
Environment: Informatica Power Centre 8.6, Oracle 11g, SQL, UNIX Shell Scripting, Toad, SQL Server 2008, DB2, IMS data, Power Connect, PL/SQL
Responsibilities:
- Analysed the Specifications and involved in identifying the source that data needs to be moved to data mart.
- Involved in analysing scope of application, defining relationship within & between groups of data, star schema, etc.
- Identified and created different source definitions to extract data from input sources like Flat files, SQL Server and load into relational tables like Oracle.
- Enhanced and created various database objects in Data Mart as per changing technical requirements.
- Extensively involved in requirement analysis and created ETL mapping design document
- Extensively Designed, Developed, Tested complexInformaticamappings and mapplets to load data from external flat files and other databases.
- Extensively Worked onInformaticatools such as Source Analyzer, Data Warehouse Designer, Transformation Designer, Mapplet Designer and Mapping Designer.
- Extensively used all the transformations like source qualifier, aggregator, filter, joiner, Sorter, Lookup, Update Strategy, Router, Sequence Generator etc. and used transformation language likes transformation expression, constants, system variables, data format strings etc.
- Involved in running the loads to the data warehouse and data mart involving different environments.
- Extensively worked on workflow manager and workflow monitor to create, schedule, monitor workflows, worklets, sessions, tasks etc.
- Extensively worked on ETL performance tuning for tune the data load, worked with DBAs for SQL query tuning etc.
- Responsible for definition, development and testing of processes programs necessary to extract data from client's operational databases, Transform and cleanse data, and Load it into data marts.
- Extensively used PL SQL programming in backend and front-end functions, procedures, packages to implement business rules.\
- Integrated various sources into the Staging area in Data warehouse
- Provide technical support to Quality Assurance team and Production group.
- Support Project Manager in estimating tasks.
- Provided production support to schedule and execute production batch jobs and analyzed log files
- Extensively involved in ETL testing, Created Unit test plan and Integration test plan to test the mappings, created test data.
- Involved in all phases of Data quality assurance.
- Developed shell scripts to automate the data loading process and to cleanse the flat file inputs
Confidential, Hartford, CT
Environment: Informatica Power Centre 8.1, Oracle 10g/9i, SQL, UNIXShell Scripting, Toad, SQL Server 2008, DB2, IMS dataPower Connect, PL/SQL
Responsibilities:
- Extensively usedInformaticapower center for loading the data from various sources involving flat files, oracle database, spread sheet etc to target (like flat file or database table).
- Creation of Mapping or mapplet using different transformations and validate the mapping.
- Creation of different transformations like lookups, Expression, joiner, Source Qualifier, Sequence Generator, Router, Filter etc.
- Creation of session for created mapping and add the session logic likeconnection, log file etc. and validate the session.
- Schedule and monitor the sessions work flows usinginformaticawork flow manager and work flow monitor.
- Analyze the existing mappings, if it is producing errors and rectify the errors.
- If any requirementchanges then modify the whole life cycle ofinformatica.
- Responsibility involved in Non-Function point Estimation, Impact Analysis, Defect Closing, preparation of HLD, LLD, Test Case etc. maintain the coding standard according to the Naming convention.
- Involved in Client Handling, fix ups the defect, peer review and status updating, defect log, finalizing the unit test cases, integration test cases and testing along with the peers.