We provide IT Staff Augmentation Services!

Data Warehouse Senior Developer Resume

2.00/5 (Submit Your Rating)

SUMMARY:

  • Data Engineer with 8 years of industry experience specialized in Data Warehousing, Dimension Modelling and Data Integration solutions using Informatica tools and scripting languages.

TECHNICAL SKILLS:

ETL and Data Mining Tools: Informatica Power Center 10.1/9.x, Admin console, Informatica Data quality( IDQ), Informatica Power Exchange 9.x, Apache Hadoop(HDFS), Sas studio, Sas e - Miner, R-studio

OLAP/Reporting Tool: SAP Business Objects XIR3, Tableau, MS Access

Databases: AWS EC2, Oracle 12c/11g, Sql server, Netezza

Data Modelling and SDLC: Ralph Kimball s methodology, Entity relationship, Dimensional modelling(star and snow flake schema), Bus Matrix, Agile, Waterfall

Programming Languages: SQL, PL/SQL, Unix Scripting, Python, R & R Mosek, C

Scheduling/Versioning/Agile tools: BMC Control- M, Clear Case, Tortoise SVN, Bit Bucket, JIRA

DB and other Tools: Toad, SQL Developer, SQL*Loader, XML spy, Putty, Winscp

Build/Deployment/Review tools: Team city, Urban deploy, Confluence

Operating Systems: Red hat-Linux, Sun Solaris 2.x, IBM-AIX, Microsoft 2000/2003/XP

PROFESSIONAL EXPERIENCE:

Confidential

Data warehouse Senior Developer

Responsibilities:

  • Worked closely with product owners to translate functional and technical requirements into a seamless data warehousing solution (Kimball approach) which can cater to both pre-defined and ad hoc reports.
  • Used Golden gate for change data/metadata capture on source database. Implemented incremental load/drip feed strategies to load the data into ODS layer using Informatica.
  • Created slowly changing dimension (SCD-2) solution using MD5 for all the dimensions including the dimensions required for the transactional fact table.
  • Created complex SQL queries and stored procedures to load data into various dimension and fact tables.
  • Created UNIX shell scripts for dynamic generation of parameter files and scheduling of Informatica workflows (catering to both serial as well as parallel order of workflow execution).
  • Created Junk dimension for multi valued attribute and Mini dimensions for rapidly changing attributes in Informatica.
  • Created summary facts and cross walks on top of transactional facts to improve the performance of reports.
  • Created accumulated facts for weekly aggregation of facts records for weekly reports.
  • Used bit bucket for version control, created test scripts for unit and integration testing and deployed the code to higher environments using deployment groups in repository manager.
  • Helped Tableau team by writing custom SQL queries for daily/weekly/quarterly reports using both live connection and tableau extract feature.
  • Extensively worked on optimizing the data seek time for reports by creating table partitions, Indexes and parallel hints.

Environment: Informatica PowerCenter (10.1.1), AWS EC2, Oracle12c, SQL, PL/SQL, UNIX shell script, Data Modelling, MS Visio, Toad, SQL Developer, Bit Bucket, Confluence, Team City, Urban Deploy, JIRA

Confidential

Informatica Senior Developer

Responsibilities:

  • Understood the functional and technical domains which the upstream legacy system handled (BAU) and created a prototype of what the new platform will look like for Data warehousing layer.
  • Created high level solution design document and engineered logical and physical models for the new platform.
  • Involved in workshops and brainstorming session with the Business Analyst and the end user to figure out the data and metadata mapping between source and target systems.
  • Developed mappings and workflows in Informatica Powercenter, processed various heterogeneous sources like Oracle table, Flat Files(delimited and fixed width), XML files and MQ source
  • Extensively used most of the basic transformation like Source Qualifier, static and dynamic Lookup cache, Aggregator, Joiner, Rank, SQL, Update strategy, Router, Sorter, Union etc. to tackle complex business logic in efficient way.
  • In addition to basic transformations, I also worked on some of the advanced transformations like Web Service consumer by importing its WSDL in Powercenter Designer. In this mapping the SOAP requests having customers’ information were getting passed onto a bank server (with two way-SSL encryption) and the response message or the SOAP fault thus received was getting captured into relational tables and error files respectively.
  • Created mappings for daily reconciliation of data for a critical asset management piece. This mapping made use of transformations like JAVA, Normalizer and Transaction Control to cater to complex logic.
  • Extensively worked on improving the performance of batch jobs by analysing explain plans, implementing indexes, database partitioning, parallel hints, Informatica partitioning, pushdown optimization and caching
  • Created a number of Unix scripts for data pre and post processing. The scripts made use of a gamut of basic commands to advanced commands like ‘Sed and Awk’.
  • Configured Secured file transfer between the hosts by exchange of public keys and created shell scripts to push and pull the files to and from the hosts.
  • Created source/target shortcuts, Mapplets and resuable Sessions in order to promote best practices and code reusability.
  • Created an error frame work (with the help of PM Tables in Informatica) using complex SQL queries and PL/SQL blocks/procedures and generated reports detailing business and technical errors in the batch run.
  • Created Control-M jobs to schedule Informatica workflows and UNIX scripts. Monitored daily batch jobs, created alerts for bridge support.
  • Worked on Informatica Admin console to create the users, groups and manage repository and Integration services. Also, managed repository back-ups and metadata- table logs’ purging /archiving and license renewal.
  • Create mappings in Informatica Data Quality by making use of transformations like Labeler, Standardizer, Address doctor, Case, Merge and Match to improve the quality of customer information.
  • Created Unit test cases and performed testing of individual code blocks as well as for the batch of workflows.
  • Followed Agile SDLC method while working on everyday tasks. Created sprints/iterations and tasks and provided estimations using planning poker, analysed burn down chart to understand the remaining work for the sprint.
  • Provided UAT/, EBF(Emergency bug fix) and production support for a number of incremental releases of the code.

Environment: Informatica PowerCenter (v9.6.1), Informatica Data Quality, Admin Console, Oracle12c/11g, SQL, PL/SQL, Unix shell script, SFTP, Python, Control-M scheduling tool, Webservice consumer, Data Modelling, MS Visio, Toad, SQL Developer, SVN, Confluence, Team City, Urban Deploy, JIRA

Confidential

Informatica Senior Develope

Responsibilities:

  • Liaised with client’s functional design team via telepresence/conference calls to understand the end to end requirement.
  • Extensively worked on refactoring a massive retail Data warehouse/Marts. Created Bus matrix and aligned the confirmed dimensions and facts for Enterprise Data warehouse model.
  • Developed mappings in Informatica catering to slowly changing dimensions (Type 1 & 2) with the help of md5 function and other basic ways.
  • Developed mapping in Informatica to load mini dimensions that were created for dimensions which were getting changed rapidly and key for the same were passed on to fact table.
  • Created a solution for Early arriving Facts in Informatica by holding the fact record in a staging/ suspense table till the dimension record is processed.
  • Worked on migrating tables in oracle to Netezza database for massive parallel processing. Made use of hash distribution for fact tables which improved the performance of the queries by 40 percent.
  • Implemented and enabled Incremental Loading for Staging Tables by using Mapping Variables Approach and created mapping for Parameter File generation
  • Created Unix Scripts and stored procedures/functions for complex business logics and called them from Informatica mappings/workflows.
  • Populated Metadata statistics for various sessions by looking up on Informatica Internal view rep sess log and OPB tables.
  • Automated Informatica deployment through PMREP utility by creating control files. Also, have created deployment groups and labels to migrate the code from Dev repository to QA and further to UAT and Prod.
  • Used Informatica Powerexchange to capture the change data on real time basis in asynchronous mode, dealt with issues pertaining to cold and warm start of PowerExchange server.
  • Designed Universes (kind of entity relationship model in BI layer) in Business Objects reporting tool which acted as the source of data and constraints for the reports and dashboards which helped the business to track the KPIs and take strategic business decisions accordingly.
  • Created high level and drilled down reports for the stores’ sales, Net cash flow, forecast of sales and profitability, top products by sales, top products by quantity etc.
  • Assisted testing teams and various go lives from application development perspective.
  • Defined the KPI’s for the project and designed the dimensional model using MS Visio, pre-processed transactional data using UNIX shell scripts & further cascaded the data into Informatica where it was extracted, quality checked, transformed & aggregated before loading to Data marts to be used for reporting purpose.
  • Created interactive Tableau dashboards for analysis of physicians’ productivity and medical billing processes.
  • Created an OLTP database and a database application in MS access for a consumer electronics store to model sales, procurement, and inventory in MS Access.
  • Migrated the database to Oracle DB hosted in AWS. Established connection to Oracle and Access with the help of linked tables by creating ODBC connections for Oracle and configuring TNS entries.
  • Forecasted the demand for different products using Exponential Smoothing Forecast for upcoming 3 days implemented with the help of analytical functions in Oracle.
  • Used the forecasted demand to manage inventory using EOQ and FOO inventory control models and arrived at reorder point.
  • Did associative and sequential rule mining in SaS E-miner and analyzed support, confidence and lift values to understand the consumer patterns in buying goods.
  • Worked on and configured University’s Hadoop cluster on distributed settings to create mapper and reducer classes for a word count program and also to deduce trigrams from various news articles.
  • Partitioned the data into test and train set, checked for skewness of the variables and applied square or log function in case of left skewed or right skewed respectively.
  • Ran Imputation on missing data before creating the regression models.
  • Created decision tree, random forest, regression (forward and stepwise) and neural network models to understand which features contribute more towards the sale
  • Also, performed market basket analysis, association and sequence rule mining to understand which products were sold together or in a sequence.

Environment: Informatica, Oracle 11g, AWS Amazon RDS for Oracle, SQL, PL/SQL, Dimension Modelling (Kimball’s approach), Tableau, SAS e-Miner, R-Studio, MS Access, Predictive Modelling (Regression, Decision Tree, Random Forest, Neural Network, PCA), Data Mining (Association and Sequential rule mining), MS Visio, Forecasting, Inventory Control models, Apache Hadoop (HDFS), UNIX shell script, Python

We'd love your feedback!