Informatica Data Integration Tech Lead (contract) Resume
Richmond, VirginiA
PROFESSIONAL SUMMARY:
- Holding TN visa, a Canadian, with 20 years IT experience, 14 years Informatica ETL Data Integration experience in all phases of SDLC/Agile cycles, multi - fields certified Data Integration Specialist, with in-depth knowledge and extensive hands-on experience in data analysis, data design, data integration development, data quality development, big data and cloud solutions, is looking for a Data Integration, Data Quality consultant/developer position.
- Broad knowledge in various data integration tools and technologies, such as Informatica PowerCenter, Informatica Developer tool, Informatica Data Quality, Informatica Analyst, Informatica Big Data Management (BDM), Informatica Cloud ICS and IICS, Informatica B2B, Informatica PowerExchange, Informatica Metadata Manager, Informatica Web Service, Informatica Integration with Teradata, Informatica DVO, Python programs, Powershell scripts, Unix scripts, Java program etc.
- 4 years Data Quality application requirement, analysis, design, development, deployment and maintenance
- 3 years Python program
- 3 years hands-on expertise with big data technologies
- 2 year hands-on expertise with Cloud, including Informatica ICS and IICS, Amazon AWS and Salesforce
- 20 years hands-on expertise with databases, including RDBMS, NoSQL and Cloud databases
- 5 years hands-on expertise with UNIX shell scripts and commands
TECHNICAL SKILLS:
ETL / Data Integration: Informatica 7.x/8.x/9.x/10.x, Informatica PowerCenter, Informatica Developer, Informatica Data Quality, Informatica Analyst, Informatica Big Data Management (BDM), Informatica Cloud ICS and IICS, Informatica B2B, Informatica PowerExchange, Informatica Metadata Manager, Informatica Web Service, Informatica Integration with Teradata, Informatica DVO etc., SSIS
Reporting & Analytics: Tableau, R, Matlab
Big Data: Informatica Big Data Management, HDFS, Sqoop, Map Reduce, Hadoop, Spark, Scala, Python, Hive, Hbase, Impala, Hue, pig, Oozie etc.
Cloud: Informatica ICS and IICS, Amazon AWS services such as Redshift, RDS, S3, EC2 etc.
Languages & Development related Tools and Packages: Python, PyCharm, Spark, Scala, R, Java, Service Now, Jira, SharePoint, Control-M, Autosys iXp, ASP, VB.NET, Microsoft IIS, Visual Studio, Perl, Linux/Unix Shells, Oracle JDeveloper, WebLogic, Tomcat, IntelliJ, Git, SNV, VMware, Quality Center etc.
Database, Data Warehouse and SQL Tools: Oracle, SQL Server, DB2, AWS Redshift, AWS RDS, Teradata, PostgreSQL, Cassandra,HBase, MongoDB, Hive, PL/SQL, T-SQL, Toad, SQL plus, Oracle Sql Developer, Teradata Studio Express, BTEQ, Teradata SQL Assistant, DB Visualizer, Erwin etc.
Platforms: Linux, Unix, Windows
PROFESSIONAL EXPERIENCE:
Confidential, Richmond, Virginia
Informatica Data Integration Tech Lead (contract)
Responsibilities:
- Led POC project, analyzed existing SSIS code and stored procedure, transformed SSIS ETL to business logic, developed Informatica ETL for business logic; implemented enterprise application with control M
- Provided guidance regarding data integration solutions; led trouble-shooting for complicated issue; supervised junior staffs on the technical resolutions; provided training to peers with my specialized technical skills; knowledge and skills transfer to developers and business analysts based on Agile and Sprint methodology, actively joined daily scrums
- Integrated data between MongoDB and file/database with Informatica
Environment: SSIS, Informatica PowerCenter 10.1.1, Control M, SQL server Management Studio, Visual Studio, Unix, MongoDB
Confidential
Senior Data Strategy Specialist (contract)
Responsibilities:
- Extracted/loaded financial transaction data from/to all kinds financial institutions and data warehouses, including: JPMorgan Chase, Bank of America, Wells Fargo, Citigroup, Goldman Sachs, Morgan Stanley, State Street, Capital One, Barclays, CRD, BNY, Bloomberg, IDW (Investment Data Warehouse), PRIME (Portfolio Repository Investment Management Enterprise) etc.
- Developed Informatica10.2 ETL to apply Python programs, Unix scripts and Power shell scripts for enterprise financial data integration
- Developed Python program to analyze file, extract financial data and load to databases
- Developed Python programs, Unix scripts and Power shell to call Informatica workflows
- Developed Autosys JIL scripts; integrated all kinds of enterprise app jobs with Autosys iXp
- Gave professional solutions for data integration, troubleshooting and performance tuning
Environment: Informatica Powercenter 10.2, Python, PyCharm, SQL Server Management Studio, Oracle SQL Developer, Autosys iXp, SVN, Unix, power shell
Confidential
Data Integration/Data Quality/Hadoop/C loud Tech Lead (contract)
Responsibilities:
Informatica ETL Data Quality Analysis and Development:
- Based on Agile and Sprint methodology, actively joined daily scrums
- As primary analyst, based on data quality standard, used Informatica 10.1.1 IDQ Analyst for analysis, such as, creation and analysis of profiles for enterprise data, creation and installation of informatica glossary desktop and business glossary service
- As primary developer, based on data quality standard, used Informatica IDQ 10.1.1 Developer to develop DQ applications for Loblaw to cleanse data, validate data and implement data exception management
- Developed scorecards to present and monitor data quality trend
- Supported data governance COE to provide best data assets, and governed data at enterprise level
- Did knowledge and skills transfer to developers and business analysts; provided guidance to use Informatica 10.1.1 IDQ tools; fixed critical bugs
- Wrote professional Data Quality documents based on data quality standard
- Did knowledge and skills transfer to analysts to implement Informatica Metadata Manager; created Informatica metadata, custom model; wrote MM view queries; browsed and analyzed metadata
Big Data Hadoop and Spark:
- Used Informatica Big Data Management to load data to Hadoop HDFS, Hive and HBase; configured Informatica Big Data Management for column profiling; cleansed data on Hadoop platform; applied Informatica Spark engine to improve performance
- Monitored Informatica Big Data workflows with YARN web user interface; troubleshooted Scala code in the developer Spark execution plan
- Executed Sqoop command to import initial data to HDFS and Hive; created Sqoop saved job to execute incremental load
- Combined Spark with Python/Scala to transform large amount of Hadoop files between all kinds of file formats, such as Parquet, Json, Avro, compressed, sequence, csv.
- Used Spark Sql, Impala, Hive query, Hue, HBase query etc. to analyze data; cleansed, transformed large datasets to small datasets for the later analysis
- Create sales aggregations per item at product group level for analysis using Spark Python/Scala code
- Wrote Python applications with PyCharm; created, deployed and executed applications in Hadoop Yarn cluster
- Wrote Scala applications with IntelliJ; created, deployed and executed application Jar assembly in Hadoop Yarn cluster with SBT
Cloud:
- Installed Informatica Cloud Secure Agent
- Developed Informatica Cloud Data Integration mapping and taskflows to exact and load data between on-premises, AWS RDS, Amazon S3, Redshift, Azure SQL Data Warehouse and Azure Data Lake Store; created and configured all kinds of cloud connections and runtime environments with Informatica IICS
- Operated AWS console to configure services and configurations
- Developed Redshift queries and RDS queries to confirm data is loaded correctly
- Created Salesforce connection, Implemented Salesforce business process with Informatica IICS data Integration
- Did knowledge and skills transfer
Analysis: Developed Tableau statistical analysis sheets and dashboards
Environment: Hadoop cluster, IntelliJ, PyCharm, Informatica powercenter; Informatica 10.1.1 IDQ package ( Informatica Developer, Informatica Analyst and Informatica Business Glossary); Informatica Metadata Manager; Informatica Big Data Management; Informatica Cloud; AWS; Azure; Workbench/j; Python, PyCharm, SQL Server Management Studio; Tableau; Oracle SQL Developer, Oracle SQLPlus, Git
Confidential
Senior Informatica Data Quality Consultant (contract)
Responsibilities:
- As primary analyst, based on data quality standard, used Informatica IDQ 9.6.1 Analyst to analyze complex financial data sets
- As primary developer, based on data quality standard, used Informatica IDQ 9.6.1 Developer to develop DQ apps to cleanse data, validate data and implement data exception management for complex financial data issues
- Developed complicated data quality rules with complicated transformations, such as Java transformation
- Wrote professional Data Quality documents based on data quality standard
- Developed scorecards to present and monitor data quality trend of financial data sets
- Did knowledge and skills transfer to developers and business analysts; provided guidance to use Informatica IDQ tools
Environment: Informatica powercenter; Informatica 9.6.1 IDQ package ( Informatica Developer and Informatica Analyst); DB Visualizer; PL/SQL
Confidential
Senior Informatica ETL Data Integration Consultant (contract)
Responsibilities:
- As primary ETL consultant, completely re-designed the data integration solutions, and successfully made the deliveries in time
- Played as a business analyst role, proactively communicated with the end users; identified, documented and analyzed the requirements; systematically created documents, such as requirement, functional, mapping and workflow documents
- Re-designed and modified target database schema; created and maintained database objects; developed stored procedures complementing the data integration solutions
- Analyzed the transformation algorithms in .NET modules; refactored these modules into Informatica workflows; generated XML documents with complicated target data structure
- Integrated with Informatica with multiple XML related skills, such as XQuery, XSLT, XPath, XSD, DTD, DOM, LINQ etc.
- Created conceptual, logical and physical data modeling with ERwin
Environment: Informatica PowerCenter 9.6.1, multiple XML skills sets, Visual Studio, Management Studio, PL/SQL and T-SQL, Erwin
Confidential
Senior Informatica Data Integration Tech Lead (contract)
Responsibilities:
- As primary consultant, designed and developed applications with Informatica PowerCenter, Java, XQuery and PL/SQL
- Provided guidance and instructions regarding data integration solutions; led the trouble-shooting for complicated issues; supervised junior staffs on the technical resolutions; providing training to peers in some of my specialized technical skills; guided project; did knowledge and skills transfer to developers and business analysts
- Implemented Informatica Web Service for system integration use cases
- Transformed and cleansed large volume of petabytes electricity/gas/water meters real time data from the whole Alberta province
- Participating in the Unit Testing, UAT and SIT based on Agile and Sprint methodology, actively joined daily scrums
Environment: Informatica 9.5 PowerCenter, Informatica Web Service, Oracle Sql Developer, Toad, PL/SQL, T-SQL, Unix and Window scripts, Java, XML, Jira
Confidential
Senior Data Integration/ Hadoop/ Cloud Tech Lead (contract)
Responsibilities:
Big Data Hadoop and Spark POC:
- Extracted/transformed/loaded data on Hadoop HDFS, Hive and Hbase with Informatica big data edition; configured Informatica big data edition for column profiling, pushdown and performance tuning
- Developed Sqoop job to import/export/incremental data
- Combined Spark/Python/Scala skills to load/transform/export a large amount of Hadoop data between all kinds of formats, such as Parquet, Json, Avro, compressed, sequence, csv etc.
- Used Spark Sql/Python/Scala/Impala/Hive/Hue/Hbase/pig/ etc. to analyze large amount of Hadoop data
Cloud POC:
- Installed Informatica Cloud Secure Agent
- Developed Informatica Cloud data integration with Informatica Cloud Services (ICS) to exact data between on-premises and Salesforce cloud
- Did knowledge and skills transfer
Informatica ETL Analysis and Development:
- Based on Agile and Sprint methodology, actively joined daily scrums; adjusted project; provided strong tech support
- Led and built four Informatica domain environments, i.e. development, test, pre-production, production, including server and client installation and configuration
- As primary consultant, with Informatica PowerCenter, developed highly sophisticated data loading applications; analyzed and loaded dada from oil & gas files (with all sorts of file formats including LLR,OGC,PRA,SIR, AER, UAL etc.) for Alberta, British Columbia and Saskatchewan provinces
- Systematically developed documents, such as requirement, functional, mapping and workflow documents
- Administrated and monitored Informatica servers, including user management, backup, upgrade, patch applications, etc.; support existing Informatica applications
- Developed Tableau statistical analysis sheets and dashboards
Environment: Informatica PowerCenter 9.5, Informatica Big Data Edition, Informatica Cloud, Salesforce cloud, multiple Big Data tools, PL/SQL, T-SQL, Unix and Window scripts, Tableau, Java, SVN
Confidential
Informatica Data Integration Consultant
Responsibilities:
- Supported whole enterprise Informatica ETL applications
- Participated client engagement and communicated with the clients for requirement analysis, application design and production system support
- Successfully designed, developed, deployed and supported Informatica applications for oil and gas industry using Informatica PowerCenter
- Tackled intricate issues, identified root causes and implemented corresponding resolutions and/or enhancements for massive and complicated Informatica workflows; designed and developed complementary codes of JAVA, XML, UNIX shells and PL/SQL and integrated the codes with Informatica workflows
- Configured Informatica PowerExchange package, Informatica Oracle Express real-time CDC; developed Informatica real-time CDC workflows
- Created tests with Informatica DVO; guided the Unit Testing, UAT and SIT
- Providing training to peers in some of my specialized technical skills
- Created conceptual, logical and physical data modeling with ERwin; created reverse and forward engineering
- Ralph Kimball, Bill Inmon, OLTP and OLAP methodology
Environment: Informatica 8.6.1 PowerCenter, Erwin, Oracle Sql Developer; Toad; PL/SQL and T-SQL, Unix and Window scripts, Java, XML, Service Now, Quality Center
Confidential
Solution Architect
Responsibilities:
ETL Data Integration and Analysis
- Wrote Python scripts to connect relational database automatically, and implement database DML operations; wrote Python scripts to parse data from XML files
- Used Informatica B2B packages to extract and transform data from any file formats to any file formats, including unstructured data, semi- structured and structured like HL7, also integrated with Powercenter package to load data
- Designed, developed and deployed large Informatica workflows with transformations and cleansing logics for complicated data integration and data quality requirements; built reverse-engineer Informatica mappings into Visio templates, published template, generated multiple mappings
- Used Informatica Data Quality plug in to analyze and clean/match/consolidate data; created reference tables, rules, profiles etc.
- Built Informatica PowerCenter development environment, including installation and configuration of PowerCenter client tools, Informatica PowerCenter server, repositories, account management, directories, services, grid, version control, metadata manager and web service applications; updated Informatica systems to higher versions
- Implemented Informatica PowerExchange package; configured Informatica Oracle Express real-time CDC; developed Informatica real-time CDC workflows and batch workflows
- Used Informatica Data Mask, for security reason, masked sensitive data to invalid data for development and test
- Configured, integrated Informatica with Teradata with Informatica Teradata Parallel Transporter connection, Informatica Teradata relational connection, and Informatica Teradata External connection; Wrote Teradata sql statements with Teradata Studio Express, BTEQ, and Teradata SQL Assistant
- Configured and Developed Informatica JMS and MQ series real time Integration workflows
- Created conceptual, logical and physical data modeling with Erwin; provided technical suggestions to data modeler and data architect for database design
- Wrote Autosys JIL scripts to schedule enterprise informatica workflows
- Developed SSIS ETL for specific projects
- Developed Tableau statistical analysis sheets and dashboards; published Tableau workbooks
- Participated the development of design, test, development and deployment documentations
- Developed data replication using GoldenGate and built database high availability (RAC and DataGuard)
Environment: Informatica PowerCenter 7.x and 8.x, Informatica B2B, Informatica Teradata Integration, Informatica PowerExchange, Informatica Data Quality plug in, Informatica JMS Integration, SSIS, Tableau, PL/SQL and T-SQL, Unix and Window scripts, Java, XML, SVN, Autosys
Confidential
Java Developer & Oracle administrator
Responsibilities:
- Developed Java web applications using JSP, Servlet, JavaBeans and Javascript
- Administrated Weblogic and Tomcat servers
- Administrated Oracle 8i/9i databases for development and test environments including backup and recovery, server & client software installation, upgrade and configuration; performed data migration, performance monitoring and tuning, trouble-shooting and problem resolution
- Performed Linux system administration tasks, such as building server, configuring kernel parameters, installing/configuring NFS server, Samba server and Apache server, VNC etc.
Confidential
SQL Server &Windows Administrator
Responsibilities:
- Administrated SQL Server databases and Windows servers for operation and maintenance