Sr Big Data /dw Architect Resume
Glendale, CA
SUMMARY
- Self - motivated DW/ETL Architect/Data Analyst with over 17 years of experience in fast-paced, delivery-focused Fortune 500 environments.
- Experience in Information Technology on handling multiple engagements in various capacities as Data Warehouse architect, ETL Architect, Project Manager, Modeler, Business Analyst, Designer and Developer.
- Experience in Banking, Mortgage, Utility, Finance and Entertainment Domains with dataset size more then 40+ tb
- Worked on various multi-site projects teams spread across US, Europe, Australia & India handling complex onsite-offshore engagements involving large data volumes.
- Strong experience on Data warehouse architecture, Modelling, Systems Integration, Database Design, ETL Design, ETL Optimization, Data Profiling, Data quality process design, Best practice implementation, ETL Development, Production support and team co-ordination.
- Data modeling experience using Erwin tool involving Dimensional and Fact tables in Operational Data Store and Enterprise Data Warehouse using Kimball or Inmon. Star and snowflake modelling.
- Strong ETL experience using Informatica Power Center 7.x & 8.6, 9.1, 9.6.1, Datastage 8.1, IBM CDC tool, IBM Balanced Optimizer, MDM, Knowledge of SAS DI 4.2
- Unix/Python Scripting experience
- Working experience on Hadoop-Hive, Greenplum 4.2, Vertica 7.2, Teradata 13.10, Netezza, Oracle 9i, and SQL 2012/2016 Server.
- Have experience of working on Business Objects 6.5, Tableau 10 and Web intelligence reports.
- Coordination with users during UAT.
- In-depth experience on process modeling, problem-solving and understanding of DW processes & best practices and onsite-offshore communication.
- Training in Mongo DB-NoSQL, Spark, Netezza & AWS
- Knowledge of Data Vault
TECHNICAL SKILLS
Databases, ETL & BI Tools: Informatica 7.x & 8.x,9.1, Datastage 8.1, IBM CDC tool 6.5, Balanced Optimizer, Business Objects, OBIEE 10.1.3.2, Oracle, SQL Server, DB2,Teradata 15.10, Visual Basic, ASP, Cold Fusion, Knowledge of SAS DI
Databases: Teradata 13.10, Greenplum 4.2, PostgresSQL, Vertica 7.2, Hive, Hbase, Mongo, Oracle 9i, SQL Server 2005/2012/2016 , HP Neoview 2.4
Operating System: Windows 2000, Windows XP/NT, Unix
Programming Languages: UNIX Shell Scripting
PROFESSIONAL EXPERIENCE
Confidential, Glendale, CA
Sr Big Data /DW Architect
Responsibilities:
- Work as primary Sr Big data/ETL Architect for Marketing team
- Working on pulling Leads, Prospects data from various source data from Hadoop hive, hbase & spark into Vertica
- Design/Development of Marketing datamart using vertica, Unix and Python framework
- Define and Improve tasks to ensure Data Quality.
- Design of database
- Created logical/physical data model using Erwin
- Source to target Mapping documents
- Extract and transform data from various source systems into Vertica MPP platform
- Attend the scrum meetings everyday
- Find performance bottle necks and provide industry standard solution.
- Standardize the process to the enterprise level.
- Work closely with the reporting team to design, develop and publish KPI reports for marketing team
- Worked on POC on AWS Redshift
- Exported data files from Linux to S3
- Created tables in postgresql in Redshift
- Copied data into tables from datafiles using COPY command
- Created few tableau reports on same
Confidential, Burbank, CA
DW Architect/Lead Developer
Responsibilities:
- Worked on Database design
- Create ETL mapping document
- Analysis and profiling of data
- Working on Unix & Python scripting and ETL development
- Created logical/physical data model using Erwin
- Work on HLD/LLD Documents along with mapping rules
- Sourced data from various source systems into Teradata platform
- Worked on creating ETL process using BTEQ scripting and Teradata db components
- Created a summary layer for reporting
- Converted existing BO reports to tableau dashboards
- Created different KPI using calculated key figures and parameters
- Developed Tableau data visualization using Cross tabs and Bar Charts
- Developed donut charts and implemented complex features in charts like creating bar charts in tooltip.
- Strong Dashboard design experience and passionate practitioner of effective data visualization. Familiarity with best practices around visualization and design.
- Provided Production support to Tableau users and Wrote Custom SQL to support business requirements.
- Utilized Tableau server to publish and share the reports with the business users.
Confidential, San Diego, CA
Lead DB/ETL Engineer
Responsibilities:
- Analysis of existing OLTP DB Source systems in SQL Server 2012
- Create ERD Diagram-Logical and Physical OLAP model for EDW using Visio 2010
- Create a source to target mapping document to create a EDW using Greenplum
- Create Architectural Design Documents
- Build EDW using postgresql/Unix/python scripting using Agile methodology
- Worked on Sending data from hdfs(Hive db) to greenplum using sqoop
- Worked on writing SSIS Stored procedure in SQL 2012
- Worked on pulling data from Hive into Greenplum platform using python scripting
- Development
- Optimize the process, data partitioning
- Set up of replication process
Confidential, Burbank, CA
DW Architect/Lead Developer
Responsibilities:
- Analyze the existing HP Neoview Database and Informatica code to understand the existing process for data load into DW.
- Provide a roadmap to migrate the existing database to a new MPP Database.
- Providing best possible solution for the migration and getting it approved from Architectural Board.
- Providing alternate solutions or approach to improve the existing process.
- Providing Architectural Design to the development team on the enhancement projects.
- Source data from HDFS file system into DW.
- Work on effort estimate for the migration project.
- Work on creating the project plan in Microsoft Project.
- Worked on converting Informatica components from existing DB to Greenplum.
- Created/modified set of Informatica mappings to do performance improvement on Greenplum DB
- Development of Greenplum DB Functions/sql as part of the conversion from current HP Neoview DB to Greenplum.
- Used python/Unix scripting
- Worked on converting the existing UNIX scripts to handle the source file movement and data load into Greenplum DB.
- Work with DBA’s to get the Database design approved and implemented.
- Define and Improve tasks to ensure Data Quality.
- Find performance bottle necks and provide industry standard solution.
- Standardize the process to the enterprise level.
- Timely communication to the Project participants and management team about the progress.
- Improve the existing process towards more automated, reduce failure risks, improve data quality and increased resource utilization.
Confidential, Irwindale, CA
ETL Architect/Project Lead
Responsibilities:
- Providing best possible solution and getting it approved from Architectural Board.
- Providing alternate solutions or approach to improve the existing process.
- Providing Architectural Design to the development team.
- Coordination with the Datastage ETL & Reporting Development team with Design/Development approach to create a Data Mart.
- Worked on developing of Teradata stored procedures.
- Worked on development of ETL components
- Data Integration from source systems into Operational Data Store-ODS.
- Worked with DBA’s to get the Database designed in Snowflake model to get approved and implemented.
- Define and Improve tasks to ensure Data Quality.
- Find performance bottle necks and provide industry standard solution.
- Standardize the process to the enterprise level.
- Closely interact with Vendors, business users and internal teams to make sure every process goes smoothly.
- Timely communication to the Project participants and management team about the progress.
- Improve the existing process towards more automated, reduce failure risks, improve data quality and increased resource utilization.
- Supported the Components testing during Teradata upgrade from 13.0 to 13.10
- Technologies Used: Data stage 8.1, IBM CDC, Teradata 13.10
Confidential, Newark, DE
Data Architect
Responsibilities:
- Gathering the Project Requirement
- Analyze changes and design database changes
- Providing best possible solution and getting it approved from Architectural Board
- Providing alternate solutions or approach to improve the existing process
- Providing Architectural Design to the development team
- Create Logical/Physical data model
- Coordination with the Datastage ETL & Reporting Development team with Design/Development approach to create a Data mart
- Working with DBA’s to get the Database design approved and implemented
- Created logical/physical data model using Erwin using Snowflake model
- Work on HLD/LLD Documents along with mapping rules
- Define and Improve tasks to ensure Data Quality
- Find performance bottle necks and provide industry standard solution
- Standardize the process to the enterprise level
- Timely communication to the Project participants and management team about the progress
- Improve the existing process towards more automated, reduce failure risks, improve data quality and increased resource utilization
- Technologies Used: Informatica 8.6.1, Data stage 8.1, Teradata, SQL Server, UNIX, Maestro.
Confidential, Newark, DE
Lead Developer
Responsibilities:
- Define roadmap for Informatica Upgrade
- Coordinate with Informatica Admin, Unix Admin, DBA’s and Upgrade team to finalize the roadmap and get it approved
- Weekly status reporting with the progress of the upgrade to management
- Identify any issues and provide support to the Upgrade team.
- Coordination with the Informatica support team for smooth upgrade.
- Technologies Used: Informatica 8.6.1, UNIX, Teradata.
Confidential, Mclean, VA
Data Architect
Responsibilities:
- Business users interaction
- Database Modeling using Erwin Tool involving Dimensional /Fact schemas using combination of start and snowflake.
- Create BRD’s
- Data Integration from source systems into Operational Data Store-ODS to create a Data mart
- Scope management and SOW submission
- Documenting the findings from POC
- Finalizing the best solution approach after approval from Technical Board.
- Status reporting to Stakeholders & PMO
- Schedule tracking
- Suggest best practices.
- Technologies Used: Informatica 8.6.1 with Web service, JMS messaging, Orchestration, Oracle, UNIX.
Confidential, Danbury, CT
Data Architect
Responsibilities:
- Requirements gathering, analysis and design
- Coordination with Business Users and Analyst to understand the requirement
- Database Design
- Creation of ETL Specs-Mapping of data from Source to Target with logic to pull data to create a data mart
- Data Integration from source systems into Operational Data Store-ODS
- Coordination and providing solution approach to Development team
- Informatica Consultation
- Data modeling using Erwin tool with Dimensional Modeling.
- Technologies Used: Informatica 8.1.1, Teradata, Business Objects, Erwin
Confidential
Informatica Developer
Responsibilities:
- Data Integration from source systems into ODS
- Worked as Module Leader for the project
- Requirements gathering
- Design and development of the ETL jobs
- Testing the jobs.
- Ensure the quality of the deliverable to the client.
- Technologies Used: Informatica 7.1, Oracle, UNIX.