We provide IT Staff Augmentation Services!

Redshift Lead Resume

3.00/5 (Submit Your Rating)

PA

SUMMARY

  • 8+ Years of Experience in analysis, design, development, implementation and troubleshooting of Data Warehouse applications.
  • Experience working in Data warehousing, Business Intelligence systems,OLAP environments.
  • Experiencing in Designing and developing a new Redshift data warehouse.
  • Good knowledge on AWS key services like Amazon Redshift, Amazon S3, DMS, Athena, Glue, Kinesis, EMR, SNS, Amazon EC2, Data Pipeline, Amazon Lambda, Amazon CloudWatch and Amazon Glacier.
  • Good experience in performing on - prem migration to AWS cloud
  • Designed, developed and implemented datalake on S3
  • Experience in identifying and utilizing AWS services wherever needed
  • Good experience in writing complex queries on the tables that contains billions of records using Redshift SQL to transform and load data to Redshift database.
  • Experience in utilizing different design and coding techniques to improve the performance of the queries on Redshift Database.
  • Experience in working on Redshift cluster and taking the snapshots on it and using those snapshots for restoring the data.
  • Experience on working with Amazon management console, to setup and manage instances through Amazon EC2 console, and manage the clusters and snapshots using Amazon Redshift console.
  • Experience in implementing performance optimization techniques.
  • Designed and developed ETL to load data from various sources like DB2, Oracle, Flat files and MS SQL Server into Amazon Redshift, Oracle, Netezza and SQL server targets.
  • Experience in creating Python scripts to extract and load data from flat files like CSV, JSON to Redshift/S3 and also moving data from different applications using APIs
  • Experience working on Informatica Cloud to extract and load data to Redshift and/or S3.
  • Extensively worked on PowerCenter Client tools - Designer, Repository manager, Workflow manager, Workflow monitor.
  • Good experience in creation of mappings, sessions, mapplets, tasks and workflows using Power Center tools to move data from multiple sources into target areas.
  • Experience in working with BI reporting tools such as Tableau, Business Objects and Crystal reports.
  • Hands on experience in identifying and resolving performance bottlenecks.
  • Worked on OLAP databases with different kinds of tables like dimensional, historical and aggregator and on OLTP databases.
  • Experience in SQL programming using tools such as Aginity, Toad, SQL Plus and SQL Developer.
  • Good in UNIX Shell scripting. Developed UNIX scripts using PMCMD utility and scheduled ETL load using utilities like CRON tab, Control-M, Automic(UC4) and Autosys.
  • Experience working on Github and PVCS version Manager to host and maintain the project related work.
  • Involved in all phases of data warehouse project life cycle with good exposure to Development, Testing, Debugging, Implementation, Documentation, End-user training and Production support.
  • Good Knowledge on Normalization and De-normalization concepts and design methodologies like Ralph Kimball and Bill Inmon’s Data Warehouse methodology.
  • Knowledge on execution of jobs in Mainframe and understandable knowledge of JCL code.
  • Performed Unit Testing, Integration Testing, System Testing and Data Validation during various stages of the project.
  • Documented design procedures, mapping specifications, operating instructions, test plans/procedures and troubleshooting procedure for ease of application maintenance.
  • Developed effective working relationships with client team to understand support requirements, develop tactical and strategic plans to implement technology solutions, and effectively manage client expectation.
  • Worked in offshore and on-site model, and acted as an on-site coordinator for many of the projects and directly interacted with the Client and the Business.
  • An excellent team member with an ability to perform individually and as team member with good interpersonal relations, strong communication skills, hardworking and high level of motivation.

TECHNICAL SKILLS

AWS Services: Amazon Management Console, AWS CLI, Amazon Redshift, Amazon S3, Amazon DMS, Amazon Kinesis, Amazon Athena, Amazon EMR, AWS Glue, Amazon EC2, Amazon SNS, Amazon Data Pipeline, Amazon Lambda, Amazon Glacier, Amazon Cloudwatch, Amazon IAM, Machine Learning

Data warehousing Tools: SQL Workbench SQL*Plus, Toad, SQL Navigator, Aginity for Netezza and Redshift, Informatica Power Center 9.x/8.x/7.x, Informatica Power Exchange, Informatica Cloud, Tableau, Crystal reports

Databases: Oracle11g/10g/9i, DB2, MSSQLServer2012/2008R2/2005, Amazon RedShift, Netezza 7.0.4, MS Access 2007/2003

Data Modeling: Star Schema, SnowFlake Schema, Kimball Methodology, Bill Inmon’s Data Warehouse methodology

Programming: SQL, PL/SQL, Redshift SQL, Unix Shell Scripting, Python.

Environment: UNIX (SunSolaris10, AIX 6.1, Linux 6.5), Windows 2008/07/XP

PROFESSIONAL EXPERIENCE

Confidential, PA

Redshift Lead

Responsibilities

  • Performed POCs and implemented the solutions to move Confidential environment from on-prem to AWS cloud.
  • Created Data lake on S3 which acts as central repository for all Confidential related data
  • Implemented various ETL and ELT flows to move data from different sources to S3 and Redshift
  • Used DMS and Schema Conversion Tool to move data from data sources like Oracle and SQL server to S3
  • Created, configured the Redshift scalable clusters that supports the data warehouse and created databases on the clusters
  • Create table structures with Distribution Keys, Sort Keys and encoding techniques to improve the performance for data retrieval on AWS Redshift.
  • Designed the tables and loaded the data to the Redshift databases using Amazon best practices.
  • Designed and developed tables in the Data warehouse based on Star Schema Model.
  • Implemented performance enhancing techniques on Redshift tables to ensure the jobs complete in schedule time.
  • Created AWS Data pipeline jobs to perform transformations on S3 data and load it to Redshift and/or S3
  • Implemented mechanism to convert CSV files to Parquet on S3 bucket to reduce space, improve data retrieval speed and decrease cost while retrieving data using AWS Athena
  • Created and used EMR clusters to implement data conversions from CSV to Parquet and automated them using data pipeline
  • Created Python scripts to pull marketing and campaign data from Eloqua application.
  • Created Python scripts, APIs and AWS Kinesis to pull data from Twitter, Facebook and Google to S3
  • Created Cloud Watch dashboards for Events, Alarms and SNS notifications
  • Created External tables on Athena and Redshift spectrum for the users to query on S3 data
  • Worked on AWS Glue to perform some ETL operations
  • Implemented Version control, Life Cycle Management and archival on S3 buckets
  • Created and streamlined IAM polices to enable complete control on AWS for the users
  • Working on POC on how AWS Machine Learning can provide predictions based on available data
  • Converted Informatica mappings to SQL scripts and used them in Data pipeline jobs to do the transformations
  • Used synchronization tasks and mappings on Informatica Cloud to move data from Salesforce to S3
  • Prepared Master Document that shows complete flow of data and tools and services that are used as part of this data flow
  • Presented POC sessions and provided knowledge transfer sessions on AWS services to the Client

Environment: Amazon Management Console, Amazon Redshift, Amazon DMS, Amazon Athena, EMR, Kinesis, Glue, Amazon EC2, AWS Athena, Amazon S3, SNS, Amazon Glacier, Amazon Lambda, Amazon Data Pipeline, AWS Machine Learning, Aginity workbench for Redshift, SQL Workbench for Redshift, Informatica Cloud, Informatica Power Center, Python, Power BI, Linux 6.5

Confidential Partners, Atlanta, GA

BI Redshift Developer

Responsibilities at CCEP:

  • Working and interacting with business to gather all the requirements for the development activities.
  • Designing the tables and loading the data to the Redshift data warehouse using Amazon best practices.
  • Designed and developing tables in the Data warehouse based on Star Schema Model.
  • Creating, configuring and supporting the Redshift scalable clusters that supports the data warehouse.
  • Working on Amazon management Console to manage the services like Amazon Redshift, Amazon S3, SNS, Data Pipeline, Glacier and Lambda.
  • Create table structures with Distribution Keys, Sort Keys and encoding techniques to improve the performance for data retrieval.
  • Creating new process flow using Redshift SQL scripts to extract transform and load the data from different sources to Amazon Redshift database.
  • Configuring, modifying and managing the pre-prod clusters of Amazon Redshift database as per the management requests.
  • Taking the snapshots and using those snapshots for restoring the data and creating other clusters.
  • Resolving the performance issues using various performance tuning techniques on the Redshift Database.
  • Creating Data pipeline process using python to load data from S3 to Redshift and transfer of files between different applications.
  • Created Tables and Views in Redshift that involves complex logic and are used for reporting purpose of the business.
  • Using commands like unload and copy to move data between different Redshift Databases and used table STL LOAD ERRORS to fix the data copy issues.
  • Implementing performance enhancing techniques on Redshift tables to ensure the jobs complete in schedule time.
  • Setting up Events, Alarms and SNS notifications and responding to them when required.
  • Worked on SNS to create and publish topics and creating subscription for various users and groups to those topics.
  • Creating AWS Data Pipeline jobs to process data between S3 and Redshift and execute different sql statements as per the requirements.
  • Archiving of old data and files to Amazon Glacier and retrieving when necessary.
  • Created Email notifications using Amazon Lambda and Python, that will invoke SNS and send email to end users
  • Creating scripts to create new tables, views, queries for new enhancement in the application using SQL Workbench and Aginity Workbench for Redshift.
  • Working on Informatica Cloud, to develop mappings and tasks to transfer data from sources like Salesforce, Flat Files to Redshift Database and S3.
  • Re-coded the existing SQLs in to speed up the run time that helped business to generate reports much faster than before.
  • Work with Business users to create workbooks with worksheets, interactive dashboards related to Forecast, Budget, actuals and daily sales data in Tableau.
  • Worked in Tableau environment to create dashboards like monthly, daily reports using tableau desktop & publish them to server.
  • Used filters, drilldowns, parameters, blending to make the views work more efficiently.
  • Created Custom Hierarchies to meet the Business requirement in Tableau.
  • Enhance and Publish dashboards developed by Business groups to desktops of the intended users.
  • Working with business in Navigating to resolve data issues faced by the business on Tableau reports and dashboards.
  • Performing testing at different phases in the life cycle of the project.
  • Creating Test documents with Test queries to perform the testing during pre-production and after the deployment.
  • Working with the production team in deploying the code to production and execution of jobs.
  • Providing support to Advanced Analytics Support whenever they have issues and required help from us.

Environment: Amazon Management Console, Amazon Redshift, Amazon EC2, Amazon S3, SNS, Amazon Glacier, Amazon Lambda, Amazon Data Pipeline, Aginity workbench for Redshift, SQL Workbench for Redshift, Informatica Cloud, Python, Tableau 10, Linux 6.5

Confidential, Bethpage, NY

BI Redshift Developer

Responsibilities at Confidential (Cable Vision):

  • Gathering project requirements and interacting with Business Users in the design of technical specification documents.
  • Worked with Business and other teams to resolve all the bottlenecks and free up the process of data migration from Netezza to Redshift.
  • Converting the existing Informatica mappings to PostgreSQL queries for the data migration from Oracle and Netezza to Amazon Redshift.
  • Experience in creating tables structures with Distribution Keys, Sort Keys and encoding to improve the performance for data retrieval.
  • Creating new process flow using PostgreSQL scripts to extract transform and load the data from different databases to Amazon Redshift database.
  • As Redshift does not support Stored Procedures, converted multiple stored procedures from Oracle and Netezza to Redshift understandable SQL queries.
  • Created shell scripts on Linux environment to execute the PostgreSQL scripts that were created.
  • Using commands like unload and copy to transfer data between different Redshift Databases and used table, STL LOAD ERRORS to fix the data copy issues.
  • Working on S3 bowser, which is an interface that connects to Amazon S3 bucket during the data transfer between different databases.
  • Used PSQL in the SQL scripts to execute the sql queries from the Linux servers.
  • Re-coded the existing SQLs in Redshift to speed up the run time that helped to meet SLAs.
  • PerformedSQL and PL/SQL tuningand Application tuning using EXPLAIN PLAN.
  • Used DTT scripts on Linux environment to run the queries and load the data to Redshift using Amazon Web Services (AWS).
  • Experience in working on Redshift cluster and taking the snapshots on it and using those snapshots for restoring the data.
  • Created new scheduling jobs in Pre-Prod environments using the scheduling tool AUTOMIC (UC4).
  • Creating scripts to create new tables, views, queries for new enhancement in the application using Aginity Workbench for Redshift.
  • Developing various Mappings, Mapplets and Transformations as per the business needs.
  • Re-designing ETL mappings to improve data quality when required.
  • Performing testing at different phases in the life cycle of the project.
  • Creating Test documents with Test queries to perform the testing during pre-production and after the deployment.
  • Worked on PVCS to move the code from Pre-Prod to Production environment.
  • Working with the production team in deploying the code to production and execution of jobs.
  • Providing support during the production run and whenever needed.
  • Co-coordinating with offshore team all through the project and making sure the project is on track.

Environment: Linux 6.5, Informatica Power Center 9.6, Power exchange 9.6, Oracle 12, Netezza, Amazon

Confidential, Chicago, IL

Informatica Developer

Responsibilities:

  • Worked as a Business analyst and Informatica developer, responsible in gathering Requirements and IT review. Interacted with Business Users in the design of technical specification documents.
  • Involved in creating logical and physical data models using MS Visio based on business requirements.
  • Worked on data warehouses and datamarts, which are built on Kimball Methodology and uses star and snowflake schema.
  • Worked on many Business need requests (BNR) and on many small and large projects.
  • Agile approach with daily scrum meetings is being followed throughput the development and deployment cycle.
  • Designed and Implemented Tables, Functions, Stored Procedures and Triggers.
  • Created scripts to create new tables, views, queries for new enhancement in the application using TOAD.
  • Loaded the tables with data from different Heterogeneous databases.
  • Created indexes on the tables for faster retrieval of the data to enhance database performance.
  • Extracted data from Heterogeneous source systems like Oracle, SQL Server and flat files into the Staging tables and then to the target databases Oracle and SQL Server.
  • Developed various Mappings, Mapplets and Transformations for datamart and data warehouse.
  • Re-designed ETL mappings to improve data quality.
  • Used Unix Shell Scripting to create scripts for automating the loading process.
  • Created Stored procedure transformations to populate targets based on business requirements.
  • Responsible for monitoring all the sessions that are running, scheduled, completed and failed. Debugged the mapping of the failed session.
  • Used Pipeline Partitioning feature in the sessions to reduce the load time.
  • Analyzed and Created Facts and Dimension Tables.
  • Used Informatica features to implement Type I, II, and III changes in slowly changing dimension tables.
  • Created Data Breakpoints & Error Breakpoints for debugging the mappings using Debugger Wizard.
  • Mapplet’s & Reusable Transformations were used to prevent redundancy of transformation usage and modularity.
  • Developed the pre & post session shell scripts, which will create the parameter file dynamically.
  • Tuned performance of Informatica session for large data files by increasing block size, data cache size, sequence buffer length and target based commit interval.
  • Created and monitored Database maintenance plans for checking database integrity, data optimization, rebuilding indexes and updating statistics.
  • Extensive use of Persistent cache to reduce session processing time
  • Written documentation to describe program development, logic, coding, testing, changes and corrections.
  • Worked on Control M, by creating the jobs and executing the scripts using them
  • Provided support in uploading the Control M jobs and managing the scheduling or making changes whenever necessary.
  • Created shell scripts to execute the SQL scripts created on AIX servers and to automate the flow of execution or to FTP the files wherever required.
  • Involved in Unit and Integration testing of Mappings and sessions.
  • Assisted Testing team in creating test plan and test cases.
  • Worked with offshore team, acted as an onsite coordinator for many of the projects, and was answerable to the Client.
  • Involved in creating multi-dimensional universe and reports in Business Objects Environment.
  • Designed and developed ad-hoc and customized reports using Crystal reports 2008.

Environment: Informatica Power Center/Exchange 9.1/8.6, Oracle 11g, MS SQL Server 2008, MS Visio, Toad, SQL plus, Business objects XiR3, Crystal Reports 2008, Control-M, BMC Remedy Tool.

We'd love your feedback!