We provide IT Staff Augmentation Services!

Snowflake Data Engineer Resume

0/5 (Submit Your Rating)

Plano, TX

PROFESSIONAL SUMMARY:

  • Professional IT with over 6+ years of experience in AWS Cloud - Snowflake DBA/S3/ /DynamoDB, Teradata Enterprise Data Warehouse (EDW/ADW) and Data Mart.
  • A talented and resourceful professional with solid SQL, Oracle and Snowflake DBA experience to leverage solid skills in database, Unix/Linux, Python and shell scripts automation to work on a wide variety of snowflake database administrator tasks.
  • Working on Amazon and GCP cloud technologies that includes developing data pipelines using Snowflake, Redshift, Teradata, AWS Glue and Python technologies.
  • Expertise Snowflake utilities such as SnowSQL, SnowPipe, Tasks, Streams, Time travel, Optimizer, Metadata Manager, data sharing, and stored procedures.
  • Extensive experience in writing Complex SQL Queries, Stored Procedures, Views, Functions, Triggers, Indexes and Exception Handling using MS-SQL Server (TSQL).
  • Good experience in Extraction, Transformation and Loading (ETL) data from multiple database sources for medium to large Enterprise Data Warehousing.
  • Designing/selecting appropriate Snowflake table structure and preparing and implementing Snowflake related change requests/deployments with Snow console access management.
  • Experience in Snowflake Performance Tuning, WLM configuration, Snowflake Cluster Resizing, launching new instance, instance type change, new volume creation and existing volume expansion using AWS Snowflake console.
  • Hands-on expertise on Data Storage, Data Ingestion, Batch Processing, Stream Processing and Real-Time Message Ingestion on Data Engineering fields including Data Lake, Datawarehouse, Reporting and Analytics.
  • Hands on Experience on Python programming for data processing and to handle Data integration between On-prem and Cloud DB or Datawarehouse.
  • Solid experience in Dimensional Data modeling, Star Schema/Snowflake modeling, FACT & Dimensions tables, Physical & logical data modeling, Oracle Designer, Data Integrator.
  • Strong programming experience in developing SQL and PL/SQL codes through various stored Procedures, Functions, Packages, Cursors, Indexes, Views and Triggers to implement the business logics of databases.
  • Experience with Snowflake cloud data warehouse and AWS S3 bucket for integrating data from multiple source systems which include loading nested JSON formatted data into snowflake table.
  • Created business driven dashboards using different chart types like Crosstab, Geo/Tree/Heat Maps, Dual Axis Charts, Pie/Bar Charts, Line/Area Charts, Gantt charts, Bubble charts, Scatter Plots, Trend lines, Bullet Graphs and Histograms in Tableau Desktop.
  • Designed and implemented live monitoring Dashboards for Tableau Server and Visualization users.
  • Experience in working with Waterfall and Agile (Scrum) Software methodologies projects.
  • Developing solutions as per business requirements using streaming/Batch processing with extreme expertise in real time data streaming with AWS technologies.
  • Experience working with clients of insurance, financial, Retail, healthcare, manufacturing industries.
  • An excellent team player with a fast-learning curve, strong analytical, problem solving, communication and interpersonal skills to work under time constraints and has capability to work with business users, project managers, team leads, architects and peers, thus maintaining a healthy environment in the project.

TECHNICAL SKILLS:

DATABASE: AWS RDS, Aurora, MongoDB, MySQL, Teradata, Oracle 10g, DB2, SQL server

CLOUD/SaaS: AWS, GCP, Azure, Snowflake, SnowSQL, Redshift, S3, Snowpipe, Google Colab

LANGUAGES/SCRIPTS: Python, PL/SQL, Unix Shell Scripting, Spark, SQL, SnowSQL, HTML, XML, C, JavaScript, JSON/Yaml.

TOOLS / IDEs: Informatica, TALEND ETL, TABLEAU, Eclipse, NetBeans, PyCharm

AWS SERVICES: AWS EMR, S3 buckets, Data Pipeline, CloudWatch, Amazon SageMaker Studio, Cloud Formation, IAM, AWS Glue, AWS RDS, Aurora.

METHODOLOGIES: Agile, Scrum, Waterfall Methodologies, Business Process, Dimensional Data Modeling, Star Join Schema Modeling, Snow-Flake Modeling, Physical and Logical Database Design.

PROFESSIONAL EXPERIENCE:

Confidential, Plano, TX

Snowflake Data Engineer

Responsibilities:

  • Served as the Snowflake Database Administrator responsible for leading the data model design and database migration deployment production releases to endure our database objects and corresponding metadata were successfully implemented to the production platform environments; (Dev, Qual and Prod) AWS Cloud (Snowflake).
  • Implementing One time Data Migration of Multistate level data from SQL server to Snowflake by using Python and SnowSQL.
  • Day to-day responsibility includes developing ETL Pipelines in and out of data warehouses, developing major regulatory and financial reports using advanced SQL queries in snowflake.
  • On call support for production job failures and lead the effort on working with various teams to resolve the issues.
  • Performed day-to-day integration with the Database Administrators (DBA) DB2, SQL Server, Oracle and AWS Cloud teams to ensure the insertion of database tables, columns and its metadata have been successfully implemented out to the DEV, QUAL and PROD region environments in AWS Cloud and Snowflake.
  • Involved in different data migration activities.
  • Performed ETL data translation using informatica of functional requirements to Source to Target Data Mapping documents to support large datasets (Big Data) out to the AWS Cloud databases and Snowflake.
  • Created Snowpipe for continuous bulk loading and unloading data into Snowflake tables using COPY command.
  • Created DWH, Databases, Schemas, Tables, write SQL queries against Snowflake by validating the data feed from the source systems to Snowflake DW cloud platform.
  • Integrated and automated data workloads to Snowflake Warehouse by analyzing the SQL scripts and designed the solution to implement using PySpark.
  • Partner with Source teams to source the data for supporting data science models by providing assistance to business users for various reporting needs.
  • Worked on data ingestion from Oracle to hive and worked on data transfer mechanism from hive to Teradata.
  • Worked with different platform teams to resolve cross dependency and involved in Code Review Discussions, Demo’s to stakeholders.
  • Involved in fixing various issues related to data quality, data availability and data stability.
  • Worked in determining various strategies related to data security.
  • Performed logical and physical data structure designs and DDL generation to facilitate the implementation of database tables and columns out to the DB2, SQL Server, AWS Cloud (Snowflake) and Oracle DB schema environment.
  • Assisted Project Managers and Developers in performing ETL solution design and development to produce reporting, dashboarding and data analytics deliverables.

Environment: AWS (EC2, VPC, ELB, S3, RDS, DynamoDB, Cloud Trail Load balancer, cloud watch), CloudFormation Templates, AWS Glue Catalog, RedShift, Python, Jira, Jenkins, GIT, Bitbucket, Linux, Shell Scripting, PUTTY, CONFLUENCE

Confidential, Plano, TX

Snowflake Developer

Responsibilities:

  • Worked in implementing Data warehouse solutions in Snowflake and AWS Redshift and worked on various projects to migrate data from Teradata to Snowflake and AWS Redshift, DynamoDB and S3.
  • Handled Redshift performance SQL Tuning, Query optimization (Explain plans, Vacuum, Distribution Keys, Sort keys, Primary and Secondary indexes).
  • Developed Business logic in semantic layer by creating views in AWS Redshift to provide transformation logic visibility.
  • Create and load staging tables based on the logic defined in views using Distribution, Sort Keys for optimal performance.
  • Used JSON schema to define table and column mapping from S3 data to Redshift.
  • Worked on Oracle Databases, RedShift, Snowflakes, Snowflake Schemas and Data Warehousing.
  • Migrated the data from Redshift data warehouse to Snowflake.
  • Developed Talend Big Data jobs to load heavy volume of data into S3 data lake and then into Redshift data warehouse by involving in design of Snowflake Continuous data Load, Data share, Swap, Cloning Multi clustering of virtual warehouse.
  • Experienced on Monitor the Redshift system, and act upon system and application automated alerts Perform diagnostics, testing, and troubleshooting of software, o/s, and database issues.
  • Developed Talend Big Data jobs to load heavy volume of data into S3 data lake and then into Redshift data warehouse.
  • Involved in creating System Roles & Custom Roles and Role Hierarchy in snowflake.
  • Developed Talend jobs to populate the claims data to data warehouse - star schema, snowflake schema.
  • Coordinated with Technical Analysts in understanding the business requirements and translated them into technical specifications.
  • Analyzed the data as per business needs to build the Interface and generated server-side T-SQL scripts for data manipulation and validation.
  • Created Snowflake Resource monitoring and sending the alters.
  • Redesigned the Views in snowflake to increase the performance.
  • Snowflake Table Design Considerations and DML and Cloning Considerations.
  • Created GLUE catalog tables, crawlers and created connections from Snowflake.
  • Snowflake Technical for data structures, transformation, dependency, workload management and metadata by involving in data integration activities including designing, coding, and testing phases in Snowflake.
  • Creating and Managing Snowflake Virtual Warehouses, Scaling Virtual Warehouses Up & Out.
  • Created Views and Index Views to reduce database complexities for the end users.
  • Responsible for designing and maintaining ETL jobs that includes extracting the data from MS-SQL Server database and transforming the data as per Business requirements and then loading into Oracle database.
  • Created DataStage parallel jobs using different stages like Aggregator, Sort, Filter, Funnel, Join, Merge, Lookup, Remove Duplicate, Sequential File and Transformer.
  • Resolved technical challenges loading large data volume efficiently.
  • Designed the Dimensional Model of the Data Warehouse Confirmation of source data layouts and needs.
  • Worked on AWS Data Pipeline to configure data loads from S3 to into Redshift
  • Performance monitoring and Optimizing Indexes tasks by using Performance Monitor, SQL Profiler, Database Tuning Advisor and Index tuning wizard.
  • Extensively worked on workflow manager and workflow monitor to create, schedule, monitor workflows, work lets, sessions, tasks etc.

Confidential, Savannah, GA

ETL / Tableau Developer / Intern

Responsibilities:

  • Experienced in Teradata Manager, which is used to create Alerts, Monitor Systems, and see historical Reports.
  • Experienced in Designing Database with prominent activities like deciding Primary / Secondary index, creating Join index, creating Partitioned Primary Index, Compare Load Utilities to determine the best Scenario, using Automatic Data protection using Recovery Journals and Fallback protection.
  • Created/Build UNIX shell scripts to pull data from vendors and dropped into Informatica environment using FTP process.
  • Teradata system, Priority Scheduler, in controlling the load of the system, where we define performance Group, Allocation groups, performance periods, Resource Partition.
  • Worked on data transfer mechanism from hive to Teradata.
  • Extensively worked on ETL performance tuning for tuning the data load, worked with DBAs for SQL query tuning etc.
  • Developed PL/SQL Stored Procedures, Views and Triggers to implement complex business logics to extract, cleanse the data, transform and load the data into the Oracle database.
  • Followed agile methodology.
  • Providing daily status to the Client about the assignment progress.
  • Created Dashboards and custom visualizations using Tableau.
  • Created DAX columns, Measures, Views, Roles and Implemented user Level security.
  • Created complex SQL Queries, Stored Procedures, Functions, Views in MS SQL Server to accomplish some of the business logic.
  • Responsible for the data mapping, ETL data structures and analysis for the traceability of legacy database data entities structures to system capabilities and business operational activities that align to the data business rules which drives the data model design development.
  • Designed ETL process using Talend Tool to load from Sources to Targets through data Transformations.
  • Talend ETL is used for MDM Designed and developed the Business Rules and workflow system.
  • Expertise with MDM, Dimensional Modeling, Data Lake & Data Governance.
  • Working on generating various dashboards in Tableau Desktop using various sources such as SQL Server/Oracle/Excel/Text data etc.
  • Involved in creating a dual-axis bar chart with multiple measures. Preparing Dashboards using calculations, parameters in Tableau. Involved in creating Ad-hoc reports.
  • Developed Tableau data visualization using Heat maps, Box and Whisker charts, Scatter Plots, Geographic Map, Cross tabs, Histograms etc.
  • Responsible for migrating the folders, mappings and sessions from development to test environment and Created Migration Documents to move the code from one Environment to the other Environment.
  • Part of the team to design a common snap logic pipeline to load all tables by reading metadata dynamically based on each subject area.
  • Collect Load Stats of Start Time, End Time, Total Records loaded and Notify production support team with load details.

We'd love your feedback!