We provide IT Staff Augmentation Services!

Senior Snowflake Developer Resume

0/5 (Submit Your Rating)

Cincinati, OhiO

SUMMARY

  • Around 8 years of experience in architecting and implementing very large - scale data intelligence solutions around Snowflake Data Warehouse also solid experience and understanding of architecting, designing and operationalization of large scale data and analytics solutions on Snowflake Cloud Data Warehouse.
  • Developing ETL pipelines in and out of data warehouse using combination of Python and Snowflake’s Snow SQL writing SQL queries against Snowflake.
  • Strong experience in migrating other databases to Snowflake.
  • Work with domain experts, engineers, and other data scientists to develop, implement, and improve upon existing systems.
  • Experience with Snowflake Multi - Cluster Warehouses.
  • Understanding of Snowflake cloud technology.
  • Developed entire frontend and backend modules using Python on Django Web Framework.
  • Experience with Snowflake cloud data warehouse and AWS S3 bucket for integrating data from multiple source system which include loading nested JSON formatted data into snowflake table.
  • Understanding data pipelines and modern ways of automating data pipelines using cloud based testing.
  • Understanding of Snowflake Internals and integration of Snowflake with other data processing and reporting technologies.
  • Experience in Configuring Snowpipe.
  • Experience with Snowflake Snow SQL and writing use defined functions.
  • In-depth knowledge of Data Sharing in Snowflake.
  • Experience in using Snowflake Clone and Time Travel.
  • Professional knowledge of AWS Redshift.
  • Participates in the development, improvement and maintenance of Snowflake database applications.
  • Experience in various methodologies like Waterfall and Agile.
  • In-depth knowledge of Snowflake Database, Schema, and Table structures.
  • Experience in developing SQL scripts and stored procedures that processes data from databases.
  • Understanding of various data formats such as CSV, XML, JSON, etc.
  • Have good knowledge of RDBMS topics, ability to write complex SQL, PL/SQL.
  • Design, develop, test, implement and support of Data Warehousing ETL using Talend.
  • Experience working in Snowflake database implementations.
  • Define roles, privileges required to access different database objects.
  • Define virtual warehouse sizing for Snowflake for different type of workloads.
  • Worked with cloud architect to set up the environment.
  • Familiar with data visualization tools (Tableau/Power BI).
  • Designs batch cycle procedures on major projects using scripting and control.
  • Hands-on experience with Snowflake utilities, Snow SQL, SnowPipe, Big Data model techniques using Python/Java.
  • Extensive experience in developing complex stored procedures/BTEQ queries.
  • Performance tuning of Big Data Workloads.
  • Working experience in Snowflake, Data warehousing technologies (Hadoop-HDFS, Hive, SQOOP, SPARK, Scala, Python, JAVA, SQL, Impala, Datastage, Tableau).
  • Experience in Design, Implement, and maintain the reusable components for Extract, Transform, and Load (ETL) jobs.
  • Operationalize data ingestion, data transformation and data visualization for enterprise use.
  • Mentor and train junior team members and ensure coding standard is followed across the project.
  • Progressive experience in the field of Big Data Technologies, Software Programming and Developing, which also includes Design, Integration, Maintenance.
  • Managed Amazon Web Service(AWS) projects while coaching the agile process and help implement agile methodology.
  • ETL pipelines in and out of data warehouse using combination of Python and Snowflakes Snow SQL writing SQL queries against Snowflake.

TECHNICAL SKILLS

Cloud Technologies: Snowflake,, SnowSQL, Snowpipe AWS.

Spark, Hive: LLAP, Beeline, Hdfs, MapReduce, Pig, Sqoop, HBase, Oozie, Flume

Reporting Systems: Splunk

Hadoop Distributions: Cloudera, Hortonworks

Programming Languages: Scala, Python, Perl, Shell scripting.

Data Warehousing: Snowflake, Redshift, Teradata

DBMS: Oracle, SQL Server,MySql,Db2

Operating System: Windows, Linux, Solaris, Centos, OS X

IDEs: Eclipse, NetBeans.

Servers: Apache Tomcat

PROFESSIONAL EXPERIENCE

Confidential, Cincinati, Ohio

SENIOR SNOWFLAKE DEVELOPER

Responsibilities:

  • Understanding data transformation and translation requirements and which tools to leverage to get the job done
  • Expertise in data privacy, security, data streaming & consumption in multi cloud environment
  • Experienced in importing data from various sources usingStreamSets.
  • Experience implementing ETL pipelines using custom and packaged tools.
  • Extensive hands-on expertise with SQL and SQL analytics.
  • AWS EC2, EBS, Trusted Advisor,S3, Cloud Watch, Cloud Front, IAM, Security Groups, Auto-Scaling.
  • Design and Develop Informatica processes to extract the data from Juris Feed on a daily basis.
  • Understanding of complete data analytics stack and workflow, from ETL to data platform design to BI and analytics tools.
  • Used DataStage director to run the jobs manually.
  • Used the DataStage Director and its run-time engine to schedule the jobs, testing and debugging its components, and monitoring the resulting executable versions.
  • Extensively worked on Spark using Scala onclusterfor computational (analytics),application by making use of Spark with Hive and SQL/Oracle/Snowflake.
  • Wrote various data normalization jobs for new data ingested into Redshift
  • Advanced knowledge on Confidential Redshift and MPP database concepts.
  • Worked on Letter generated programs using C and UNIX shell scripting.
  • ConfiguredStreamsetsto store the converted data to SQL SERVER using JDBC drivers.
  • Oversee the migration of data from legacy systems to new solutions
  • Monitor the system performance by performing regular tests, troubleshooting and integrating new features
  • Implemented Change Data Capture technology in Talend in order to load deltas to a Data Warehouse.
  • Collaborate with IT Security to ensure necessary controls to Cloud services are deployed and tested
  • Experience with Agile and DevOps concepts
  • Wrote ETL jobs to read from web APIs using REST and HTTP calls and loaded into HDFS using java and Talend
  • Understanding and hands on experience of operating system, e.g. Unix/ Linux
  • Experience in PL/SQL development and performance tuning.
  • Worked on Oracle Databases, RedShift and Snowflakes

Confidential, New York, NY

SNOWFLAKE DEVELOPER

Responsibilities:

  • Driving replacing every other data platform technology using Snowflake with lowest TCO with no compromise on performance, quality and scalability.
  • Worked on SnowSQL and Snowpipe
  • Designed cloud based Datawarehouse using Snowflake architecture that includes data load using ELT tools like SnowPipe that being used for data analytics in dashboarding.
  • Developed end to end cloud-based data warehouse which includes loading data from various third party applications.
  • Created SnowPipes, databases, schemas, objects, UDFs and Shares using Snowflake.
  • Designing and working on proof of concept for integrating new bolt on applications.
  • Supporting and ensuring smooth deployment of releases.
  • Converted Talend Joblets to support the snowflake functionality.
  • Created data sharing between two snowflake accounts(Prod—Dev). Redesigned the Views in snowflake to increase the performance.
  • Unit tested the data between S3 and Snowflake.
  • Snowflake SQL Writing SQL queries against Snowflake Developing scripts Unix, Python etc
  • Extract, Load and Transform of experience in Data management (e.g. DW/BI) solutions.
  • Practicing consulting on Snowflake Data Platform Solution Architecture, Design, development and deployment focused to bring the data driven culture across the enterprises
  • Experience in various data ingestion patterns to hadoop.
  • Created data access modules in python.
  • Troubleshooting Datastage users login issues.
  • Used informatica to parse out the xml data into the datamart structures that is further utilized for the reporting needs.
  • InstalledData Stage clients /Serversand maintained metadata in repositories.
  • Used DataStage Manager for importing metadata from repository, new job categories and creating new data elements
  • Automated JIRA processes using Python and bash scripts.
  • To fetch data of definite options that are selected, Python routines were written to log into websites.
  • Automated AWS S3 data upload / download using Python scripts.
  • Worked on AWS Data Pipeline to configure data loads from S3 to into Redshift
  • Used JSON schema to define table and column mapping from S3 data to Redshift
  • Experience in Sqoop ingesting data from relational to hive.
  • Worked on data streaming tools and ETL tools likeStreamsetsData Collector, Attunity Replicate.
  • Strong hands-on onAWScloud services likeEC2, S3, RDS, ELB,andEBSfor installing, configuring
  • Heavily involved in testing Snowflake to understand best possible way to use the cloud resources.
  • Cloned Production data for code modifications and testing.
  • Used FLATTEN table function to produce lateral view of VARIENT, OBECT and ARRAY column.
  • Created Talend Mappings to populate the data into dimensions and fact tables.
  • Used Temporary and Transient tables on diff datasets.
  • Developed data warehouse model in snowflake for over 100 datasets using whereScape.
  • Experience in designingStar, Snowflakesschemas anddatabase modelingusing Erwin tool.
  • Converted the SQL Server Stored procedures and views to work with Snowflake.
  • Migrated Data from SQL Server 2016 to Snowflake using ELTMaestro.
  • Experience with Snowflake cloud data warehouse and AWS S3 bucket for integrating data from multiple source system which include loading nested JSON formatted data into snowflake table.
  • Performed bulk load of JSON data from s3 bucket to snowflake.
  • Converted Talend Joblets to support the snowflake functionality
  • Used Snowflake functions to perform semi structures data parsing entirely with SQL statements.

Confidential, Mountain View, CA

SENIOR DATA ENGINEER

Responsibilities:

  • Developed Spark jobs in Java to perform ETL from SQL Server to Hadoop.
  • Designed HBase schema based on data access patterns. Used Phoenix to add SQL layer on HBase table. Created indexes on Phoenix tables for optimization.
  • Integrated Spark with Drools Engine to implement business rule management system.
  • Benchmarked compression algorithms and file formats (Avro, ORC, Parquet, and Sequence File) for Hive, MapReduce, Spark, and HBase.
  • Analyzed Stored Procedures to convert business logic into Hadoop jobs
  • Used SSIS and SSRS for BI projects
  • Worked on various POCs including Apache NiFi as a data flow\orchestration engine, Talend as ESB and Big Data solutions, Elasticsearch as an indexing engine for MDM, and SMART on FHIR with MongoDB for FHIR based app interface.
  • Used Cassandra to build next generation storage platform. Employed CQL for data manipulation.

Confidential

SQL DEVELOPER

Responsibilities:

  • GeneratedSQL and PL/SQL scriptsto install create and drop database objects, including tables, views, primary keys, indexes, constraints, packages, sequences, grants and synonyms.
  • Database design and development includingTables, Primary and Foreign Keys, IndexesandStored Procedures.
  • Involved in performance tuning using withTKPROF.
  • Developed Shell scripts for job automation and daily backup.
  • Developed highly optimized stored procedures, functions, and database viewsto implement the business logic and also created clustered and non-clustered indexes.
  • Used various SSIS tasks such as Conditional Split, Multicast, Fuzzy Lookup etc., which did Data Scrubbing, including data validation checks during Staging, before loading the data into the Data warehouse.
  • Created SSIS package to load data from Flat Files, Excel and XML Files to Data warehouse and Report-Data mart using Lookup, Derived Columns, Sort, Aggregate, Pivot Transformation, and Slowly Changing Dimension.
  • Designed, deployed, and maintained complex canned reports using SQL Server 2008 Reporting Services (SSRS).
  • Created and maintained data flow diagrams for existing and future ETL processes.
  • Involved in Data Extraction, Transformation and Loading (ETL) between Homogenous and Heterogeneous systems using SQL tools (SSIS, Bulk Insert).
  • Design and Deploy SSIS Packages
  • Creating SSIS Packages, Configure SSIS packages for different environments using DTSConfig
  • Extract-Transform-Load data into SQL Server 2008 Normalized database from various sources
  • Involved in error handling of SSIS packages by evaluating error logs.
  • Extensively used high level backendPL/SQLprogramming involving different databases.

We'd love your feedback!