We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume

0/5 (Submit Your Rating)

Sfo, CA

SUMMARY:

  • 7 years of professional experience in IT Industry that includes Hadoop Developing, Implementing, configuring, testing and maintenance.
  • 3+ years' experience in Hadoop Framework, and its ecosystem.
  • Development of standalone and web applications in multi - tiered environment using Java,Hadoop, Hive, HBase, Impala, Pig, Sqoop, J2EE Technologies (Spring, Hibernate), Oracle, HTML, Java Script
  • Extensive experience on BigData Analytics with hands on experience in writing MapReduce jobs on Hadoop Ecosystem including Hive and Pig
  • Excellent knowledge onHadoop architecture; as in HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm
  • The system consists of several applications, highly distributive, scalable and large in nature using ClouderaHadoop
  • BigData frame work and Eco system with evolving technologies
  • Experience with distributed systems, large scale non-relational data stores, MapReduce
  • Systems, data modeling, and big data systems
  • Experience on Apache, Cloudera, Hortonworks Hadoop distributions
  • Involved in developing solutions to analyze large data sets efficiently
  • Excellent hands on with importing and exporting data from different Relational Database Systems like Mysql and Oracle into HDFS and Hive and vice-versa, using Sqoop
  • Hands-on experience in writing Pig Latin scripts, working with grunt shells and job scheduling with Oozie
  • Experience in analyzing data using Hive QL, Pig Latin, and custom Map Reduce programs in Java
  • Experience with web-based UI development using jQuery, ExtJS, CSS, HTML, HTML5, XHTML and Java script
  • Knowledge of job workflow scheduling and monitoring tools like oozie and Zookeeper
  • Experience with databases like DB2, Oracle 9i, Oracle 10g, MySQL, SQL Server, lambda and MS Access
  • Experience in creating complex SQL Queries and SQL tuning, writing PL/SQL blocks like stored procedures, Functions, Cursors, Index, triggers and packages
  • Very good understanding on NOSQL databases like MongoDB and HBase
  • Have good Knowledge in ETL and hands on experience in Informatica ETL
  • Extensive experience in creating Class Diagrams, Activity Diagrams, Sequence Diagrams using Unified Modeling Language (UML)

WORK EXPERIENCE:

Confidential, SFO, CA

Sr. Hadoop Developer

Responsibilities:

  • Developed multiple MapReduce programs for analyzing the insurance data of the customer and produce summary results fromHadoopto downstream systems
  • Worked on importing and exporting data from Oracle and DB2 into HDFS using Sqoop
  • Developed data pipeline using Flume to ingest customer behavioral data and financial histories into HDFS for analysis
  • Prepared the best practices in writing map reduce programs and hive scripts
  • Scheduled a workflow to import the weekly transactions in the revenue department from RDBMS database using Oozie
  • Built wrapper shell scripts to hold these Oozie workflow
  • Developed PIG Latin scripts to transform the log data files and load into HDFS
  • Used Pig as ETL tool to do transformations, event joins and some pre-aggregations before storing the data onto HDFS
  • Hands on experience with NoSQL databases like Cassandra for POC (proof of concept) in storing URL's and images
  • Developed hive UDF for functions that were not preexisting in Hive like the rank etc.
  • Created External Hive tables and involved in data loading and writing Hive UDFs
  • Experienced in implementing POC's to migrate iterative map reduce programs into Spark transformations using Scala
  • Created concurrent access for hive tables with shared and exclusive locking that can be enabled in hive with the help of Zookeeper implementation in the cluster
  • Wrote the shell scripts to monitor the health check ofHadoopdaemon services and respond accordingly to any warning or failure conditions
  • Developed Unit test cases using MR unit for map reduce code
  • Involved in creating Hadoop streaming jobs
  • Involved in Installing, Configuring Hadoop Eco System, Cloudera Manager using CDH4 Distribution

Confidential, Tampa, FL

Hadoop Developer

Responsibilities:

  • Involved in design and development phases of Software Development Life Cycle (SDLC) using Scrum methodology
  • Developed data pipeline using Flume, Sqoop, Pig and MapReduce to ingest customer behavioral data and purchase histories into HDFS for analysis
  • Involved in creating Hive tables, loading with data and writing hive queries that will run internally in map reduce way
  • Used hive optimization techniques during joins and best practices in writing hive scripts.
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting
  • As a Hadoop Administration responsibilities include software installation, configuration, software upgrades, backup and recovery, commissioning and decommissioning data nodes, cluster setup, cluster performance and monitoring on daily basis, maintaining cluster on healthy on different Hadoop distributions (Hortonworks & Cloudera)
  • Experienced in managing and reviewing the Hadooplog files
  • Working with Apache Crunch library to write, test and run HADOOP MapReduce pipeline jobs
  • Involved in joining and data aggregation using Apache Crunch
  • Used Pig as ETL tool to do Transformations, even joins and some pre-aggregations before storing the data onto HDFS
  • Worked on Oozie workflow engine for job scheduling
  • Developed custom implementation for Partioner, Input / Output Formats, Record Reader and Writers
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting on the dashboard
  • Loaded the aggregated data onto DB2 for reporting on the dashboard
  • Project also involved building analytical reports using Sql server and Excel
  • Monitoring and Debugging Hadoopjobs/Applications running in production
  • Worked on Providing User support and application support on Hadoop Infrastructure
  • Reviewing ETL application use cases before on boarding to Hadoop
  • Worked on Evaluating, comparing different tools for test data management withHadoop
  • Helped and directed testing team to get up to speed onHadoop Application testing
  • Worked on Installing 20 node UAT Hadoopcluster
  • Created ETL jobs to generate and distribute reports from MySQL database using Pentaho Data Integration
  • Created ETL jobs using Pentaho Data Integration to handle the maintenance and processing of data

Confidential, Franklin Lakes, NJ

Java/J2EE Developer

Responsibilities:

  • Extensively used Java multi-threading to implement batch Jobs with JDK 1.5 features
  • Configured the project on Web Logic 10.3 application servers
  • Implemented the online application using Core Java, JDBC, JSP, Servlets, spring, Hibernate, Web Services, SOAP, and WSDL
  • Communicated with other Health Care info using Web Services with the help of SOAP, WSDL
  • Tested the web services with SOAP UI tool
  • Involved in gathering business requirements, analyzing the project and created UML diagrams such as Use Cases, Class Diagrams, Sequence Diagrams and flowcharts for the optimization Module using Microsoft Visio.
  • Designed and developed Optimization UI screens for Rate Structure, Operating Cost, Temperature and Predicted loads using JSF myfaces, JSP, JavaScript and HTML.
  • Configured faces-config.xml for the page navigation rules and created managed and backing beans for the Optimization module.
  • Developed JSP web pages for rate Structure and Operating cost using JSF HTML and JSF CORE tags library.
  • Implemented Singleton, Factory Design Pattern, DAO Design Patterns based on the application requirements

Confidential

Java/J2EE Developer

Responsibilities:

  • Used Microsoft Visio for designing the Use Case Diagrams, Class model, Sequence diagrams, and Activity diagrams for SDLC process of the application
  • Implemented GUI pages by using JSP, JSTL, HTML, DHTML, XHTML, CSS, JavaScript, AJAX
  • Configured the project on WebSphere 6.1 application servers
  • Implemented the online application by using Core Java, Jdbc, JSP, Servlets and EJB 1.1, Web Services, SOAP, WSDL
  • Communicated with other Health Care info by using Web Services with the help of SOAP, WSDL JAX-RPC
  • Implemented Singleton, factory design pattern, DAO Design Patterns based on the application requirements
  • Used SAX and DOM parsers to parse the raw XML documents
  • Used RAD as Development IDE for web applications
  • Used Clear Case for Version Control tool and ClearQuest for bug tracking tool
  • Configured job scheduling in Linux using shell scripts and Crontab
  • Developed test plan documents for all back end database modules

We'd love your feedback!