We provide IT Staff Augmentation Services!

Hadoop Developer Resume

2.00/5 (Submit Your Rating)

New York, NY

SUMMARY:

  • 8+ years of overall experience in IT industry, which includes hands on experience in Big data eco - system related technologies like Map Reduce, Hive, HBase, Pig, SQOOP, Oozie, flume, Zookeeper and HDFS.
  • 3 years of comprehensive experience as a Big Data & Analytics (Hadoop).
  • Very Strong Object-oriented concepts with complete software development life cycle experience - Requirements gathering, Conceptual Design, Analysis, Detail design, Development, System and User Acceptance Testing.
  • Experience in working with MapReduce programs using Apache Hadoop for working with Big Data.
  • Hands on experience in big data ingestion tools like Flume and SQOOP.
  • Integrated the HIVE and HBase to access the hive data randomly using NOSQL database from UI.
  • Strong knowledge of Pig and Hive's functions, extending Hive and Pig core functionality by writing UDFs.
  • Proficiency in different Hadoop distributions Cloudera and Hortonworks Distributions.
  • Experience in analyzing data using HiveQL, Pig Latin, Impala and custom Map Reduce programs in Java.
  • Hands on NoSQL database experience with Apache HBase& Mongo DB.
  • Knowledge in job workflow scheduling and coordinating tools like Oozie and Zookeeper.
  • Experience with various scripting languages like Linux/Unix shell scripts, Python 2.7 and Scala.
  • Good Knowledge on Apache Spark & Scala.
  • Good experience on EMS, RV, JDK 5, JDBC, Servlets, JSP, Struts 2.1, Spring 2, AJAX, DOJO, JSON Hibernate 3, XML Processing (JAXB), JMS, and TIBCO
  • Strong front-end UI development skills using scripting languages like JSP, HTML, JavaScript, jQuery and CSS.
  • Good experience on application server WebSphere 6.0 and JBoss 4.x.
  • Experience to build tool like Ant.
  • Good Knowledge on life cycle of ETL (Extraction, Transformation and Loading) using Informatica Power Center (Repository Manager, Mapping Designer, Workflow Manager, Workflow monitor).
  • Good database experience using SQL Server, Stored Procedures, Cursors, Constraints and Triggers.
  • Highly motivated, adaptive and quick learner.
  • Domain Knowledge: Credit Cards domain experience with expertise in Non-Monetary & Monetary process, Payments, Fees, Collections, Statements functionalities, Banking, SMS Alerts & Email gateway solutions, IVR, Cards Origination, GL Accounting and middleware integration.
  • Strong Analytical and Problem-Solving skills, Multi-Tasking abilities, with proven experience in utilizing people and process knowledge to assist enterprises in making critical decision.

PROFESSIONAL EXPERIENCE:

Hadoop Developer

Confidential, New York,NY

Responsibilities:

  • Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Installed and configured Pig and also written Pig Latin scripts.
  • Developing Scripts and Batch Job to schedule various Hadoop Program.
  • Written Hive queries for data analysis to meet the business requirements.
  • Creating Hive tables and working on them using Hive QL. Importing and exporting data into HDFS from Oracle Database and vice versa using Sqoop.
  • Implemented test scripts to support test driven development and continuous integration.
  • Responsible to manage data coming from different sources.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Experience in managing and reviewing Hadoop log files.
  • Worked on Hive for exposing data for further analysis and for generating transforming files from different analytical formats to text files.
  • Wrote MapReduce job using Pig Latin.
  • Have solid understanding of REST architecture style and its application to well performing web sites for global usage.
  • Involved in ETL, Data Integration and Migration. Imported data using Sqoop to load data from Oracle to HDFS on regular basis.
  • Managing and scheduling Jobs on a Hadoop cluster.
  • Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Involved in creating Hive tables, loading with data and writing hive queries, which will run internally in map, reduce way.
  • Used Pig as ETL tool to do transformations, event joins, filter bot traffic and some pre-aggregations before storing the data onto HDFS.
  • Written Hive queries for data analysis to meet the business requirements.
  • Involved in writing Hive scripts to extract, transform and load the data into Database.
  • Used JIRA for bug tracking.
  • Used CVS for version control.

Environment: Hadoop, Hive, Linux, MapReduce, HDFS, Hive, Pig, Sqoop, Shell Scripting, Java (JDK 1.6), Java 6, Eclipse, Oracle 10g, PL/SQL, SQL*PLUS, Toad 9.6, Linux, JIRA 5.1, CVS, JIRA 5.2.

Hadoop Developer

Confidential,Philadelphia,PA

Responsibilities:

  • Worked extensively with Amazon Web Services.
  • Created Amazon Elastic MapReduce cluster in 1.0.3.
  • Developed Pig program for loading and filtering the streaming data into HDFS.
  • Imported data from Oracle database to HDFS using Sqoop.
  • Worked on Data cleansing using apache Avro schema and implementing it in Pig.
  • Developed Hive scripts to de-normalize and aggregate the disparate data.
  • Automated workflows using shell scripts and Oozie jobs to pull data from various databases into Hadoop.
  • Implemented external tables and dynamic partitions using Hive.
  • Worked on custom Pig Loaders and Storage classes to work with a variety of data formats such as JSON, Compressed CSV, etc.
  • Loaded the created HFiles into HBase for faster access of large customer base without taking performance hit.
  • Used Oozie scheduler system to automate the pipeline workflow.
  • Actively participated in software development lifecycle (scope, design, implement, deploy, test), including design and code reviews.
  • Implemented data serialization using apache Avro.
  • Involved in story-driven agile development methodology and actively participated in daily scrum meetings.
  • Worked on reporting using Microstrategy.

Environment: Hortonworks, Hadoop, MapReduce, HDFS, Hive, Sqoop, Avro, Oozie, Java (jdk1.6), Informatica, Oracle 4.0.8.

Hadoop Developer

Confidential,Livingston, NJ

Responsibilities:

  • Installed, implemented and administered Hadoop & Hive cluster Setup Flume to stream application web log and twitter data to HDFS
  • Parsed JSON data using Apache Pig Latin scripts Developed Pig Latin scripts to parse JSON data and create analytics reports
  • Processed data in form of tuples, bags and char arrays Used Secondary Sort feature to tune Map Reduce applications
  • Configured Flume agent with Flume receive the data from twitter Auto Populate HDFS with data coming from Flume sink
  • Developed Map/Reduce Jobs of various complexity using Hive and Pig Analyzed and troubleshot Hadoop logs
  • Used Secondary Sort feature to tune MapReduce applications Sqoop incremental loads, import and export between Oracle and HDFS
  • Worked on Sqoop to populate data from Oracle RDBMS to HDFS Transformation of data within Hadoop framework.

Environment: Hadoop Map Reduce, Hive QL, Pig Latin, ETL, Flume, Administration, Shell Script.

Hadoop Developer

Confidential,Charlotte, NC

Responsibilities:

  • Worked on Hortonworks (HDP 2.X) platform to perform Hadoop operations.
  • Installed and configured Hive, Pig, Sqoop and Flume on the Hadoop cluster.
  • Developed Simple to complex Map/reduce Jobs using Hive and Pig.
  • The logs and semi structured content that are stored on HDFS were preprocessed using PIG and the processed data is imported into Hive warehouse which enabled business analysts to write Hive queries.
  • Created Hive tables to store the processed results in a tabular format.
  • Developed Hive Scripts for implementing dynamic Partitions.
  • Developed Pig scripts for data analysis and extended its functionality by developing custom UDF's.
  • Extensive knowledge on PIG scripts using bags and tuples.
  • Experience in managing and reviewing Hadoop log files.
  • Developed workflow in middleware to automate the tasks of loading the data into HDFS and pre-processing with PigInvolved in loading data from UNIX/LINUX file system to HDFS.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, and loaded data into HDFS.
  • Continuous monitoring and managing the Hadoop cluster through Cloudera Manager.
  • Installed and configured Hadoop MapReduce, HDFS, developed multiple Map Reduce jobs in java for data cleaning and processing.
  • Importing and exporting data into HDFS, Pig, Hive and HBase using SQOOP.
  • Injected the data from logs and relational databases using Flume and SQOOP.
  • Involved in gathering the requirements, designing, development and testing.
  • Worked on loading and transformation of large sets of structured, semi structured data into Hadoop system.
  • Worked on Hue interface for querying the data.
  • Exported analyzed data to relational databases using SQOOP for visualization to generate reports for the BI team.

Environment: Hadoop (HDP 2.X), UNIX, Eclipse, TIBCO, HDFS, Java, MapReduce, Apache Pig, Hive, Flume, Python 2.7, SQOOP and MySQL.

Java Developer

Confidential,IN

Responsibilities:

  • Developed Servlets and back-end Java classes using Web Sphere application server.
  • Developed an API to write XML documents from a database.
  • Performed usability testing for the application using JUnit Test.
  • Maintenance of a Java GUI application using JFC/Swing.
  • Created complex SQL and used JDBC connectivity to access the database.
  • Involved in the design and coding of the data capture templates, presentation and component templates.
  • Part of the team that designed, customized and implemented metadata search and database synchronization.
  • Developed JavaScript behavior code for user interaction.
  • Created database program in SQL server to manipulate data accumulated by internet transactions.
  • Wrote Servlets class to generate dynamic HTML pages.
  • Used Oracle as Database and used Toad for queries execution and also Involved in writing SQL scripts, PL SQL code for procedures and functions.

Environment: Java, Web Sphere 3.5, EJB, Servlets, JavaScript, JDBC, SQL, JUnit, Eclipse IDE and Apache Tomcat 6.

SQL Server/Oracle Developer

Confidential,IN

Responsibilities:

  • Involved in various stages of Enhancements in the Application by doing the required analysis, development, and testing.
  • Prepared the High and Low-level design document and Generating Digital Signature
  • For analysis and design of application created Use Cases, Class and Sequence Diagrams.
  • For the registration and validation of the enrolling customer developed logic and code.
  • Developed web-based user interfaces using struts frame work.
  • Handled Client-side Validations used JavaScript and • Involved in integration of various Struts actions in the framework.
  • Used Validation Framework for Server-side Validations
  • Created test cases for the Unit and Integration testing.
  • Front-end was integrated with Oracle database using JDBC API through JDBC-ODBC Bridge driver at server side.

Environment: Java Servlets, JSP, Java Script, XML, HTML, UML, Apache Tomcat, JDBC, Oracle, SQL.

We'd love your feedback!