We provide IT Staff Augmentation Services!

Senior Hadoop Developer Resume

5.00/5 (Submit Your Rating)

Sunnyvale, CA

SUMMARY

  • 7+ years of working knowledge in Hadoop and its stack like HDFS, Map Reduce, Apache Pig, Hive, Hbase and Sqoop.
  • 4+ years of experience in Web development using Java.
  • Extensive knowledge in Hadoop eco - system technologies like Apache Pig, Apache Hive, Apache Scoop, Apache Flume and Apache Hbase.
  • Experience in UNIX shell scripting.
  • Experiences in analyzing data using HiveQL and Pig latin and custom MapReduce programs in Java.
  • Experienced in writing Pig UDFs and Hive UDFs and UDAFs in the analysis of data.
  • Worked with Hbase which is a NoSQL database.
  • Experienced in importing and exporting data from relational database into HDFS using Scoop.
  • Experienced in configuring flume to stream data into HDFS.
  • Good knowledge of Hadoop architecture and its components like HDFS, MapReduce, Job Tracker, Task Tracker, Name Node and Data Node.
  • Hands on experience in Application Development using Java, Hadoop, RDBMS and Linux shell scripting.
  • Excellent knowledge in Java and SQL in application development and deployment.
  • Good knowledge on relational databases like MySQL, Oracle and NoSQL databases like Hbase.
  • Experience in Extraction, Transformation and Loading (ETL) of data from multiple sources.
  • Basic knowledge in Apache spark for fast large scale in memory MapReduce.
  • Experience in using open source projects like Apache commons, ANT, Log4J and Junit to increase the productivity and software quality.
  • Diverse experience in utilizing Java tools in business, web and client server environments including Java platform, JSP, Servlet, Java beans, JSTL, JSP custom tags, EL, JSF and JDBC.
  • Expertise in developing the presentation layer components like HTML, CSS, JavaScript, JQuery, XML, XSLT, XSD, JSON and AJAX.
  • Experience in source control repositories like SVN and Git.

TECHNICAL SKILLS

Big Data Ecosystem: HDFS, HBase, Hadoop MapReduce, Zookeeper, HivePig, Sqoop, Flume, Oozie, Cassandra

Languages: C, C++, Java, PHP, SQL/PLSQL

Methodologies: Agile, waterfall.

Database: Oracle 10g, DB2, MySQL, MongoDB, CouchDB, MS

SQL server, Amazon EC2:

Web Tools: HTML, Java Script, XML, ODBC, JDBC, Java BeansEJB, MVC, Ajax, JSP, Servlets, Java Mail, Struts, Junit

IDE / Testing Tool: Eclipse.

Operating Systems: Windows, UNIX, Linux

Scripts: JavaScript, Shell Scripting

PROFESSIONAL EXPERIENCE

Confidential, Sunnyvale, CA

Senior Hadoop Developer

Responsibilities:

  • Used Scoop to dump data from relational database into HDFS for processing.
  • Configured flume to capture the news from various sources for testing the classifier.
  • Wrote extensive MapReduce Jobs in Java to train the classifier.
  • Wrote MR jobs using various Input and Output formats. Also used custom formats whenever necessary.
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing, analyzing and training the classifier using MapReduce jobs, Pig jobs and Hive jobs.
  • Used OpenNLP in the removal of stop words and Stemming of the words.
  • Used Pig and Hive in the analysis of data.
  • Created/modified UDF and UDAFs for Hive whenever necessary.
  • Loaded and transformed large sets of structured, semi structured and unstructured data.
  • Supported Map Reduce Programs those are running on the cluster.
  • Managed and reviewed Hadoop log files to identify issues when job fails.
  • Developed Pig UDFs for preprocessing the data for analysis.
  • Involved in writing shell scripts in scheduling and automation of tasks.
  • Worked on Hive for further analysis and for generating transforming files from different analytical formats to text files.
  • Recently started using Mahout for machine learning in identifying a more subtle classifier.

Environment: Hadoop (Gen 1), HDFS, MapReduce (MR1), Pig, Hive, Scoop, Oozie, Mahout, Java, Linux Shell Scripting and Big Data.

Confidential, Melville, NY

Hadoop/Big data Developer

Responsibilities:

  • Installed and configured Hadoop and Hadoop stack on a 7 node cluster.
  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables.
  • Involved in data ingestion into HDFS using Scoop and Flume from variety of sources.
  • Responsible for managing data from various sources.
  • Got good experience with NoSQL database Hbase.
  • Designed and implemented MapReduce-based large-scale parallel relation-learning system.
  • Worked with NoSQL databases like Hbase in creating Hbase tables to load large sets of semi structured data coming from various sources.
  • Evaluated the use of Zookeeper in cluster co-ordination services.
  • Installed and configured Hive and also wrote Hive UDAFs that helped spot market trends.
  • Used Hadoop streaming to process terabytes data in XML format.
  • Involved in loading data from UNIX file system to HDFS.
  • Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
  • Involved in creating Hive tables, loading the data using it and in writing Hive queries to analyze the data.
  • Gained very good business knowledge on health insurance, claim processing, fraud suspect identification, appeals process etc.

Environment: CDH4 with Hadoop 1.x, HDFS, Pig, Hive, Hbase, MapReduce, Java, Sqoop, Flume, Oozie, Linux, UNIX Shell Scripting and Big Data.

Confidential, Monroe, MI

Hadoop/Java Developer

Responsibilities:

  • Worked extensively in creating MapReduce jobs for search and analytics in the identification of various trends.
  • Worked on data analytics using Pig and Hive on Hadoop.
  • Evaluated Oozie for workflow orchestration in the automation of MapReduce jobs, Pig and Hive jobs.
  • Installed and configured Hadoop Map reduce, HDFS, Developed multiple Map Reduce jobs in java for data cleaning and preprocessing.
  • Captured data from existing relational database into HDFS using Scoop.
  • Captured the data logs from web server into HDFS using flume for analysis.
  • Experienced in writing Pig scripts and Pig UDFs to pre-process the data for analysis.
  • Experienced in managing and reviewing Hadoop log files.
  • Built front end using JSP, Servlets, HTML and JavaScript to create user friendly and appealing interface.
  • Used JSTL and built custom tags whenever necessary.
  • Used Expression Language to tie beans to UI components.
  • Gained very good business knowledge on health insurance, claim processing, fraud suspect identification, appeals process etc.

Environment: CDH4 with Hadoop 1.x, HDFS, Pig, Hive, Oozie, Scoop, flume, Servlets, JSPs, JSTL, HTML, JavaScript, JQuery, CSS, Linux Shell Scripting.

Confidential, Woburn, MA

Application Developer J2EE

Responsibilities:

  • Wrote Servlets and JSPs to generate UI for an internal application.
  • Developed user interface using Java Server Faces UI component framework.
  • Developed POJOs and Java beans to implement business logic.
  • Managed data to and from the database using JDBC connections.
  • Used Spring JDBC to write some DAO classes to interact with the database to access account information.
  • Involved in creation of tables and indexes and wrote complex SQL queries.
  • Designed, customized and implemented the search feature using Lucene internally.
  • Used expression language to wire UI components to the business objects in the model layer.
  • Used Git as version control system to manage the progress of the project.
  • Used Junit framework for unit testing of the application.

Environment: Servlets, JSPs, Java Beans, POJO, JSF, HTML, CSS, JavaScript, JQuery, SQL, JDBC, JSTL, EL, Stored Procedures, Lucene, Git, Junit.

Confidential, Billerica, MA

Software Engineer

Responsibilities:

  • Involved in designing, coding, debugging, documenting and maintaining a number of applications.
  • Participated in Hadoop training and development as a part of cross training program.
  • Prepared use cases and designed class diagrams and object models.
  • Involved in the creation of SQL tables and indexes and also wrote queries to read/manipulate data.
  • Used JDBC to establish connection between the database and the application.
  • Implemented controllers layer using servlets and JSPs.
  • Implemented view layer using JSPs, JSTL and EL and also made custom JSP tags.
  • Created the user interface using HTML, CSS and JavaScript.
  • Created/modified shell scripts for scheduling and automating tasks.
  • Wrote unit test cases using Junit framework.

Environment: Java (Jdk 1.6), Servlets, JSPs, Java Beans, POJO, JSF, HTML, CSS, JavaScript, JQuery, SQL, JDBC, JSTL, EL, Custom JSP tags, Junit, Oracle 9i/10g.

We'd love your feedback!