We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

Virginia Beach, VA

PROFESSIONAL SUMMARY:

  • 8+ years of overall experience in IT Industry which includes experience in Big Data
  • Ecosystem related technologies Hadoop 2.0.
  • 3 years of comprehensive experience in Big Data Technology Stack.
  • Good knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, and MapReduce/Yarn concepts.
  • Well versed in installation, configuration, supporting and managing of Big Data and underlying infrastructure of Hadoop Cluster along with CDH3&4 clusters.
  • Experience in installing, configuring, and administrating Hadoop cluster for various distributions like Apache, Cloudera, Map R and Horton works.
  • Very good understanding on NOSQL databases like Hbase and Cassandra.
  • Designed and implemented a Cassandra NoSQL based database and associated RESTful web service that persists high - volume user profile data for vertical teams.
  • Experience in building large scale highly available Web Applications.Working knowledge of web services and other integration patterns.
  • Experience in managing and reviewing Hadoop log files.
  • Experience in using Pig, Hive, Scoop and Cloudera Manager.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
  • Generated Java APIs for retrieval and analysis on No-SQL database such as HBase and Cassandra.
  • Analyzed the data on Cassandra cluster by running queries for searching, sorting and grouping.
  • Hands-on experience with message brokers such as Apache Kafka and RabbitMQ.
  • Hands on experience in RDBMS, and Linux shell scripting.
  • Extending Hive and Pig core functionality by writing customUDFs.
  • Experience in analyzing data using HiveQL, Pig Latin, Map Reduce and Yarn.
  • Developed MapReduce jobs to automate transfer of data from HBase.
  • Knowledge in job work-flow scheduling and monitoring tools like Oozie and Zookeeper.
  • Knowledge of data warehousing and ETL tools like Informatica and Pentaho.
  • Experience with Eclipse/ RSA.
  • Knowledge on coding skills PYTHON, Kafka and Spark.
  • Knowledge on Nifi.
  • Expertise in RDBMS like MS SQL Server, MySQL, Greenplum and DB2.
  • Knowledge of job workflow scheduling and monitoring tools like Oozie and Zookeeper, of NoSQL databases such as HBase, Cassandra, and of administrative tasks such as installing Hadoop, Commissioning and decommissioning, and its ecosystem components such as Flume, Oozie, Hive and Pig.
  • Knowledge in Software Development Life Cycle (Requirements Analysis, Design, Development, Testing, Deployment and Support).
  • Excellent Java development skills using J2EE, Spring, J2SE, Servlets, JUnit, MRUnit, JSP, JDBC.
  • Techno-functional responsibilities include interfacing with users, identifying functional and technical gaps, estimates, designing custom solutions, development, leading developers, producing documentation, and production support.
  • Excellent interpersonal and communication skills, creative, research-minded, technically competent and result-oriented with problem solving and leadership skills.

TECHNICAL SKILLS:

Programming Languages: C, C++, Java, Shell Scripting, PL/SQL

J2EE Technologies: Spring, Servlets, JSP, JDBC, Hibernate.

BigData Ecosystem: HDFS, HBase, MapReduce, Hive, Pig, PythonSqoop, Impala, Kafka, Cassandra, Oozie, Zookeeper, Flume.

DBMS: Oracle 11g, SQL Server, MySQL.

Modeling Tools: UML on Rational Rose 4.0.

Web Technologies: HTML, JavaScript, XML, jQuery, Ajax, CSS.

Web Services: Web Logic, WebSphere, Apache, Cassandra,TomcatIDEs:Eclipse, Netbeans, WinSCP.

Operating systems: Windows, Unix, Linux (Ubuntu), Solaris, Centos.

Version and Source Control: CVS, SVN.

Servers: Apache Tomcat.

Frameworks: MVC, Struts, Log4J, Junit, Maven, ANT, WebServices.

WORK EXPERIENCE:

Confidential - Virginia Beach, VA

Hadoop Developer

Responsibilities:

  • Design, deploy, Manage cluster nodes for our data platform operations (racking/stacking)
  • Install and configure cluster. Setting up puppet for centralized configuration management.
  • Monitoring Cluster using various tools to see how the nodes are performing.
  • Managing the received data using Pentaho ETL tool and upload the same to the database.
  • Expertise in cluster task like Adding Nodes, Removing Nodes without any effect to running jobs and data.
  • Write scripts to automate application deployments and configurations. Monitoring YARN applications. Troubleshoot and resolve cluster related system problems.
  • Wrote map reduce programs to clean and pre-process the data coming from different sources.
  • Implemented various output formats like Sequence file and parquet format in Map reduce programs. Also, implemented multiple output formats in the same program to match the use cases.
  • Developed Hadoop streaming Map/Reduce works using Python.
  • Performed benchmarking of the No-SQL databases, Cassandra and HBase.
  • Hands on experience with Lambda architectures.
  • Created data model for structuring and storing the data efficiently. Implemented partitioning and bucketing of tables in Cassandra.
  • Implemented test scripts to support test driven development and continuous integration.
  • Converted text files into Avro then to parquet format for the file to be used with other Hadoop eco system tools.
  • Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
  • Exported the analyzed data to HBase using Sqoop and to generate reports for the BI team.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Participate in requirement gathering and analysis phase of the project in documenting the business requirements by conducting workshops/meetings with various business users.
  • POC work is going on using Spark and Kafka for real time processing.
  • Developed a data pipeline using Kafka and Storm to store data into HDFS.
  • Populated HDFS and Cassandra with huge amounts of data using Apache Kafka.
  • POC work is going on comparing the Cassandra and HBase NoSQL databases.
  • Hands-on experience in Pentaho.

Environment: MapReduce, HDFS, Hive, Pig, Hue, Oozie, Core Java, Eclipse, Hbase, Flume, Spark, Kafka, Cloudera Manager, Cassandra, Python, Greenplum DB, IDMS, VSAM, SQL*PLUS, Toad, Putty, Windows NT, UNIX Shell Scripting, Pentaho, Talend, Bigdata, YARN

Confidential - Omaha, NE

Hadoop Developer

Responsibilities:

  • Developed the application using Struts Framework that leverages classical Model View Controller (MVC) architecture.
  • Extensively worked on User Interface for few modules using JSPs, JavaScript and Ajax
  • Created Business Logic using Servlets, POJO's and deployed them on Web logic server
  • Involved in running Hadoop jobs for processing millions of records of text data.
  • Responsible for Cluster maintenance, adding and removing cluster nodes, Cluster Monitoring and Troubleshooting, manage and review data backups and log files.
  • Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop Clusters.
  • Monitored multiple Hadoop clusters environments using Ganglia.
  • Experience with Pentaho Data Integration.
  • Managing and scheduling Jobs on a Hadoop cluster.
  • Involved in defining job flows, managing and reviewing log files.
  • Monitored workload, job performance and capacity planning using Cloud era Manager.
  • Installed Oozie workflow engine to run multiple Map Reduce, Hive and Pig jobs.
  • Implemented Map Reduce programs on log data to transform into structured way to find user information.
  • Responsible for loading and transforming large sets of structured, semi structured and unstructured data.
  • Collected the log data from web servers and integrated into HDFS using Flume.
  • Responsible to manage data coming from different sources.
  • Extracted files from Couch DB and placed into HDFS using Sqoop and pre-process the data for analysis.
  • Gained experience with NoSQL database.
  • Unit tested and tuned SQLs and ETL Code for better performance.
  • Performed major role in understanding the business requirements and designing and loading the data into data warehouse (ETL).
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.

Environment: JDK 1.5, J2EE 1.4, Struts 1.3, JSP, Spring, Servlets 2.5, WebSphere 6.1, HTML, XML, JavaScript, Informatica Power Center 9.1,Hadoop, HDFS, Pig, Hive, MapReduce, HBase, Sqoop, Python,Oozie, Ganglia and Flume.

Confidential

Java/J2EE Interface Developer

Responsibilities:

  • Created Use case, Sequence diagrams, functional specifications and User Interface diagrams using Star UML.
  • Involved in complete requirement analysis, design, coding and testing phases of the project.
  • Participated in JAD meetings to gather the requirements and understand the End Users System.
  • Developed user interfaces using JSP, HTML, XML and JavaScript.
  • Generated XML Schemas and used XML Beans to parse XML files.
  • Created Stored Procedures & Functions. Used JDBC to process database calls for DB2/AS400 and SQL Server databases.
  • Developed the code which will create XML files and Flat files with the data retrieved from Databases and XML files.
  • Created Data sources and Helper classes which will be utilized by all the interfaces to access the data and manipulate the data.
  • Developed web application called iHUB (integration hub) to initiate all the interface processes using Struts Framework, JSP and HTML.
  • Heavily used bug trackers Pivotal, jira, Bugzilla and Target Process
  • Developed the interfaces using Eclipse 3.1.1 and JBoss 4.1 Involved in integrated testing, Bug fixing and in Production Support.

Environment: Java 1.3, Servlets, JSPs, Java Mail API, Java Script, HTML, MySQL 2.1, Java Web Server 2.0, JBoss 2.0, RMI, Rational Rose, Red Hat Linux 7.1.

Confidential

Java/J2EE Java developer

Responsibilities:

  • Implemented Services using Core Java.
  • Developed and deployed UI layer logics of sites using JSP.
  • Developed the XML data object to generate the PDF documents and other reports.
  • Used Hibernate, DAO, and JDBC for data retrieval and medications from database.
  • Messaging and interaction of Web Services is done using SOAP.
  • Developed JUnit Test cases for Unit Test cases and as well as System and User test scenarios.
  • Designed and developed Struts like MVC 2 Web framework using the front-controller design pattern, which is used successfully in a number of production systems.
  • Spearheaded the “Quick Wins” project by working very closely with the business and end users to improve the current website’s ranking from being 23rd to 6th in just 3 months.
  • Normalized Oracle database, conforming to design concepts and best practices.
  • Resolved product complications at customer sites and funneled the insights to the development and deployment teams to adopt long term product development strategy with minimal roadblocks.
  • Convinced business users and analysts with alternative solutions that are more robust and simpler to implement from technical perspective while satisfying the functional requirements from the business perspective.
  • Applied design patterns and OO design concepts to improve the existing Java/JEE based code base.
  • Identified and fixed transactional issues due to incorrect exception handling and concurrency issues due to unsynchronized block of code.

Environment: Java 1.2/1.3, Swing, Applet, Servlets, JSP, custom tags, JNDI, JDBC, XML, XSL, DTD, HTML, CSS, Java Script, Oracle, DB2, PL/SQL, Weblogic, JUnit, Log4J and CVS.

We'd love your feedback!