We provide IT Staff Augmentation Services!

Hadoop Developer Resume

4.00/5 (Submit Your Rating)

MI

SUMMARY:

  • Around 8+ years of professional experience this includes Design, Development, Integration, Deployment and Maintenance of quality software applications in big data Hadoop technologies using Java/J2EE Technologies.
  • Hands on experience in installing, configuring, monitoring and integration of Hadoop ecosystem components like MapReduce, HDFS, HBase, Pig, Hive, Oozie, Sqoop, Flume, Spark, kafka and Zookeeper.
  • Experience working with NoSQL database including Mongo DB, Cassandra and Hbase.
  • Experience in developing NoSQL database by using CRUD, Shading, Indexing and Replication.
  • Exported the analyzed data to various Databases like Teradata (Sales Data Warehouse), SQL - Server using Sqoop.
  • Developed stored procedures and queries using PL/SQL.
  • Working with relative ease with different working strategies like Agile, Waterfall and Scrum methodologies.
  • Hands-on programming experience in various technologies like JAVA, J2EE, HTML, XML.
  • Hands on experience in application development using Java, RDBMS and Linux Shell Scripting.
  • Developed enterprise applications using Scala.
  • Has extensively worked in Java technologies such as Java Beans, Servlets, and JDBC.
  • Hands on experience working with Java project build managers Apache MAVEN and ANT.
  • Maintenance/ Implementation of Commercial Software. Extensive work experience with Java/J2EE technologies such as Servlets, JSP, EJB, JDBC, JSF, Struts, spring, SOA, AJAX, XML/XSL, Web Services (REST, SOAP), UML, Design Patterns and XML Schemas.
  • Quick Learner with problem solving skills, willing to learn and implement new technologies
  • Ability to work effectively in cross-functional team environments, excellent communication and interpersonal skills.
  • Understand continuous integration tools and Java IDEs

TECHNICAL SKILLS:

Big Data Technologies: Hadoop, HDFS, Hive, MapReduce, Pig, Sqoop, Flume, Oozie, kafka Spark, Tableau

Programming Languages: C, C++, Java, Scala

Java/J2ee Technologies: Java, Java Beans, J2EE (JSP, Servlets, EJB), JDBC

Databases/ETL: Oracle 11g, SQL Server 2000, MySQL, SQL /PL SQL, Informatica v 8.x

Nosql Databases: HBase and Cassandra

Web Technologies: HTML, JSP, JavaScript, Ajax, XML, PHP, AWS

Servers: Web Sphere, Apache Web Serve, Tomcat Server 7.0

Methodologies: UML, OOP, OOA, OOD and Agile

Version Controls: Tortoise CVS Client, SVN

Mapping Tools: Hibernate 3.0

Operating Systems: LINUX (Centos and Ubuntu), Windows XP, 7, MS DOS, UNIX

Scripting Languages: Perl, Python, Shell scripts

Build Tools: Ant, Maven 2.2.1

IDE/Tools/Utilities: Eclipse Helios, MS Visio, MS Office 2010, Control M, SQL Programmer

PROFESSIONAL EXPERIENCE:

Confidential, MI

Hadoop Developer

Roles & Responsibilities:

  • Developed MapReduce and Hive jobs for analyzing large and critical datasets.
  • Imported data into HDFS from various SQL databases and files using Sqoop and from streaming systems using Storm into Big Data Lake.
  • Collected and aggregated large amounts of log data using Apache Flume and staged data in HDFS for further analysis.
  • Developed custom aggregate functions using Spark SQL and performed interactive querying.
  • Worked with HBase databases to create tables and store the processed data.
  • Wrote Pig scripts to store the data into HBase.
  • Created Hive tables, dynamic partitions, buckets for sampling, and worked on them using Hive QL.
  • Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
  • Involved in creating calculated fields and dashboards in Tableau for visualization of the analyzed data.
  • Worked with using different kind of compression techniques to save data and optimize data transfer over network using Snappy.
  • Configured deployed and maintained multi-node Dev and Test Kafka Clusters
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processed the data with Pig.
  • Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, manage and review data backups, manage and review Hadoop log files.

Environment: Hadoop, HDFS, Hive, Pig, Sqoop, Flume, Eclipse, Data Lake, Spark, MapReduce, Cloudera, Tableau, Snappy, Zookeeper, HBase, Scala 2.10/2.11, Windows 7/Vista/ XP, Linux, Unix, NoSQL, MySQL, Shell Scripting, Ubuntu, Teradata.

Confidential, Baltimore, MD

Hadoop Developer

Roles & Responsibilities:

  • Primary responsibilities include building scalable distributed data solutions using Hadoop ecosystem.
  • Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
  • Developed Simple to complex Map/reduce streaming jobs using Python language that are implemented using Hive and Pig.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Handled importing of data from various data sources, performed transformations using Hive, Map Reduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop.
  • Analyzed the data by performing Hive queries (Hive QL) and running Pig scripts (Pig Latin) to study customer behavior.
  • Tested Apache(TM) Tez, an extensible framework for building high performance batch and interactive data processing applications, on Pig and Hive jobs.
  • Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources.
  • Continuous monitoring and managing the Hadoop cluster using Cloudera Manager.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Used Mahout to understand the machine learning algorithms for an efficient data processing.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Created AWS Route53 to route traffic between different regions.
  • Written multiple Map Reduce programs in Java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other compressed file formats.

Environment: Hadoop, PIG, Hive, Apache Sqoop, Oozie, HBase, Zoo keeper, Cloudera manager, JDK1.6, HDFS, Map-Reduce, Apache Crunch, Unix, Linux, SQL server, Flume, Spark, lambda, DB2, HBase and Python, Windows 7/XP.

Confidential, Atlanta, GA

Hadoop Developer

Roles & Responsibilities:

  • Involved in capacity planning architecture of the Big Data platform.
  • Developed Map-Reduce programs in Java for data cleaning and pre-processing.
  • Created MapReduce jobs for data transformations and data parsing.
  • Created Hive scripts for extracting the summarized information from hive tables.
  • Created Hive UDFS to extract data from staging tables.
  • Involved in creating Hive tables, loading data and querying data.
  • Design and implementation of High Availability feature for Search Engine.
  • Volume testing to calculate cluster's throughput.
  • Helped the team to increase the Cluster size from 22 to 30 Nodes.
  • Solution architecture for proposals in Web technologies, log file analysis of storage equipment data.
  • Maintain System Integrity of all sub-components (primarily HDFS, MR).
  • Monitor System health, logs and respond accordingly to any warnings or failures.
  • Unit testing, Volume testing and Bug fixing.
  • Co-ordination with Client and offshore counterparts.

Environment: JDK1.6, CentOS, FLUME, HBase, Maven, Map-Reduce, Hadoop, Hive, Pig, Sqoop, Oozie, Zookeeper, SQL Server, Oracle, Windows 7/Vista, AWS,UNIX, Linux.

Confidential

Java Developer

Roles & Responsibilities:

  • Using OOAD Technology classes are designed in UML with the help of Rational Rose tool.
  • Created user-friendly GUI interface and Web pages using HTML and DHTML embedded in JSP.
  • JavaScript was used for the client side validations.
  • Designing and developing generic validator framework for modules and injecting these validators using hibernate framework.
  • Creating Hibernate POJOs, Hibernate mapping files for all database tables.
  • Developing GUI Screens using JSF (IBM Implementation) and for Ajax functionality.
  • Developed and deployed EJB's (Session and Entity) to implement the business logic and to handle various interactions with the database.
  • Involved in debugging the application.
  • Developed Servlets using JDBC for storing and retrieving user data into and from the SQL database.
  • Used Web Logic Application Server to deliver a new class of enterprise applications that enhance business interactions and transactions between a company and its key constituencies.
  • Used Web Logic Application Server to deliver high performance and scalability.
  • Written Database objects like Triggers, Stored procedures in SQL.
  • Interacted with the users and documented the System.
  • Used HP QA to manage the defects and issues.

Environment: JSP 2.0, JDBC, HTML, OOAD, Servlets, Web Services, Rational Rose, WSAD 5.0, UML, Java 1.6, EJB 2.0, JSF, QA, Hibernate, AJAX, Windows 7/XP, CVS, XML/XSL.

Confidential

Java Developer

Roles & Responsibilities:

  • Worked as software developer for Confidential on developing a supply chain management system.
  • The application involved tracking invoices, raw materials and finished products.
  • Gathered user requirements and specifications.
  • Developed the entire application on Eclipse IDE.
  • Developed and programmed the required classes in Java to support the User account module.
  • Used HTML, JSP and JavaScript for designing the front end user interface.
  • Implemented error checking/validation on the Java Server Pages using JavaScript.
  • Developed Servlets to handle the requests, perform server side validation and generate result for user.
  • Used JDBC interface to connect to database.
  • Used SQL to access data from Microsoft SQL Server database.
  • Performed User Acceptance Test.
  • Deployed and tested the web application on Web Logic application server.

Environment: JDK 1.4, Servlet 2.3, JSP 1.2, JavaScript, HTML, JDBC 2.1, SQL, MySQL Server, UNIX and BEA Web Logic Application Server.

We'd love your feedback!