We provide IT Staff Augmentation Services!

Hadoop Developer/ Admin Resume

3.00/5 (Submit Your Rating)

Chicago, IL

PROFESSIONAL SUMMARY:

  • Software Developer, having 7+ years of extensive experience in delivering challenging technology solutions, working with geographically distributed teams.
  • Over 3+ years of experience in dealing with Apache Hadoop components like BIG Data, HDFS, MapReduce, Hive, Pig, Sqoop, Oozie, and Big Data Analytics.
  • Solid experience in designing, implementing, and improving analytic solutions for Big Data on Apache Hadoop . Experience in Map Reduce programming model and Hadoop Distributed File Systems.
  • Deep understanding of data import & export from relational database into Hadoop cluster.
  • Good Knowledge in NoSQL databases like MongoDB, Cassandra and HBASE.
  • Experience in writing UDFS in java for hive and pig. Experience in using HCatalog for Hive, Pig, and HBase.
  • Extensive hands on administration with Cloudera.
  • Extensive knowledge on Flume and HBase technologies. Experience in working with flume to load the log data from multiple sources directly into HDFS.
  • Experience in Hadoop Shell commands, writing MapReduce Programs, verifying managing and reviewing Hadoop Log files.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
  • Hands on experience in Import/Export of data using Hadoop Data Management tool SQOOP.
  • Provided technical assistance for configuration, administration and monitoring of Hadoop clusters.
  • Experience in developing customized UDF’s in java to extend Hive and Pig Latin functionality.
  • Experience in test environment setup and test infrastructure development in both manual and automation.
  • Experience with database testing using various complex set of queries and UDF’s.
  • Experience in all phases of Software Development Life Cycle for maintaining and supporting the Java, and J2EE applications.
  • Good experience in designing, developing database to create its Objects like Tables, Stored Procedures, Triggers, Cursors and writing SQL queries .
  • Quick learner and self-starter with effective communication, motivation and organizational skills combined with attention to details and business process improvements.
  • Excellent team player as well as an individual with strong analytical, communication and interpersonal skills.

TECHNICAL SKILLS:

Hadoop Ecosystems: HDFS, MapReduce, HBase, Hive, Pig, Sqoop, Oozie, Zookeeper, Flume, Cascading.

Scripting Languages: PL/SQL, Shell Scripting.

Java Technologies: Java, J2EE, Servlets, Java Beans, JSP, JDBC.

Database: MYSQL, Oracle 12c, Mongo DB, Cassandra, Teradata.

Tools: Eclipse, NETBEANS, RAD, Maven, IntelliJ, Visual Studio, Log4J, UML.

Testing Tools: Selenium IDE and Web Driver, HP Quick test professional (QTP).

Servers: Apache Tomcat, IBM WebSphere, Oracle WebLogic server, SQL Server.

Operating Systems: DOS, Windows, Linux, Ubuntu, Fedora, MAC OSX.

PROFESSIONAL EXPERIENCE:

Confidential, Chicago, IL

Hadoop Developer/ Admin

Responsibilities:

  • Involved in ETL, Data Integration and Migration. Imported data using Sqoop to load data from Oracle to HDFS on regular basis.
  • Responsible for implementation and ongoing administration of Hadoop infrastructure.
  • Installed and configured Hadoop Mapreduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Worked with data delivery teams to setup new Hadoop users.
  • Worked on setting up Pig, Hive, and Hbase on multiple nodes.
  • Developed SQL Queries to retrieve or create test data from different databases.
  • Used Pig as ETL tool to do transformations, event joins, filter bot traffic and some pre-aggregations before storing the data onto HDFS.
  • Written Hive queries for data analysis to meet the business requirements.
  • Created Hive tables and working on them using Hive QL. Importing and exporting data into HDFS from Oracle Database and vice versa using Sqoop.
  • Performed Cluster coordination services through Zoo Keeper.
  • Installed Oozie workflow engine to run multiple Hive.
  • Responsible to manage data coming from different sources.
  • Experience in managing and reviewing Hadoop log files.
  • Created jobs to load data from MongoDB into Data warehouse.
  • Developed Scripts and Batch Job to schedule various Hadoop Program.
  • Managing and scheduling Jobs on a Hadoop cluster.
  • Installed and configured Pig and also written PigLatin scripts.
  • Wrote MapReduce jobs using Pig Latin.
  • Involved in analyzing system design specifications and developed test plans, test scenarios and test cases to cover overall quality testing.
  • Implemented test scripts to support test driven development and continuous integration.

Environment: Hadoop, Hive, Linux, MapReduce, HDFS, Pig, Sqoop, Shell Scripting, Java (JDK 1.7), Java 7, Eclipse, Oracle 10g, PL/SQL, SQL*PLUS, Toad 9.6, Linux, QTP.

Confidential, Atlanta, GA

Hadoop Developer/ Admin

Responsibilities:

  • Involved in building a real-time analytics platform based on web-crawled data stored in Enterprise Data Warehouse (EDW).
  • Installed raw Hadoop and NoSQL applications and develop programs for sorting and analyzing data.
  • Responsible for Data in the warehouse is processed and transformed using Map Reduce and loaded into Hadoop Distributed File System (HDFS).
  • Analyzed data using Hive that exposes data in Hadoop File Systems in a distributed Query enabled platform. Sqoop is used to extract data from internal structured data stores and load into HDFS.
  • Understand customer analytic needs and translate them into pluggable Hadoop Hive UDFs/UDAFs. Involved in writing UDF functions for Hive.
  • Create and update Hive schema to support a highly-available data warehouse for BI teams.
  • Controlled Data volumes ranging in Petabytes, which makes it challenging for regular analytics platform.
  • Installing, Upgrading and Managing Hadoop Cluster on Cloudera.
  • Installed and configured Pig.
  • Used MongoDB extensively to filter required data for data scientists.
  • Used Flume to transport logs to HDFS.
  • Involved in loading and maintenance of NoSQL database.
  • Setup Hive with MySQL as a Remote Metastore.
  • Modified required UNIX scripts and other SQL validation scripts, and writing scripts in SQL to validate the outputs.
  • Automated several jobs for pulling data from FTP server and load data into Hive tables, using Oozie workflows.
  • Developed business requirements by reaching the customer directly.
  • Collected specifications and requirements and developed Test Plan accordingly.
  • Involved in modifying existing Test Cases as per new requirements.

Environment: MapReduce, HDFS, Hive, Pig, Hue, Oozie, Spark, Core Java, Eclipse, HBase, Flume, Cloudera Manager, NOSQL, Mongo DB, PL/SQL, SQL*PLUS, Toad, Putty, UNIX Shell Scripting.

Confidential, Atlanta, GA

Hadoop Developer

Responsibilities:

  • Installed and configured various components of Hadoop ecosystem like JobTracker, TaskTracker, Name Node and Secondary Name Node.
  • Used Sqoop to dump data from relational database into HDFS for processing and exporting data to RDMS.
  • Configured to capture the news from various sources for testing the classifier.
  • Wrote extensive MapReduce Jobs in Java to train the classifier.
  • Wrote MR jobs and pig scripts using various Input and Output formats. Also used custom formats whenever necessary.
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing, analyzing and training the classifier using MapReduce jobs, Pig jobs and Hive jobs.
  • Used Pig and Hive in the analysis of data.
  • Created/modified UDF and UDAFs for Hive whenever necessary. Developed Pig UDFs for preprocessing the data for analysis.
  • Loaded and transformed large sets of structured, semi structured and unstructured data.
  • Managed and reviewed Hadoop log files to identify issues when job fails.
  • Involved in writing shell scripts in scheduling and automation of tasks.
  • Used Hue for UI based PIG script execution, Oozie scheduling and creating tables in Hive.
  • Assisted admin team in setting up additional nodes in the cluster.
  • Supported extensively Load, Stress and Performance testing using LoadRunner.
  • Identified the requirements and objectives of the project and designed a plan to work out through the project.
  • Engaged in analyzing, reviewing and tracking bugs using a defect tracking database in Test Director.
  • Involved in weekly walkthroughs and inspection meetings, to verify the status of the testing efforts and the project as a whole.

Environment: Hadoop, HDFS, Map Reduce, Pig, Hive, H-base, Sqoop, Oozie, Cloudera manager, Flume, Zookeeper, LoadRunner and Linux Ubuntu operating system, Shell scripting.

Confidential

Java/ J2EE Developer

Responsibilities:

  • Developed E-commerce product allowing websites to accept online payments and establish credit relationships.
  • Developed application based on SDLC (Software Development Life Cycle). Designed application with UML (Unified Modeling Language).
  • Developed complete Web tier of the application with spring framework.
  • Used JMS for sending asynchronous data between the different session beans.
  • Developed the presentation layer using JSP and Servlets with a RAD tool.
  • Used JavaScript for client side validations.
  • Used Spring Core for middle tier development to achieve inversion of control.
  • Developed complete Business tier with Stateless Session Beans. Used Hibernate for Persistence layer. DAO wrapped all the Hibernate functions.
  • Developed the automatic build scripts using Ant for the application to deploy and test.
  • Used Clear Case for version control.
  • Created JUNIT test cases.
  • Used CVS for version control.
  • Implemented Log4J for Logging Errors, debugging and tracking.
  • Involved in testing the complete flow of the modules.

Environment: Java EE 5, Struts, spring, Unix, EJB, Http, HTML, XML, XSL, CSS, Servlets, JavaScript, Dojo js, GWT, AJAX, EXT JS Multicast, Hibernate, Cloud computing, JUNIT, JSP, Web services, UML.

Confidential

Java Developer

Responsibilities:

  • Involved in design, development and testing phases of the application.
  • Implemented the application using Java object oriented principles.
  • Involved in Administering and Configuring of BEA WebLogic Server v8.1 that runs application Servlets, JSPs, EJBs, etc. Used XSLT to transform test-specific XML scripts into HTML.
  • Supported the application by resolving tickets and reviewing the patrol monitoring reports.
  • Worked closely with project managers, analysts and other developers to meet tight timeframes for issue resolutions during production deployment.
  • Raise and track Problem Reports for the bugs found during testing using PRISM (defect tracking tool provided by BA).

Environment: Java, Servlet, JSP, JavaScript, JUnit, HTML, XSLT, XML, BEA WebLogic Server v8.1, Eclipse, Putty.

We'd love your feedback!