We provide IT Staff Augmentation Services!

Hadoop Developer Resume

5.00/5 (Submit Your Rating)

Southfield, MichigaN

PROFESSIONAL EXPERIENCE

  • Around Seven years of professional IT experience with Two years of Big data Hadoop ecosystems experience in ingestion, storage, querying, processing and analysis of Big data.
  • Excellent understanding / knowledge of Hadoop architecture and various components such as HDFS, JobTracker, TaskTracker, NameNode, DataNode and MapReduce programming paradigm.
  • Hands on experience in installing, configuring, and using Hadoop ecosystem components like Hadoop MapReduce, HDFS, Oozie, Hive, Sqoop, Pig, Spark and Flume.
  • Expertise in writing Hadoop Jobs for analyzing data using Hive and Pig.
  • Experience in working with MapReduce programs using Apache Hadoop for working with Big Data.
  • Experience in managing and reviewing Hadoop log files.
  • Hands on experience with Apache Spark and exposure to stream data processing Storm.
  • Worked on MapReduce programs using Python libraries.
  • Experience in data management and implementation of Big Data applications using Hadoop frameworks.
  • Experience in understanding the security requirements for Hadoop and integrate with Kerberos authentication and authorization infrastructure.
  • Experience in designing, developing and implementing connectivity products that allow efficient exchange of data between our core database engine and the Hadoop ecosystem.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
  • Strong working experience in application development using JDBC, Servlets, JSP, Spring Framework, Spring MVC, Hibernate 3.0 ORM.
  • Expertise in J2EE Design Patterns like MVC, DAO, DTO, Front controller, Fascade and Singleton.
  • Proficient in writing ANT, Maven builds script to automate the application build and deployment.
  • Expertise in writing Junit and unittest test cases for different kind of applications and also development experience in mock object to simulate real objects.
  • Experienced in configuring and deploying over Oracle WebLogic, WebSphere and Tomcat server.
  • Worked in TCP/IP and UDP based network programming (sockets, datagrams and multicasting) in Java as well as in Python.
  • Experience in using Python libraries NumPy, Scipy, Pandas.
  • Techno-functional responsibilities include interfacing with users, identifying functional and technical gaps, estimates, designing custom solutions, development, leading developers, producing documentation, and production support.
  • Excellent interpersonal and communication skills, creative, research-minded, technically competent and result-oriented with problem solving and leadership skills.
  • Active participate in team to accomplish targets.

TECHNICAL SKILLS:

Hadoop/Big Data: HDFS, MapReduce, Hive, Flume, Pig, Zookeeper, Impala, HBase Sqoop, Oozie, SparkProgramming LanguagesC, C++, Java, Python, Pig Latin, Hive QL, Shell Script, SQL, PL/SQL

Methodologies: Agile, Design Patterns (Core Java and J2EE) Testing APIJunit, unittest

Frameworks: MVC, Hibernate, Spring, Hadoop

Databases: Oracle 7.x, 8.1.6, SQL Server7.0, SQL, PL SQL

Web Related: Servlets, JDBC, JSP, HTML, XML, JavaScript, AJAX, Web Services

Network Protocols: TCP/IP, UDP, HTTP, DNS, ICMP, DHCP

PROFESSIONAL EXPERIENCE

Confidential, Southfield, Michigan

Hadoop Developer

Responsibilities:

  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Experience with Amazon web services: EC2, S3, and EMR (Elastic Map Reduce) equivalent cloud computing approaches.
  • Installed and configured Flume, Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
  • Used HiveQL to store data into tables and to perform queries.
  • Used Hive to transform large datasets into required format.
  • Handling structured and unstructured data and applying ETL processes.
  • Used Pig for data deduplication.
  • Used Distributed Cache to store reference data and Improve performance of MR.
  • Extensively used Map-Reduce to process the available terabytes of data.
  • Worked extensively in creating map reduce jobs to the financial services such as passing the raw data, populating staging tables and storing the refined data in partitioned tables.
  • Worked on merging and updating the data using MapReduce programs in both Java and Python.
  • Exported the analyzed data to RDBMS using Sqoop for visualization and to generate reports for the BI team.
  • Used Spark API over Cloudera Hadoop YARN to perform analytics on data in Hive.
  • Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs, Python.
  • Exposure to stream data processing Storm.
  • Integrated Oozie with the rest of the Hadoop stack supporting several types of Hadoop jobs out of the box (like Java MapReduce, Pig, Hive, Sqoop) as well as system specific jobs (such as Java programs and shell scripts).
  • Involved in installing and configuring Kerberos for the authentication of users and Hadoop daemons.
  • Developed projects using Maven to build and deploy.

Environment: CDH5, Hive, Sqoop, Pig, Oozie, Cloudera Manager, Java, Python, Linux, Eclipse

Confidential, Pennsylvania

Java/J2EE Developer

Responsibilities:

  • The Premier Provider Application is a Supplier Services application.
  • Supplier Services application frameworks used Spring Framework, Spring MVC, Hibernate Object/Relational Mapper.
  • Implemented authentication and authorization for the web layer using the Spring Security.
  • The application uses the Hibernate as the JavaPersistanceAPI to perform all the database transactions.
  • Used the Log4j as the logging tool and unittest framework in python.
  • Wrote python scripts to parse XML documents and load the data in database.
  • Worked on importing the data and generating pivot tables using Pandas and weekly reports for subscribers.
  • Productively analyze and report data using Python (Pandas, NumPy, matplotlib).
  • Implemented SOAP Web services and Restful.
  • Hands on experience in using NumPypythonAPI.
  • Experience in using collections inPythonfor manipulating and looping through different user defined objects.
  • Developed Merge jobs inPythonto extract and load data into database.
  • Involved in writing Pig scripts and loading data into Hive.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way.

Environment: RAD 7, Java 6.0, Python, Pandas, Spring MVC Framework, Hibernate, Oracle 9i, TOAD, UNIX scripting, log4j, unittest, Oracle web logic, Webservices, Serena Dimensions, Pig

Confidential

Java/J2EE Developer

Responsibilities:

  • The Application is developed based on the Agile Development process, so responsible for gathering requirements from the clients regularly.
  • Developed the Application as a Reusable and highly customizable set of agile, stand alone components.
  • The Web Application is developed using Spring MVC and used Spring Dependency Injection and AOP.
  • The GUI is developed using the Servlets, JSP, JavaScript, HTML, CSS and AJAX.
  • Used the Built in Tag Libraries in the Framework, which ease the code development.
  • Performed dependency injection and configured spring beans for modules Spring IOC.
  • Used Maven to build and package the application.
  • Enhance and modify the presentation layer and GUI framework that are written using JSP and client-side validations done using JavaScript & design enhanced wireframe screens.
  • Excessive work in writing SQL Queries, Stored procedures using TOAD.
  • Deployed the Application on Tomcat server.
  • Wrote unit and system test cases for modified processes and Continuous Integration with the help of QC team and Configuration team on timely manner.
  • Using the log4j logging tool for logging all our jobs.
  • Using the Web Services for User Authentication for enabling the Authentication for the report to be accessible as External.
  • SOAP based XML messages are used for communicating with the Web Services.
  • Have been using the CVS as the version Control Tool.
  • Wrote Specification for the development.
  • Participated in Server side and Client side programming.
  • Wrote SQL stored procedures, used JDBC to connect to database.
  • Created a user Interface through HTML and JSP.
  • Involved in developing various Servlets.
  • Tested and validated the application on different testing environments.
Environment: Windows NT, Java (JDK 1.2), HTML, JSP, Servlets, Eclipse 3.0, Oracle 8.0, JDBC, Apache Tomcat.

We'd love your feedback!