We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

IA

PROFESSIONAL SUMMARY:

  • 7+years of professional experience that includes development, deployment, maintenance and support of various projects at major organizations
  • Comprehensive working experience in Software Development with proficiency in development of applications using Hadoop ecosystems such as ApacheHadoop, Pig, Hive, Impala, Zookeeper, Oozie, and also Core Java, JSP, Servlets, and JDBC
  • Experience in developing and implementing Map Reduce jobs using java to process and perform various analytics on large datasets
  • Hands on experience in Hibernate, JPA, and Spring framework
  • Knowledge of NoSQL databases such as HBase, MongoDB & Cassandra
  • Experience in developing PigLatin and HiveQL scripts for Data Analysis and ETL purposes and also extending the default functionality by writing User Defined Functions (UDFs) for data specific processing
  • Experience in job scheduling and monitoring through Oozie and ZooKeeper
  • Experience in migrating data to and from RDBMS and unstructured sources into HDFS using Sqoop & Flume
  • Experience in writing SQL queries, Stored procedures in PL/SQL for Oracle, DB2, MySQL databases
  • Strong web development skills with HTML, JavaScript, Angular JS, DHTML, XHTML, XSLT, XSL, VSS, Eclipse, Tomcat 5.0, 6.0 and UNIX, Linux, Windows
  • Experience with application of System Development Life Cycle (SDLC) for projects

TECHNICAL SKILLS:

Hadoop/Big Data: Apache Hadoop, YARN, HDFS, MapReduce, HBase, Pig, Hive, Sqoop, Oozie, Flume, Cassandra, Zookeeper

Java & J2EE Technologies: Core Java, Servlets, JSP, JDBC, Java Beans

Frameworks: Spring, Hibernate, Spring MVC, Spark

Databases: Oracle, DB2, MySQL

Web Technologies: Java Script, Ajax, HTML, DHTML, XML, CSS

Programming Languages: C, C++, Java, JQuery, Python, Linux shell script

IDE: Eclipse, Netbeans

Web Servers: WebLogic, WebSphere, Apache Tomcat

Network Protocols: TCP/IP, UDP, HTTP, DNS

ETL Tools: Informatica, Pentaho

Testing: Win Runner, Load Runner, QTP

PROFESSIONAL EXPERIENCE:

Hadoop Developer

Confidential, IA

Responsibilities:

  • Written MapReduce jobs and pig scripts using various input and output formats. Also designed custom format as per the business requirements
  • Used Pentaho for creating dashboards, to display Key performance indicators and business data movements, alongside with Business Intelligence team
  • Used Sqoop to transfer data between MySQL Relational Database and HDFS for processing and exporting data to RDBMS
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre - processing, analyzing and testing the classifier using MapReduce, Pig and Hive jobs
  • Integrated Oozie with the rest of theHadoopstack, supporting several types ofHadoopjobs like, Pig, Hive, Sqoop, and as well as system specific jobs such as Perl and shell script
  • Automation of jobs, for pulling netflow data from relational databases to load data into Hive tables, using Oozie workflows and enabled email alerts on any failure cases
  • Implemented Database access through JDBC, using Oracle at the server end
  • Used Spring Aspect Oriented Programming (AOP) for addressing cross cutting concerns
  • Developed request/response paradigm by using Spring Controllers, Inversion of Control and Dependency Injection with Spring MVC
  • Used Web Services like SOAP and WSDL to communicate over Internet
  • Involved in implementation of the JMS Connection Pool, including publish and subscribe using Spring JMS
  • Used CVS tool as version control system and Log4j to create logs of various error and informational messages
  • Use of tools like JProbe and JConsole to debug application from memory leaks and resource utilization
  • Developed test classes in JUnit for implementing unit testing
  • Deployed the application using WebLogic Application Server
  • Developed Pig UDFs for preprocessing of the data for analysis
  • Created/modified UDF and UDAF's for Hive, whenever necessary
  • Used Pig and Hive in the analysis of data
  • Extracted files from NoSQL databases like CouchDB, and Cassandra using Sqoop
  • Worked with Flume to import the log data from the reaper logs and syslog's into theHadoopcluster
  • Used complex data types like bags, tuples, and maps in Pig for handling data
  • Involved in managing and running pending tasks on Map Reduce through Cloudera manager console
  • Involved in writing shell scripts for scheduling and automation of tasks
  • Managed and reviewedHadooplog files to identify issues, when job fails
  • Hands on experience with NoSQL databases like HBase, Cassandra for POC (Proof of Concept) in storing URL's, images, products and supplements information at real time
  • Worked on Hive for analysis, generating, and transforming files from different analytical formats to text files
  • Used Hue for UI based PIG script execution and Oozie scheduling
  • Involved in writing Hive queries for data analysis, with respect to business requirements

Environment: ApacheHadoop(Gen 1), Java, JDK1.6, J2EE, JDBC, Servlets, JSP, JSTL, HTML, Ajax, JAX-WS, Spring 2.0, Hibernate 4.2, Java Script, CSS, XSLT, XML. JUnit, WebServices, SOAP, WSDL, WebLogic, CVS, Rational Application Developer RAD, Rational Rose, JMS, Eclipse, Hive, Pig, Sqoop, Oozie, Cassandra, NoSQL, HBase, CouchDB, MapReduce (MR1), Cloudera, HDFS, Flume, Hue, Pentaho, Oracle 10.2, Tableau, Agile, SVN, Git, SourceTree, Jira, and Linux shell scripting.

Hadoop Developer

Confidential, Durham, NC

Responsibilities:

  • Installed and configuredHadoopMapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and preprocessing
  • Importing and exporting data into HDFS from Oracle 10.2 database and vice versa using SQOOP
  • Experienced in defining and coordination of job flows
  • Gained experience in reviewing and managingHadooplog files
  • Extracted files from NoSQL database like CouchDB, HBase through Sqoop and placed in HDFS for processing
  • Involved in Writing Data Refinement Pig Scripts and Hive Queries
  • Good knowledge in runningHadoopstreaming jobs to process terabytes of xml format data
  • Load and transform large sets of structured, semi structured and unstructured data
  • Coordinated cluster services using ZooKeeper
  • Designed applications using Struts with Tiles and Validator, implementing MVC design pattern and writing Custom Tag Libraries, JSP, Java Beans, Struts Controller, Action and Action Form classes using Struts tag libraries
  • Used XML Technologies like DOM for transferring data
  • Object relational mapping and Persistence mechanism is executed using Hibernate ORM
  • Developed custom validator in Struts and implemented server side validations using annotations
  • Created struts-config.xml file for the Action Servlet to extract the data from specified Action form so as to send it to specified instance of action class
  • Used Oracle for the database and WebLogic as the application server
  • Involved in coding for DAO Objects using JDBC (using DAO pattern)
  • Used Flume to transport logs to HDFS
  • Experienced in moving data from Hive tables into Cassandra for real time analytics on hive tables
  • Organize documents in more useable clusters using Mahout
  • Configured connection between HDFS and Tableau using Impala for Tableau developer team
  • Responsible to manage data coming from different sources
  • Got good experience with various NoSQL databases
  • Experienced with handling administration activations using Cloudera manager
  • Supported MapReduce programs those are running on the cluster
  • Involved in loading data from UNIX file system to HDFS
  • Installed and configured Hive and also written Hive UDFs
  • Involved in creating Hive tables, loading with data and writing Hive queries which will run internally in map reduce way

Environment: ApacheHadoop, Java, JDK1.6, J2EE, JDBC, Servlets, JSP, Struts 2.0, Spring 2.0, Hibernate 3.0, Linux, XML, WebLogic, SOAP, WSDL, HBase, Hive, Pig, Sqoop, ZooKeeper, NoSQL, HBase, R, MAHOUT Map-Reduce, Cloudera, HDFS, Flume, Impala, Tableau, and MySQL

Hadoop Developer

Confidential, Bloomington, IL

Responsibilities:

  • Involved in the requirements review meetings and partnered with business analysts to clarify any specific scenarios
  • Analyzed the business requirements and created use cases and prepared low and high level design documents
  • Involved with the application teams to installHadoopupdates, patches and version upgrades as required
  • Developed Map Reduce jobs for data cleaning and transformation
  • Developed Hive queries and Pig scripts to analyze large datasets
  • Collected the logs data from the web servers and integrated it to HDFS using flume
  • Involved in importing and exporting the data from RDBMS to HDFS and vice versa using Sqoop
  • Created Hive tables and analyzing the loaded data in the hive tables using hive queries
  • Experienced in integrating hive and HBase for better performing the Map Reduce algorithms
  • Loaded large sets of structured, semi - structured and unstructured data
  • Used Oozie job scheduler to automate the job flows
  • Developed Pig UDFs in java for pre processing the data
  • Cluster coordination services using ZooKeeper
  • Created and maintained Technical documentation for all the tasks performed like executing Pig scripts and Hive queries
  • Involved in fixing QA defects and tracked then using HP Quality Center

Environment: HadoopEcosystem, HDFS, MapReduce, Pig, Hive, Sqoop, Oozie, Eclipse, RDBMS, Oracle 9i, and HP Quality Center

Java Developer

Confidential

Responsibilities:

  • Coded the java programs in development phase and deployed in dev, stage and production environments.
  • Developed many front end phases and integrated with backend using the customized service architecture
  • Maintained the code resource using the version controlling tools CVS
  • Drafting the business requirements, analyzing converting them into the functionalities
  • Performed functional testing for the new Iterations
  • Documentation of the test procedures for the performance testing
  • Involved in the day-to-day support for the entire application
  • Responsible for identifying problem areas in application and for providing recommendations to speed up the process
  • Developed the screen and code design for different modules

Environment: Core Java, J2EE, JSP, Servlets, Struts, Java Script, Tomcat, Eclipse, JUnit, Oracle, SQL, PL/SQL, and Log4J

Java Developer

Confidential

Responsibilities:

  • Involved throughout the software development life cycle
  • Key involvement in Java code changes for the application, and performed Regression testing in Unit, Integration, and Acceptance regions, and staging to Production environment
  • Actively involved in designing the application for performance tuning
  • Designed UNIX shell scripts to run as scheduled jobs for many automated application process
  • Actively incorporated various change requests in the application within a limited timeframe and yet delivered a quality code
  • Sustained and enhanced the existing services, providing immediate fixes to several issues in production environment
  • Provided production support for the application. Came up with quick solutions and work around for various production incidents
  • Involved in weekly status and monthly status meetings
  • Developed the support documents of the application and maintaining up to date support procedure to off shore and on shore teams
  • Responsible for identifying arising defects and providing permanent fix, thereby reducing the number of defects arising and stabilizing the application
  • Developed SQL Queries, Stored Procedures, and Sub Queries
  • Handled configuration and change management for the source code of the application in Visual Source Safe CVS

Environment: Core Java, Eclipse, JUnit, XML, PL/SQL, Log4J, UNIX, Putty, and ASE-iSQL, Autosys, Sybase, SQL Server, CVS

We'd love your feedback!