Hadoop Developer Resume
NC
SUMMARY
- 7+ years of professional experience in IT and Enterprise Application Development in multiple industries which includes hands on experience in Big Data/Hadoop ecosystem and related technologies. 3 years of comprehensive experience on Big Data Technologies.
- Excellent understanding / knowledge of Hadoop Architecture, its components Map Reduce & HDFS. Good understanding on various deamon processes like Job Tracker, Task Tracker, Name Node, Data Node.
- Experience in working with Hadoop/BigData storage and analytical frameworks over Amazon AWS Cloud using tools like SSH, Putty.
- Experience in installing, configuring, and administrating Hadoop cluster for distributions like Cloudera and Horton works.
- Very good understanding of NOSQL database HBase
- Designed and implemented Cassandra, a NoSQL based database and associated RESTful web service that persists high - volume user profile data for vertical teams.
- Very well experienced in designing and developing both server side and client side applications.
- In-depth knowledge on Hadoop Eco system components like PIG, Hive, Sqoop, HBase, Oozie, ZooKeeper, Hue, Cloudera Manager, Flume, Spark, Scala.
- Hands-on experience writing MapReduce Jobs
- Experience in managing and reviewing Hadoop log files.
- Experience in using YARN,Pig, Hive, Scoop.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Hands-on experience in RDBMS, and Linux shell scripting
- Experience in extending Hive and Pig core functionality by writing customUDFs
- Experience in analyzing data using HiveQL, Pig Latin and Map Reduce.
- Knowledge in job work-flow scheduling and monitoring tools like oozie and Zookeeper.
- Knowledge of data warehousing and ETL tools like Informatica and Pentaho.
- Experience coding with Eclipse.
- Explored Spark,Kafka,Stormalong with other open source projects to create a realtime analytics framework
- Expertise in RDBMS like MS SQL Server, MySQL, and DB2.
- Extensive experience in using MVC architecture,Struts,Hibernate for developing web applications using Java, JSPs, JavaScript, HTML, jQuery, AJAX, XML and JSON
- Excellent Java development skills using J2EE, Spring, J2SE, Servlets, JSP
- Extensively worked in Core Java and object oriented design with strong understanding of Collections, Multithreading and Exception handling.
- Experience in using Agile software methodology (scrum).
- Good in Database concepts and writing finely tuned queries.
- Certified Harvard ManageMentor Professional.
- Strong knowledge in writing advanced Shell Scripts in Linux/Unix
- Excellent project documentation skills which includes creation of design documents which constitute UML diagrams.
- Skilled in establishing strong relations among project’s teams and team members.
TECHNICAL SKILLS
Bigdata Ecosystem: MapReduce, HDFS, Hive, PigHBase, Zookeeper, Sqoop, Oozie,Flume,Spark,Kafka,Storm
Java/J2EETechnologies: Java, J2EE, Servlets, JSP, XML, AJAX, SOAP, WSDL
SDLC Methodologies: Agile, UML, Design Patterns (Core Java and J2EE)
Enterprise Frameworks: Ajax, MVC, Struts 2/1, Hibernate 3, Spring 3/2.5/2
Version and Source Control: CVS,SVN
Programming Languages: C,C++,Java, XML, Unix Shell scripting, SQL and PLSQL
Web Technologies: HTML, DHTML, XML, XSLT, JavaScript, CSS
Modeling Tools: UML on Rational Rose 4.0
IDE Tools: Eclipse, Netbeans, WinSCP.
WebServices: WebLogic, WebSphere, JBoss
Databases: Oracle DB2, MS - SQL Server, MySQL, MS - Access, Apache Cassandra
Frameworks: MVC, Struts, Log4J, Junit, Maven, WebServices.
Operating Systems: Windows 9x/NT/XP, UNIX, Linux,Centos
PROFESSIONAL EXPERIENCE
Confidential, NC
Hadoop Developer
Responsibilities:
- Co-ordinated with business customers to gather business requirements
- Extensively involved in Design phase and delivered Design documents
- Good working knowledge of YARN Map Reduce V2 and its daemons Resource Manager and Application Master.
- Experience in loading data from Teradata database into HDFS using SQOOP
- Experience in moving large amount of log data from different servers
- Experience in writing transformer/mapping Map-Reduce pipelines
- Wrote map reduce programs to cleanse and pre-process the data coming from different sources.
- Implemented various output formats like Sequence file and ORC format in Map reduce programs.
- Responsible for creating Hive tables, loading the structured data resulted from Map Reduce jobs into the tables and writing Hive Queries to further analyze the data
- Implemented Hive Generic UDF’s to implement business logic. Also designed and implemented incremental imports into Hive tables
- Implemented Partitioning, Dynamic Partitions, and Buckets in Hive for efficient data access. Also designed both Managed and External tables in Hive for optimized performance
- Developed Hadoop streaming Map/Reduce works using Python.
- Developed Pig Latin Scripts and used ETL tool, informatica to do transformations, event joins, filter and some pre-aggregations
- Good hands-on experience using Informatica.
- Used Spark for iterative processing
- Explored Spark,Kafka,Stormalong with other open source projects to create a realtime analytics framework
- Implemented test scripts to support test driven development and continuous integration.
- Developed oozie workflow for scheduling & orchestrating the ETL process
- Involved in Unit testing and delivered Unit test plans and results documents using JUnit and MRUnit
- Performed data analysis with HBase using Hive External tables
- Exported the analyzed data to HBase using Sqoop and to generate reports for the BI team.
- As a POC we have installed Cassandra on the Amazon EC2
- Created tables, inserted data and executed various Cassandra Query language (CQL 3) commands on tables using cqlsh
- Used DataStax OpsCentre, a web based tool to monitor and to simplify administration tasks.
Environment: Cloudera,Python, MapReduce, Cassandra, AWS, HDFS, Hive, Pig, Hue, Spark, Oozie, Core Java, Eclipse, Hbase, Flume, Informatica, SQL*PLUS, Toad, Putty, UNIX Shell Scripting,Linux 5, YARN, Eclipse, Java SDK 1.6
Confidential, NC
Senior Hadoop Developer
Responsibilities:
- Good understanding and related experience with Hadoop stack-internals, Hive, Pig and Map/Reduce.
- Involved in requirement gathering of the enhancement for the project
- Deep understanding of schedulers, workload management, availability, scalability and distributed data platforms.
- Installed and configured HadoopMapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and pre-processing.
- Involved in loading data from UNIX file system to HDFS.
- Wrote MapReduce jobs to discover trends in data usage by users.
- Involved in managing and reviewing Hadoop log files.
- Involved in running Hadoop streaming jobs to process terabytes of text data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Developed Hive and PigUDF’s.
- Scripted HIVE queries for the analysts.
- Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
- Exported the result set from HIVE to MySQL using Shell scripts.
- Used Zookeeper for various types of centralized configurations.
- Involved in maintaining various Unix Shell scripts.
- Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
- Automated all the jobs starting from pulling the Data from different Data Sources like MySQL to pushing the result set Data to Hadoop Distributed File System using Sqoop.
- Used SVN for version control.
- Helped the team to increase Cluster from 25 Nodes to 40 Nodes.
- Maintain System integrity of all sub-components (primarily HDFS, MR, HBase, and Flume).
- Monitor System status and logs and respond accordingly to any warning or failure conditions.
Environment: Hadoop, Map Reduce, Hive QL, Hive, HBase, Sqoop, Kafka, Storm, Solr, Cassandra, Flume, Tableau, Impala, Oozie, Informatica, MYSQL, Oracle SQL, Java, Unix Shell, YARN, Pig Latin.
Confidential, MI
Java Developer
Responsibilities:
- Interacted with business analysts to study the requirements and design interfaces for the various use cases.
- Responsible for developing and modifying the existing service layer based on the business requirements.
- Implemented the view layer using JSP, Business layer using Spring 3.0, and persistent layer using DAO and Hibernate 3.2.
- Developed the necessary front end user interfaces in JSPs, HTML, Java Script, CSS that will be integrated using Spring MVC Framework.
- Worked on Spring Rest web service implementation.
- Developed Form Beans/ JavaBeans, Spring Controller Classes and configured with the corresponding JSP pages in spring-config.xml file.
- Developed HTML and JSP pages for user interaction and data presentation.
- Used Spring Framework for Dependency injection and integrated with Hibernate Framework.
- Participated in coding Spring AOP components for the Transactional Model to handle many requests. Involved in writing JSP and Servlet components. Used Spring's Bean Factory for initializing services.
- Implementation of Thread pool and locks for concurrency management.
- Used Spring Core Annotations for Dependency Injection.
- Created the Hibernate mapping and properties files using spring modules like MVC, Web and ORM. Used Hibernate template to integrate Spring DAO business objects with respective tables from the database.
- Involved in configuring hibernate to access database and retrieve data from the database.
- Wrote HBM Files for mapping Java objects to tables in relational database and updated it as required.
- Created tables, stored procedures in SQL for data manipulation and retrieval, Database Modification using SQL, PL/SQL, Stored procedures, triggers, Views in Oracle 9i.
- Created XML schema documents based on the requirements and used JAXB for marshalling and unmarshalling of the XML documents.
- Implemented various standard J2EE design patterns (Singleton, Factory, Data Access Object, MVC, Front Controller)
- Used Weblogic Server 8.2 for deploying the application.
- Used MAVEN for project management and build automation.
- Used Message Driven Beans (MDB)to read the messages fromJMSqueues for asynchronous communication.
- Used HP Quality center for defect tracking.
- Used Subversion as a version control for implementing the application.
- Controlled the log information by using framework Log4j.
- Performed Unit Testing Using JUnit.
Confidential
Java Developer
Responsibilities:
- Developed application layout and composing tiles definitions, managed beans, Transfer Objects to pass data over layers, Business Delegates for invoking business methods of EJBs, Stateless Session Beans which will access
- Business Objects, which will in turn call DAO's to retrieve & update data in the database.
- Performed Use case design, object modeling using UML, like transformation of the Use Cases into Class Diagrams, Sequence Diagrams.
- Participated in coding Spring AOP components for the Transactional Model to handle many requests. Involved in writing JSP and Servlet components. Spring Framework at Business Tier and uses Spring's Bean Factory for initializing services.
- Developed the AJAX asynchronous calls to the server to validate and populate the pages that are directed through view resolver.
- Used Spring Core Annotations for Dependency Injection.
- Developed Backing beans to handle UI components state and stores that state in a scope.
- Wrote EJB Stateless session to communicate with Controller and Service layer i.e. make the EJB calls from Backing Beans.
- Involved in writing the database integration code using Hibernate.
- Used Spring Framework with Hibernate to map to Oracle database.
- Wrote HBM Files for mapping Java objects to tables in relational database and updated it as required.
- Hibernate used as Persistence framework mapping the ORM objects to tables and developed HQL, SQL queries.
- Build the application using TDD (Test Driven Development) approach.
- Used Oracle as database and Involved in the development of PL/SQL backend implementation and using SQL created Select, Update and Delete statements.
- Used MAVEN for project management and build automation.
- Used Subversion as a version control for implementing the application.
Environment: Java 1.6, JSF 1.2, Spring 2.5, Hibernate 3.0, EJB 3.0, JMS, AJAX, Web services, IBM Web Sphere Application Server 8.0, Jboss, Java Beans, UML, XML, HTML, JavaScript, CSS, XSL, Oracle 10g, SQL, PL/SQL, Apache Maven, subversion, TFS, JIRA, Remedy(Incident Management Tool).
Confidential
Java Developer
Responsibilities:
- Developing Online panels and application using EJB,JavaServlets, Session & Entity Beans
- Handled the database persistence using JDBC.
- Implemented the project using the Spring Framework (used Spring IOC and Dependency Injection, Spring MVC)
- Used Spring Framework and created the Dependency injection for the Action classes using ApplicationContext.xml.
- Using Java Script functions for the custom validations.
- Expertise in RedhatLinuxCapacity and Performance Planning.
- Developed using Struts, JSPs and Tiles definitions.
- Developed JSP pages using HTML, JSTL Tag Library, Struts Tab Library,& JSP Custom Tags.
- Writing JSP form bean validations by using Struts Validation framework (validation.xml, validator-rules.xml and message Resources).
- Designed and developed REST web service for validating address.
- Developed the application usingJSFfor front end, spring boot to manage business layer components.
- Used Hibernate for object Relational Mapping and used JPA for annotations.
- Used Criteria API and HQL for data extraction.
- Performed Validations on UI data usingJSFvalidations and JavaScript.
- Involved in implementing the rich user interface using JSP Standard Tag Libraries and worked with custom tags.
- Performed client side validations using java script functions to implement various functionality
- Worked on ancillary technologies/tools (portal/portlet development, BPM, rules engines, security/SSO, and UML)
- Worked on designing/developing of large, transactional, enterprise class systems
- Worked on externally facing web based applications.
- Involved in implementation of the presentation layer (GUI) for the application using JSF, HTML, XHTML, CSS and JavaScript.
Environment: Core Java, Java EE, Spring, Struts, Hibernate, IBM WebSphere, WebLogic, Apache Tomcat 6.0, Web Services, HTML, XML, XSL, JSTL, JSP, AJAX, PL/SQL