Senior Hadoop Developer Resume
Houston, TX
SUMMARY
- Over 8+ years experience in Big Data Analytics, Hadoop, Java, Database Administration and Software development expertise.
- Provided 24x7 operational support for Production and Non production environments.
- Strong hands on experience in Hadoop Framework and its ecosystem including but not limited to HDFS Architecture, MapReduce Programming, Hive, Pig, Sqoop, Hbase, Oozie etc.
- Potential experience in (SDLC) Analysis, Design, Development, Integration and Testing in diversified areas of Client - Server/Enterprise applications using Java, J2EE technologies.
- Good work experience on JAVA, JDBC, Servlets, JSP
- Proficient in Java, J2ee, JDBC, Collections, Servlets, JSP, JSON, XML, XSLT, XSD, WSDL, WADL, REST, SOAP Web services, Gradle and Eclipse Link
- Strong experience in ETL tools on Oracle, DB2 and SQL Server Databases.
- Hands on experience working with Java project build managers Apache MAVEN and ANT.
- Worked with HBase and MongoDB
- Experienced in configuring Flume to stream data into HDFS.
- Familiarity with Hadoop architecture and its components like HDFS, Map Reduce, Job Tracker, Task Tracker, Name Node and Data Node.
- Experienced in Application Development using Java, Hadoop, RDBMS and Linux shell scripting and performance tuning.
- Proficient in Amazon WebServices.
- Familiarity with distributed coordination system Zookeeper.
- Excellent knowledge in Java and SQL in application development and deployment.
- In-depth understanding of Data Structures and Algorithms and Optimization.
- Worked with relational databases like MySQL, Oracle and NoSQL databases like HBase and Cassandra
- Well versed with databaseslike MS SQL Servers 2012 and 2008, Oracle 11g/10g/9i, MySQL.
- Versatile experience in utilizing Java tools in business, web and client server environments including Java platform, JSP, Servlets, Java beans and JDBC.
- Expertise in developing the presentation layer components like HTML, CSS, JavaScript, JQuery, XML, JSON, AJAX and D3.
- Experienced in source control repositories viz. SVN, GitHub.
- Experienced in detailed system design using use case analysis, functional analysis, modelling program with class and sequence, activity and state diagrams using UML.
- Experience in strong and analyzing data using HiveQL, Pig Latin, HBase and custom Map Reduce programs in Java.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Hands on experience in installing, configuring Cloudera's ApacheHadoop ecosystem components like Flume-ng, Hbase, Zoo Keeper, Oozie, Hive, Sqoop, Hue, Pig, Hue with CDH3&4 clusters
- Extending Hive and Pig core functionality by writing custom UDFs.
- Architected, Designed and maintained high performing ELT/ETL Processes.
- Good knowledge in integration of various data sources like RDBMS, Spreadsheets, Text files, JSON and XML files.
- Participated in an Agile SDLC to deliver new cloud platform services and components
- Developing and Maintenance the Web Applications using the Web server Tomcat.
- Exceptional ability to learn new technologies and to deliver outputs in short deadlines.
- Having Experience on UNIX commands, shell scripting and Deployment of Applications in Server.
TECHNICAL SKILLS
Hadoop: Hadoop 2.1, HDFS, MapReduce, PIG 0.8, Hive0.13, Hbase 0.94, Sqoop 1.4.4, Zoopkeeper 3.4.5, Yarn
Hadoop management & Security: Hortonworks Ambari, Cloudera Manager, Apache Knox, XA Secure
NoSQL Databases: MongoDB, Hbase, Redis and Cassandra
Web Technologies: DHTML, HTML, XHTML, XML, XSL (XSLT, XPATH), XSD, CSS, JavaScript
Server Side Scripting: UNIX Shell Scripting
Database: Oracle 11g/10g/9i/8i, MS SQL Server 2008/2005, DB2 v8.1, Teradata, MySQL.
Programming Languages: Java, J2EE, JSTL, JDBC 3.0/2.1, JSP 1.2/1.1, Java Servlets, JMS, JUNIT,Log4j, COBOL, C
Scripting Languages: Javascript
Web Servers: Apache Tomcat 5.x, BEA Weblogic 8.x, IBM Websphere 6.0/5.1.1
IDE: WSAD5.0, IRAD 6.0, Eclipse3.5, Dreamweaver13.2.1
OS/Platforms: Windows7/2008/Vista/2003/XP/2000/NT,Linux(All major distributions, mainly Centos and Ubuntu), Unix.
Client side: JavaScript, CSS, HTML, JQuery
XML: XML, HTML, DTD, XML Schema, XPath
Build tools: Maven and ANT
Data Modeling tools: Erwin, MS Visio
Cloud Platforms: Amazon EC2
Methodologies: Agile, UML, Design Patterns, SDLC
Office Tools: MS-OFFICE - Excel, Word, PowerPoint.
PROFESSIONAL EXPERIENCE
Confidential, Houston, TX
Senior Hadoop Developer
Responsibilities:
- Involved in loading data from Teradata, Oracle database into HDFS using Sqoop queries.
- Responsible for building scalable distributed data solutions using Hadoop.
- Developed multiple Map Reduce jobs in java for data cleaning and preprocessing.
- Developed Map Reduce pipeline jobs to process the data and create necessary HFiles.
- Involved in loading the created HFiles into HBase for faster access of large customer base without taking Performance hit.
- Created Hbase tables to store various data formats of PII data coming from different portfolios.
- Collecting and aggregating large amounts of log data using Apache Flume and staging data in HDFS for further analysis
- Involved in managing and reviewing Hadoop log files.
- Responsible to manage data coming from different sources.
- Involved in creating Pig tables, loading with data and writing Pig Latin queries which will run internally in Map Reduce way.
- Documented the systems processes and procedures for future references.
- Provide batch processing solution to certain unstructured and large volume of data by using Hadoop Map Reduce framework.
- Involved in writing Unix/Linux Shell Scripting for scheduling jobs and for writing pig scripts and hive QL.
- Developed Scripts and automated data management from end to end and sync up between all the clusters.
- Involved in creating Hive Tables, loading with data and writing Hive queries which will invoke and run MapReduce jobs in the backend.
- Assisted in performing unit testing of Map Reduce jobs using MRUnit.
- Assisted in exporting data into MongoDB and writing column families to provide fast listing outputs.
- Used Oozie Scheduler system to automate the pipeline workflow and orchestrate the map reduce jobs that extract the data on a timely manner.
- Actively participated in software development lifecycle (scope, design, implement, deploy, test), including design and code reviews, test development, test automation.
- Used Zookeeper for providing coordinating services to the cluster.
- Involved in story-driven agile development methodology and actively participated in daily scrum meetings.
- Worked with Hue GUI in scheduling jobs with ease and File browsing, Job browsing, Metastore management.
- Used Maven for Project building and management.
Environment: Hadoop, Map Reduce, HDFS, Hive, Hue, Pig, HBase, Teradata, Oracle 11g/10g, MongoDB, Oozie, Java (jdk1.6), UNIX, SVN and Zookeeper, Maven.
Confidential, Detroit, MI
Sr. Hadoop Developer
Responsibilities:
- Work with the Teradata analysis team to gather the business requirements.
- Worked extensively on importing data using scoop and flume.
- Responsible for creating complex tables using hive and developing Hive queries for the analysts.
- Created partitioned tables in Hive for best performance and faster querying.
- Transportation of data to HBase using pig.
- Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
- Experience with professional software engineering practices and best practices for the full software development life cycle including coding standards, code reviews, source control management and build processes.
- Worked collaboratively with all levels of business stakeholders to architect, implement and test Big Data based analytical solution from disparate sources
- Involved in source system analysis, data analysis, data modeling to ETL
- Written multiple MapReduce procedures to power data for extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV & other compressed file formats.
- Handling structured and unstructured data and applying ETL processes.
- Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS
- Developed the Pig UDF'S to pre-process the data for analysis.
- Prepare Developer (Unit) Test cases and execute Developer Testing.
- Create/Modify shell scripts for scheduling various data cleansing scripts and ETL loading process.
- Supports and assist QA Engineers in understanding, testing and troubleshooting.
- Written build scripts using ant and participated in the deployment of one or more production systems
- Production Rollout Support which includes monitoring the solution post go-live and resolving any issues that are discovered by the client and client services teams.
- Designed, documented operational problems by following standards and procedures using a software reporting tool JIRA.
- Involved in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries and Pig scripts
- Assisted in Cluster maintenance, Cluster Monitoring and Troubleshooting, Manage and review data backups and log files
- Participated in requirement gathering form the Experts and Business Partners and converting the requirements into technical specifications
Environment: Apache Hadoop, Sqoop, Hive, Oozie, Java (jdk1.6), Flat files, Oracle 11g/10g, MySQL, Windows NT, UNIX, Zoo Keeper, Cloudera, FLUME, CentOS, Maven.
Confidential, Dublin, OH
Sr. Hadoop Developer
Responsibilities:
- Identifying, Designing & developing the POC solution
- Implementation of various Mapreduce Jobs and configuring Oozie appropriately to manage the flow
- POC implemented on Amazon S3 cloud servers
- Writing and executing queries in Hive and Using Pig to read the content from Content Management System.
- Giving demo of Developed Applications to different project managers and clients
- Creating High Level Design, Functional Design and Technical Design documents as per the requirement.
- Training the team on project architectures and functioning of the software applications, Assuring better
Environment: Hive, Java/J2EE, Pig, Hbase, Hadoop Framework
Confidential, Kansas City, MO
Sr. Java/J2EE Consultant
Responsibilities:
- Involved in Design the Project Structure, System Design and Every phase in the project.
- Responsible for developing Platform related logic and Resource classes, controller classes to access the Domain and Service classes.
- Involved in the Development of Interfaces and services for the App Builder functionality.
- Involved in technical discussions, design and workflow.
- Participate in the requirement gathering and analysis.
- Used Jaxb to unmarshall XML into Java Objects.
- Developed Unit test cases using Junit Framework.
- Implemented the DataAccess Using Hibernate and Wrote the Domain Classes to generate the database tables.
- Designed and developed the communication tier to exchange data to Xpress Services through JMS and XML over HTTP.
- Involved in implementation of view pages based on XML Attributes using normal java classes.
- Involved in Integration of App builder and UI modules with the platform.
- Used GitHub as a code repository.
- Used Gradle as a build tool.
- Implemented JQuery and Ajax for form submissions and design.
- Developed Maven scripts to build the application.
- Executed unit test cases using JUnit and Mock Objects.
Environment: Jaxb, Junit Framework, XML, GitHub, Gradle, JQuery, Ajax, Maven, Mock Objects, Junit.
Confidential, Houston, TX
Java/J2EE Developer
Responsibilities:
- Involved in the process Design, Coding and Testing phases of the software development cycle.
- Designed use-case, sequence and class diagram (UML).
- Developed rich web user interfaces using JavaScript (pre-developed library).
- Created modules in Java and C++, python.
- Developed JSP pages with Struts framework, Custom tags and JSTL.
- Developed Servlets, JSP pages, Beans, JavaScript and worked on integration.
- Developed SOAP/WSDL interface to exchange usage and Image and terrain information from Geomaps.
- Developed Unit test cases for the classes using JUnit.
- Developed stored procedures to extract data from Oracle database.
- Developed and maintained Ant Scripts for the build purposes on testing and production environments.
- Designed and developed user interface components using AJAX, JQuery, JSON, JSP, JSTL & Custom Tag library.
- Involved in building and parsing XML documents using SAX parser.
- Application developed with strict adherence to J2EE best practices.
Environment: Java, C++, Python, Ajax, JavaScript, Struts, Spring, Hibernate, SQL/PLSQL, Web Services, WSDL, Linux, Unix.
Confidential
Java Programmer
Responsibilities:
- Involved in Requirements gathering, Requirement analysis, Design, Development, Integration and Deployment.
- Involved in Tax module and Order Placement / Order Processing module.
- Responsible for the design and development of the application framework
- Designed and Developed UI’s using JSP by following MVC architecture.
- Developed the application using Struts framework. The views are programmed using JSP pages with the struts tag library, Model is the combination of EJB’s and Java classes and web implementation controllers are Servlets.
- Used EJB as a middleware in designing and developing a three-tier distributed application.
- The Java Message Service (JMS) API is used to allow application components to create, send, receive, and read messages.
- Used JUnit for unit testing of the system and Log4J for logging.
- Created and maintained data using Oracle database and used JDBC for database connectivity.
- Created and implemented Oracle stored procedures and triggers.
- Installed WebLogic Server for handling HTTP Request/Response. The request and response from the client are controlled using Session Tracking in JSP.
- Reported Daily about the team progress to the Project Manager and Team Lead.
Environment: Core Java, J2EE 1.3, JSP 1.2, Servlets 2.3, EJB 2.0, Struts 1.1, JNDI 1.2, JDBC 2.1, Oracle 8i, UML, DAO, JMS, XML, WebLogic 7.0, MVC Design Pattern, Eclipse 2.1, Log4j and JUnit.