Hadoop Developer Resume
Charlotte, NC
PROFESSIONAL SUMMARY:
- 6+ years of experience in Software Development Life Cycle process which includes analysis, design, development and maintenance of Hadoop and Java/ J2EE based applications.
- 3+ years of experience in Hadoop development.
- Proficient in development methodologies such as Agile, Scrum and Waterfall
- Proven ability to excel in fast paced development environment using latest frameworks/tools.(Struts, Springs, Hibernate, Web services, Jira, Bugzilla, Quality Centre, JUnit, Mockito, SVN, CVS, VSS, IBM Clear case, IBM clear quest, GIT Hub, SOAP GUI, LDAP )
- Extensive experience of development using Hadoop ecosystem covering Map Reduce, HDFS, YARN, Hive, Pig, Hbase, Apache Kafka, Spring XD framework, Sqoop, Oozie, Splunk, Perforce, IBM Big insights, Cloudera.
- Experience in working with various Cloudera distributions (CDH4/CDH5) and have knowledge on Hortonworks
- Hands on experience in writing MapReduce jobs using Java.
- Expert level understanding of Hadoop HDFS and Map Reduce internals.
- Experience in working with NoSQL database HBase.
- Expertise in working with RDBMS databases like Oracle, MSSQL and DB2.
- Experience in strong and analyzing data using HiveQL, Pig Latin, HBase and custom Map Reduce programs in Java.
- Worked on designing and implementing complete end - to-end Hadoop Infrastructure including PIG, HIVE, Sqoop, Oozie.
- Performed Importing and exporting data into HDFS and Hive using Sqoop.
- Experience in writing shell scripts to dump the Shared data from MySQL servers to HDFS.
- Experience in developing PIG queries and writing custom UDF’s.
- Experienced in creating Avro, ORC, RC, Text files.
- Experienced in creating views on top of hive ORC tables.
- Manage and review Hadoop log file and Knowledge on Spark & Storm
- Extensive Experience in Design and Building Complex Business components, Developing Multi-Tier Web Applications and Server Side Business Logic using J2EE, UML, Java JDK 1.4/1.5/1.6 Servlets, JSP and Tag Library, Struts, Spring, Hibernate, Restful Web Services, Ant, Web logic 10 and Web sphere 6.1, Apache Tomcat 6.0.
- Experience in developing Front-End using JavaScript, HTML, XML and CSS
- Experience in designing highly transactional websites using J2EE technologies and handling design/implementation-using RAD/Eclipse/WSAD
- Experience in designing Low Level and High Level design documents
- Good exposure to MVC and design patterns
- Experience in working with multi/ multiple Operating System like Windows, Linux and strong knowledge with troubleshooting, finding and fixing critical problems
- Experience in Project Coordination along with technical development with strong communication and interpersonal competencies
TECHNICAL SKILLS:
Operating Systems: Windows, Linux & Unix
Languages: Jdk 1.4/1.5/1.6/1.7 , JavaScript, Servlet, PL/SQL, Unix Shell
Hadoop Distribution: Apache, CDH, Pivotal
Big Data Technologies: Apache Hadoop (MRv1, MRv2), Hive, Pig, Sqoop, HBase, MongoDB, Flume, Zookeeper, Oozie, Cassandra
Web Technologies: HTML, JSP, JSF, CSS, JavaScript, JSON & AJAX
ServerSide Frameworks: Struts 2, Spring, Active MQ, Apache Camel, Hadoop
IDEs: Eclipse, JBOSS, IBM Web Sphere
Build Tools: Maven, Ant, Hudson
Web Services: SOAP & RESTful Web Services
Web Servers /App Servers: Apache Tomcat 6.0/7.0, IBM WebSphere 6.0/7.0, JBoss 4.3
CI Tools: Hudson/Jenkins
Production Support Tools: iDesk, TSG Support Central
Static Code Analysis: Find Bug/CheckStyle/PMD
Configuration Tool: SVN, GitHub, CVS, VSS, IBM Rational Clear Case 7.0, IBM Rational Clear Quest 7.0, Demandware, SOAPUI
Database: Oracle 8i/9i/10g/11g, MySQL, HDFS, MongoDB(NoSQL), LDAP
Cloud Solution: Amazon Web Services (AWS), Qubole
Defects Triage: Quality Center 9.2, BugZilla, JIRA
PROFESSIONAL EXPERIENCE:
Confidential, Charlotte, NC
Hadoop Developer
Responsibilities:
- Designed and developed data movement framework for multiple sources like Sql Server,Oracle, and MySQL.
- Created Sqoop import and export jobs for multiple sources.
- Developed scripts to automate the creation Sqoop jobs for various workflows.
- Developed Hive scripts to alter the tables and perform required transformations.
- Developed a java Map Reduce and PIG cleansers for data cleansing
- Worked on Hive UDFS to mask confidential information in the data.
- Designed and developed Map Reduce programs for data lineage.
- Designed and Developed HBase Sink Handler module in the data ingestion framework. HBase Sink Handler ingests data from various input sources into target HBase tables.
- Writing the Hadoop Job workflows & scheduling using Oozie
- Contributed to design and performance testing of various HBase Schemas for the metrics tables. Languages/Tools/Frameworks: Java, Python, Hive, Pig, Oozie, Hbase
- Designed and developed the framework to log information for auditing and failure recovery.
- Closed worked with the web application development team to develop the user interface for data movement framework.
- Designed Oozie workflows for job automation.
- Experience in troubleshooting in MapReduce jobs by reviewing log files.
- Optimized our Hadoop infrastructure at both the Software and Hardware level
- Ensured our Hadoop clusters are built and tuned in the most optimal way to support the activities of our Big Data teams.
- Developed MapReduce programs to extract and transform the data sets and results were exported back to RDBMS using Sqoop.
- Maintaining the MySQL server and Authentication to required users for Databases. Appropriately documented various Administrative & technical issues.
- Running Cron-tab to back up Data. Applied Operating System updates, patches and configuration changes.
- Experience in collecting metrics for Hadoop clusters using Ganglia and Ambari
Environment: Cloudera Manager, HBase, flume, pig, Linux, Oozie, MapReduce, Sqoop, SQL Server, MySQL
Confidential, Chesapeake, VA
Hadoop Developer
Responsibilities:
- Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
- Developed the Pig UDF’S to pre-process the data for analysis.
- Develop Hive queries for the analysts.
- Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig.
- Generated reports using Pentaho report designer.
- Cluster co-ordination services through ZooKeeper.
- Job management using Fair Scheduler.
- Managed and reviewed Hadoop log files.
- Installed Oozie workflow engine to run multiple Hive and Pig jobs.
- Developed Hive queries to process the data and generate the data cubes for visualizing.
- Hands on experience in installing, configuring and using ecosystem components like Hadoop MapReduce, HDFS, HBase, Zoo Keeper, Oozie, Hive, Cassandra, Sqoop, Pig, Flume, Avro and Horton Works Talend.
- Experience in deploying applications in heterogeneous Application Servers TOMCAT, WebLogic, IBM WebSphere and Oracle Application Server.
- Monitored multiple Hadoop clusters environments using Ganglia and Nagios. Monitored workload, job performance and capacity planning using Ambari
Environment: MapReduce, HDFS, Hive, Java (jdk1.6), Pig, Oozie, Pentaho, Tomcat, WebLogic IBM WebSphere, Oracle Application Server, Apache Phoenix
Confidential
Hadoop Developer
Responsibilities:
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Configured Sqoop Jobs to import data from RDBMS into HDFS using Oozie workflows.
- Involved in creating Hive Internal and External tables, loading data and writing hive queries, which will run internally in map, reduce way.
- Involved in Migrating the Hive queries to Impala.
- Created batch analysis job prototypes using Hadoop, Pig, Oozie, Hue and Hive.
- Assisted with data capacity planning and node forecasting.
- Integrated Oozie with the rest of the Hadoop stack supporting several types of Hadoop jobs out of the box (such as Map-Reduce, Pig, Hive, and Sqoop) as well as system specific jobs (such as Java programs and shell scripts).
- Involved in Analyzing system failures, identifying root causes, and recommended course of actions.
- Documented the systems processes and procedures for future s.
- Monitored workload, job performance and capacity planning using Cloudera Manager.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Monitoring, Performance tuning of Hadoop clusters, Screening Hadoop cluster job performances and capacity planning Monitor Hadoop cluster connectivity and security Manage and review Hadoop log files.
- Load and transform large sets of structured, semi structured and unstructured data.
Environment: Hadoop, MapReduce, HDFS, Hive, Java, SQL, Cloudera Manager, Pig, Sqoop, Oozie, Hadoop, HDFS, Map Reduce, Hive, HBase, Linux, Cluster Management
Confidential
Java Developer/ Web Developer
Responsibilities:
- Coded the business methods according to the IBM Rational Rose UML model.
- Implement the code as per design. Developed the front-ends in JSP, HTML and JavaScript. Created Java Beans to capture business logic. Written SQL scripts. Debug the implemented code as per the code design.
- Develop the site and integrating the dynamic pages with the static designs
- Extensively used Core Java, Servlets, JSP and XML.
- Used Struts 1.2 in presentation tier.
- Generated the Hibernate XML and Java Mappings for the schemas.
- Used DB2 Database to store the system data
- Used Rational Application Developer (RAD) as Integrated Development Environment (IDE)
- Used unit testing for all the components using JUnit
- Used Apache log 4j Logging framework for logging of trace and Auditing
- Used Asynchronous JavaScript and XML (AJAX) for better and faster interactive Front-End
- Used IBM Web-Sphere as the Application Server
- Used IBM Rational Clearcase as the version controller
- Used Git for controlling the version of the project
Environment: Java 1.6, Servlets, JSP, Struts1.2, IBM Rational Application Developer (RAD) 6, WebSphere 6.0, iText, AJAX, Rational Clearcase, Rational Rose, Oracle 9i, log4j, Java Server Pages and XML, Java, J2EE, Java Helper Class, Java Servlets (Controller), JNI, JAXP, SQLJ, JavaScript
Confidential
Junior Java Developer
Responsibilities:
- Involved in Analysis, Design, Implementation, and Testing of the project.
- Implemented the presentation layer with HTML, XHTML, JavaScript, and CSS.
- Developed web components using JSP, Servlets and JDBC.
- Implemented database using SQL Server.
- Designed Tables and Indexes.
- Wrote complex T-SQL and Stored Procedures.
- Involved in fixing defects and unit testing with test cases using JUnit.
- Developed user and technical documentation.
Environment: Java, JSP, Servlets, JDBC, JavaScript, CSS, MySQL, JUnit, Eclipse, JBoss