Hadoop Developer Resume
PA
PROFESSIONAL SUMMARY:
- Over 8+ years of experience in IT and 4+ years of experience Hadoop ecosystem's implementation, maintenance, ETL and Big Data analysis operations.
- Excellent understanding / knowledge of BigData and Hadoop ecosystems.
- Experience with configuration of Hadoop ecosystem components such as MapReduce, HDFS, HIVE, HBase, PIG, Sqoop, Zookeeper, Oozie and Flume.
- Experienced in using NFS (network file systems) for Name node metadata backup.
- In depth knowledge of Hadoop architecture and various components such as HDFS, JobTracker, NameNode, DataNode and MapReduce concepts.
- Good Exposure to Map Reduce programming, PIG Scripting and Distributed Application and NoSQL Databases like HBase, MongoDB
- Experience in maintaining multiple Hadoop clusters (prod, dev etc.).
- Experience in performance tuning of the Hadoop Cluster by gathering and analyzing the existing infrastructure.
- Experience in automating the Hadoop installation, configuration and maintaining the cluster by using the tools like puppet.
- Experience in managing and monitoring Hadoop cluster using Cloudera Manager.
- Knowledge on designing both time driven and data driven automated workflows using Oozie.
- Excellent analytical, multi - tasking, problem-solving, time-management & communication skills with particular emphasis on clearly communicating & documenting detailed Test Requirements & Tests.
- Experience in support analysts by administering and configuring Hive.
- Experience in providing support to data analyst in running Pig and Hive queries.
- Extending Hive and Pig core functionality by writing custom UDFs.
- Developed MapReduce programs to perform analysis.
- Performed importing and exporting data into HDFS and Hive using Sqoop.
- Experience in writing shell scripts to dump the Shared Data from MySQL server to HDFS.
- Experience in Core Java, Java Virtual Machine (JVM) and multi-thread processing.
- Over 3 years of extensive experience as a Netcool Management Service (NMS) Consultant, involved in developing architectural solutions for customers in the areas of Fault Management, Performance Management, Operational Support Systems (OSS) and other related disciplines of included IBM Netcool.
- Responsible for architecting, implementing and administration of monitoring solutions using IBM Tivoli Product Suite Netcool/OMNIbus (Object Server, Probes, Gateways, Process Agents), Netcool/Webtop, Netcool/Impact, ITNM, TBSM, Tivoli Monitoring - ITM, Tivoli Common Reporting.
- Experience in Web based portal system such using Edge tool (JARVIS) that displays and consolidates web enabled network management applications into a single view.
- Experience in Unix Shell Scripting, SQL, Reporting and validating complex Stored Procedures, Triggers.
- Worked on UNIX platform and got a clear understanding of the VI editor and UNIX commands.
- Developed the User Interface using JSP/AJAX/ HTML / CSS/ Java Script
- Capable of writing programs in JDBC programs for callable statements, prepared statements written in Oracle andused TOAD toolto interact with database
- Excellent knowledge over the Restful API's like Elastic Search
- Excellent communication and inter-personal skills, self-motivated and detail oriented.
TECHNICAL SKILLS:
Operating Systems: Linux, Windows XP, Server 2003, Server 2008
Databases: Oracle, MySQL, MS SQL Server 2000, DB2
Hadoop Ecosystem Development: HDFS, HBase, MapReduce, YARN, Hive, Pig,Sqoop, Flume, Oozie, Zookeeper
Fault Management Tool: IBM Tivoli suite of products (OMNIbus, ITNM, Web-GUI, TBSM, TCR, Impact, ITM
Languages: C, Java, SQL, Pig Latin, UNIX Shell Scripting
IDE/ Tools: Eclipse
Others: JARVIS, HP-BSM
PROFESSIONAL EXPERIENCE:
Confidential, PA
Hadoop Developer
Responsibilities:
- Analyzing the requirement to setup a cluster.
- Installed and configured Hadoop, MapReduce, HDFS (Hadoop Distributed File System), developed multiple MapReduce jobs in java.
- Worked with the infrastructure and admin team in designing, modelling, sizing and configuring Hadoop cluster of 15 nodes.
- Developed Map Reduce programs in Java for parsing the raw data and populating staging Tables.
- Created Hive queries to compare the raw data with EDW tables and performing aggregates
- Importing and exporting data into HDFS and Hive using Sqoop.
- Experienced in analyzing data with Hive and Pig.
- Experienced knowledge over the Restful API's like Elastic Search.
- Writing Pig scripts to process the data.
- Developed PIG Latin scripts to extract the data from the web server output files to load into HDFS.
- Integrating bulk data into Cassandra file system using MapReduce programs.
- Got good experience with NOSQL database.
- Involved in HBASE setup and storing data into HBASE, which will be used for further analysis.
- Experienced in managing and reviewing Hadoop log files.
- Experienced in defining job flows.
- Experienced in managing and reviewing Hadoop log files.
- Installed and configured Hive and also written Hive UDFs.
- Involved in creating Hive tables, loading with data and writing hive queries using the HiveQL which will run internally in map reduce way.
- Extracted the data from MySQL into HDFS using Sqoop.
- Used HiveQL to analyse the partitioned and bucketed data and compute various metrics for reporting.
- Deployed Hadoop Cluster in Fully Distributed and Pseudo-distributed modes.
- Experience in managing and monitoring Hadoop cluster using Cloudera Manager.
- Supported in setting up QA environment and updating configurations for implementing scripts with Pig, Hive and Sqoop.
- Unit tested a sample of raw data and improved performance and turned over to production.
Environment: Java (JDK1.7), Java 7, Hadoop 2.6.0, MapReduce, HDFS, Hive 0.13.0, Sqoop 1.4.4, HBase, Pig 0.12.0, Oozie, Kerberos, Linux, Shell Scripting, Oracle 11g, PL/SQL, SQL*PLUS, HDInsight
Confidential
Hadoop Developer
Responsibilities:
- Installed and configured Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and pre-processing.
- Installed and Configured Apache Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase and Sqoop.
- Good understanding and related experience with Hadoop stack - internals, Hive, Pig and Map/Reduce.
- Wrote MapReduce jobs to discover trends in data usage by users.
- Involved in defining job flows.
- Involved in database connection by using SQOOP
- Involved in creating Hive tables, loading with data and writing hive queries using the HiveQL.
- Experienced in SQL DB Migration to HDFS
- Got good experience with NOSQL database.
- Used Oozie to orchestrate the workflow.
- Installed and configured Pig for ETL jobs.
- Used big data technologies with hands on experience on Hadoop Distributed File System.
- Worked on Map Reduce programming and Pig latin scripts.
- Experience with Core Distributed computing and Data Mining Library using Apache Spark.
- Integrating bulk data into Cassandra file system using MapReduce programs.
- Involved in creating data-models for customer data using Cassandra Query Language.
- The data is collected from distributed sources into Avro models. Applied transformations and standardizations and loaded into HBase for further data processing.
- Developed the custom writable JAVA programs to load the data into the HBase by using the Apache Crunch Java API to data pipelines.
Environment: Hadoop, Hive, Linux, MapReduce, HDFS, Hive, Python, Pig, Sqoop, Cloudera, Cassandra, Spark, Shell Scripting, Java (JDK 1.6), Java 6, Oracle 10g, PL/SQL, SQL*PLUS
Confidential, Oklahoma, OK
Hadoop Developer/ Administrator
Responsibilities:
- Gathered the business requirements from the Business Partners and Subject Matter experts.
- Involved in installing Hadoop ecosystem components.
- Responsible to manage data coming from different sources.
- Monitoring the jobs to analyze performance statistics.
- Performing Unit Testing of completed jobs.
- Involved in database connection by using SQOOP.
- Applying optimization techniques at both Hadoop and Database level.
- Involved in creating Hive tables, and loading and analyzing data using hive queries.
- Developed Simple to complex MapReduce Jobs using Hive and Pig.
- Involved in running Hadoop jobs for processing millions of records of text data.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Experienced in defining job flows.
- Experienced in managing and reviewing Hadoop log files.
- Extracted files from MongoDB through Sqoop and placed in HDFS and processed.
- Experienced in running Hadoop streaming jobs to process terabytes of XML format data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Automation script to monitor HDFS and HBase through cronjobs.
- Plan, design, and implement processing massive amounts of marketing information, complete with information enrichment, text analytics, and natural language processing.
- Prepare multi-cluster test harness to exercise the system for performance and failover.
Environment: Hadoop HDFS, Map Reduce, Hive, Pig, Sqoop, Oozie, HBase, Linux, Java, Xml, MongoDB.
Confidential, CA
Java Developer
Responsibilities:
- Actively participated in the Analysis, Design, Development and Testing phases of various modules.
- Scrupulously involved in Agile Standup meetings throughout the project with team members on day to day basis for the update on project status using Rally Software.
- Estimated Level1 and Level2 and Designed Class, Sequence Diagrams.
- Worked on J2EEdesign patternslike DAO, Front Controller, Session Façade, Factory Pattern and Singleton.
- Created mobile web applications using HTML5, CSS3, JSON, and JavaScript.
- Worked on OOPS concepts, Collections Framework using Core Java.
- Extensively used Spring MVC architecture for the application development.
- Developed application service components and configured beans using Spring IOC .
- Used JMS for Asynchronous communication by consuming messages in the Message queue.
- Used Hibernate as data persistence layer and configured service and persistence object using Hibernate mapping files.
- Involved in frameworks integration of Spring and Hibernate.
- Produced and Consumed REST based webservices, did unit testing using SOAPUI tool and generated json.
- Worked on Database queries using Oracle instance.
- Used JAXB to convert XML content into a Java Object.
- Involved in unit testing, test case reviews and code reviews in accordance with applicable development standards.
- Involved in integration system testing and user acceptance testing (UAT)
- Worked on Web deployments in DEV, QA and UAT environments.
- Support the Application whenever encountered Production issues.
- Deployed application in Weblogic Application server.
- SVN is used as Code Versioned control.
- Firebug is used as debugger.
- Used Maven script to build application.
Environment: Core Java1.6, Oracle10g, Toad, JSP, Spring3.0, Hibernate4.0, JavaScript, HTML5, jQuery, Ajax, Apache CXF Framework, REST Webservices, JAXB, Linux, Shell scripting, WinSCP, log4J, Maven, SVN, Weblogic Server11g, Eclipse, Microsoft Visio.
Confidential
Java Developer
Responsibilities:
- Participates in the full SDLC including collaborative review meetings, estimations, design sessions, development and testing.
- Developed the User Interface using JSP/AJAX/ HTML / CSS/ Java Script.
- Designed and developed Use Cases, Class diagrams and Sequence Diagram.
- Implementedvarious J2EEdesign patternslikeDAO, DTO, Factory Pattern and Singleton.
- Developed the application using Java Server Faces (JSF) and Spring MVC.
- Implement GET and POST operations using Restful web services.
- Totally involved in building components using DOM and JaxB which helps to parse the XML.
- Written JDBC programs for callable statements, prepared statements written in Oracle andused TOAD toolto interact with database.
- Designed and developed Hibernate Dao’s using spring framework’s Hibernate Dao Support & Hibernate Template
- Used CVS for version controlling.
- Used log4j for logging statements.
Environment: Core Java 1.5, JSF1.1, Spring 2.0, Restful webservices, Rest Client, XML, XSD, JAXB, DOM, Tomcat, JSP, Oracle 10g, SQL/PL SQL, Hibernate2.0, Java script, CSS, HTML, Ajax, log4j, My Eclipse, Toad, Ant1.7, Microsoft Visio.
Confidential
Java Developer
Responsibilities:
- Responsible for reviewing business user requirements and also participated in meeting the users with Business Analysts
- Design and developed applications with Eclipse in Java using a Struts 1.2 framework running on Weblogic application servers.
- Developed the User Interface using JSP/AJAX/ HTML / CSS/ Java Script
- Widely Used Design pattern like DAO, Singleton, Business delegate and Service Locator in the process of system designing and development
- Used Message Driven Beans and JMS to process the requests from the customer asynchronously
- Developed stored procedures, cursors and database Triggers and implemented Scrollable Result sets
- Consumed Web Services (WSDL, SOAP, and UDDI) from third party to verify the credit score of applicants
- Developed Web services using top-down approach and coded required WSDL files
- Used XSL/XSLT for transforming common XML format into displayable format
- Involved in testing the system using JUnit
- Maintained the source code versions in Subversion repository
- Used Log4J for logging and tracing the messages
- Deployed application in Wesphere Application Server and developed using RAD
Environment: Core Java 1.4, Websphere6.0, RSA6.0,SQLServer2005, Linux, Servlets 2.5, JSP 2.1, Struts1.2, EJB2.0, AJAX, HTML, JMS, JNDI, JAAS, JTA, XML, XSL, JAXB, SOAP, UDDI, WSDL, JUnit, Log4J, ANT, VSS