Hadoop Developer Resume
Phoenix, AZ
PROFESSIONAL SUMMARY:
- Around 7 years of Experience in Enterprise Application Development, Web Applications, Client - Server Technologies using various languages and tools like Java, J2EE, JSP and Servlets.
- 3+ years of Experience in design, development, maintenance and support of Big Data Analytics using Hadoop Ecosystem tools like HDFS, Hive, Sqoop and Pig.
- Experience working Agile Environment and Participation in Scrum meetings
- Experienced in processing Big data on the Apache Hadoop(Yarn) framework using MapReduce programs.
- Excellent understanding and knowledge of NOSQL databases like HBase and Mongo DB.
- Good knowledge of Hadoop ecosystem, HDFS, Big Data, RDBMS.
- Experienced in installation, configuration, supporting and monitoring Hadoop clusters using Apache, Cloudera distributions and AWS.
- Good knowledge on Hadoop, HBase, Hive, Pig Latin Scripts, MR, Sqoop, Flume, Hive QL.
- Experience in analyzing data using Pig Latin, HiveQL and HBase.
- Capturing data from existing databases that provide SQL interfaces using Sqoop.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Implemented Proofs of Concept on Hadoop stack and different big data analytic tools, migration from different databases (i.e Teradata, Oracle, MYSQL) to Hadoop.
- Worked on NoSQL databases including HBase, Cassandra and MongoDB
- Successfully loaded files to Hive and HDFS from MongoDB, HBase
- Experience in configuring Hadoop Clusters and HDFS.
- Worked extensively in Java, J2EE, XML, XSL, EJB, JSP, JSF, JDBC, MVC, Jakarta struts, JSTL, Spring2.0, Design Patterns and UML.
- Extensive experience in Object Oriented Programming, using Java & J2EE (Servlets, JSP, Java Beans, EJB, JDBC, RMI, XML, JMS, Web Services, AJAX).
- Excellent analytical and problem solving skills and ability to quickly learn new technologies.
- Good communication and interpersonal skills. A very good team player with the ability to work independently.
- Experience working with various Hadoop distributions like Cloudera and Horton works (HDP)
- Experience working in Agile/Scrum Environment.
- Experience with Amazon Web Services like S3, EC2, Lambda, Dynamo DB and Redshift
- Experience using Sequence files, AVRO , RCFile and HAR file formats
- Excellent understanding of Apache Storm Architecture and topologies
- Very familiar with Play 2.1, including the Iteratee / Enumeratee streaming patterns in Scala
- Comfortable with functional programming idioms in Scala
- Familiar with Akka Concurrency Collections Closures Traits
- Extensive experience of designing and developing Essbase applications with balance of Dense and Sparse Dimensions.
- Experience coding Map Reduce jobs in Scala on Spark
- Experience working with Cubes and Druid
- Excellent understanding on hosting Big Data services on Microsoft Azure and Amazon Web Services .
- Experience working with Jenkins plugin for Hadoop with Namenode, Jobtracker as Masters and Datanode, Tasktracker as Slave nodes.
- Experience working with Apache Apex for real-time stream and batch processing for Big Data
- Worked on Kafka and streaming processing to ensure the data ingested using both streaming and batch ways.
- Had good experience working on Datameer Cloud Platform for Data Analytics and visualization purpose
TECHNICAL SKILL:
Hadoop/Big Data: HDFS, MapReduce, Hive, Pig, Sqoop, Flume, Oozie, Ambari, Cloudera Manager, Zookeeper, Apache Storm and TEZ, Horton Works, Apache Kafka, Apache Storm, Apache Phoenix.
No SQL Databases: HBase, Cassandra, MongoDB, DynamoDB
Languages: C, Python, Java, J2EE, PL/SQL, Pig Latin, HiveQL, Unix shell scripts, R Programming, Scala, Python, HBase, Scala
Java/J2EE Technologies: Applets, Swing, JDBC, JNDI, JSON, JSTL, RMI, JMS, Java Script, JSP, Servlets, EJB, JSF, JQuery, RESTful APIs, OLAP cubes
Frameworks: MVC, Struts, Spring, Hibernate, mrjob, MRUnit
Operating Systems: Sun Solaris, HP-UNIX, RedHat Linux, Ubuntu Linux and Windows XP/Vista/7/8, CentOS, Ubuntu
Web Technologies: HTML, DHTML, XML, Node.js, WSDL, SOAP
Web/Application servers: Apache Tomcat, WebLogic, JBoss
Databases: Oracle 9i/10g/11g, DB2, SQL Server, MySQL, Teradata
Tools: and IDE: Eclipse, NetBeans, Toad, Maven, ANT, Hudson, Sonar, JDeveloper, Assent PMD, DB Visualizer, Spark, QlikView, Kibana, Tableau, Nagios, Ganglia, Akka, Amazon Redshift, Datameer
Network Protocols: TCP/IP, UDP, HTTP, DNS, DHCP
PROFESSIONAL EXPERIENCE:
Confidential, Phoenix, AZ
Hadoop Developer
Responsibilities:
- Developing parser and loader map reduce application to retrieve data from HDFS and store to HBase and Hive.
- Experience using Jenkins Plugin to Deploying all the jar files need to run Hadoop in a package.
- Involved in Agile SDLC- interacting daily with product owners and scrum team members,
- Importing the unstructured data into the HDFS using Flume.
- Used Oozie to orchestrate the map reduce jobs that extract the data on a timely manner.
- Written Map Reduce java programs to analyze the log data for large-scale data sets.
- Involved in using HBase Java API on Java application.
- Automated all the jobs for extracting the data from different Data Sources like MySQL to pushing the result set data to Hadoop version 2.1.7 (Yarn) Distributed File System using Oozie Workflow Scheduler.
- Implemented Map Reduce jobs using Java API and Python using Spark
- Participated in the setup and deployment of Hadoop cluster
- Hands on design and development of an application using Hive (UDF).
- Responsible for writing Hive Queries for analyzing data in Hive warehouse using Hive Query Language (HQL).
- Provide support data analysts in running Pig and Hive jobs on TEZ.
- Involved in HiveQL and Involved in Pig Latin.
- Importing and exporting Data from MYSQL/Oracle to HIVEQL Using SQOOP.
- Configured Highly Available cluster for both Manual failover and Automatic failover.
- Continuous monitoring and managing the Hadoop cluster through Ganglia and Nagios.
- Designed and built many applications to deal with vast amounts of data flowing through multiple Hadoop clusters, using Pig Latin and Java-based map-reduce.
- Specifying the cluster size, allocating Resource pool, Distribution of Hadoop version 2.7.1(Yarn) by writing the specification texts in JSON File format.
- Experience in writing SOLR queries for various search documents
- Responsible for defining the data flow within Hadoop eco system and direct the team in implement them.
- Involved in a version upgrade of Hadoop Components
- Experience Integrating Apache Storm with Yarn (Hadoop 2.0) using Apache Slider
- Involved in using Apache Storm for processing millions of records of streaming data from the cluster.
- Experience working with Amazon Redshift Data Warehouse Service.
- Configured the Security using Kerberos
- Continuous monitoring and managing the Hadoop cluster through Apache Ambari
- Created Map Reduce jobs to load JSON data and server data into MongoDB and transported MongoDB into HDFS
- Experience developing fault tolerant scalable clusters on commodity hardware using KAFKA
- Involved working with Hortonworks Data Platform (HDP)
- Experience working with Apache Phoenix for OLTP and operational analytics on Hadoop.
- Have done optimization of Star Schema and Interleaved Sorting using Amazon RedShift Data Warehouse.
- Experience working on Datameer to deliver optimal performance for our analytic workloads and produce answers faster .
Environment: Linux, Apache Ambari, Big Data Apache Hadoop, Hive, Hue Tool, Zookeeper, Map Reduce, Sqoop, Nagios, crunch API, Pig 0.10 and 0.11, HCatalog, Unix, Java, JSP, Eclipse, Maven, Oracle, SQL Server, MYSQL, Oozie, Python 3.4, TEZ, Datameer.
Confidential, Atlanta, GAHadoop Developer
Responsibilities:
- Processed Big Data using a Hadoop cluster consisting of 40 nodes on Amazon EC2.
- Designed and configured Flume servers to collect data from the network proxy servers and store to HDFS.
- Loaded the customer profiles data, customer spending data, credit from legacy warehouses onto HDFS using Sqoop.
- Built data pipeline using Pig and Java/Scala Map Reduce to store onto HDFS.
- Applied transformations and filtered both traffic using Pig.
- Used Pattern matching algorithms to recognize the customer across different sources and built risk profiles for each customer using Hive and stored the results in DynamoDB.
- Performed unit testing using MRUnit.
- Responsible for building scalable distributed data solutions using Hadoop
- Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster
- Setup and benchmarked Hadoop/DynamoDB clusters for internal use
- Developed Simple to complex Map/reduce Jobs using Scala in EMR.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop
- Created several Hive unit tests using JUnit and executed these tests in Hive
- Analyzed the data by performing Hive queries and running Pig scripts to study employee behavior
- Installed Oozie workflow engine to run multiple Hive and Pig jobs
- Involved in managing and reviewing the Hadoop log files
- Continuous monitoring and managing the Hadoop cluster through Cloudera Manager.
- Involved in migration of cluster from one data center to another
- Implemented and configured FAIR Scheduler
- Configured Security using Apache Sentry
- Involved working with AWS components like EC2, EMR, S3 etc.
- Developed a Oozie workflow for Scheduling and Orchestrating the ETL Process
- Experience working on AT&T threat management system.
- Worked on 24x7 threat monitoring coverage from attackers, cyber-hackers and organized crimes.
- Worked on Cloudera platform Developed data pipeline using Flume and Sqoop to ingest customer behavioral data and financial histories from traditional databases into HDFS for analysis.
- Implemented design patterns in Scala for the application.
- Experience designing high volumes of write using fan-out Architecture in KAFKA
- Created and tested several java classes in JUnit4
- Experience working on JMS (Java Messaging Service) for solving producer -consumer problems.
Environment: Hadoop, Hive, Zookeeper, Map Reduce, Sqoop, Pig 0.10 and 0.11, JDK1.6, HDFS, Flume, Oozie, DB2, DynamoDB, AWS, Mahout, Scala, Cloudera Manager, Apache Sentry, Tableau.
Confidential, Hartford, CTSr Java Developer
Responsibilities:
- Involved in defining the business rule according to client specific and convert them into High Level Technical Design.
- Performed High Performance Market Data Distribution around a large network using ActiveMQ
- Designed entire system according to OOPS & UML by using Rational Tools.
- Elaborated use cases, interface definition specifications in collaboration with Business.
- Converted handful EJB services to RESTful Web services using JAX-RS API.
- Used Backend as the Oracle database & used JDBC technologies for integration.
- Experience Performing cross-language, cross-platform connectivity using ActiveMQ
- Developed persistence module with Hibernate, Spring with Oracle.
- Extensively used TOAD for all DB related activities & integration testing.
- Used build and deploy scripts in ANT and UNIX shell scripting.
- Developed User interface screens using Servlets, JSP, JavaScript(Node.js), CSS, HTML.
- Involved in unit testing of developed business units & used the JUnit for specifics.
- Worked along with the Development team & QA team to resolve the issues in SIT/UAT/Production environments.
- Developed Controller Classes, Command Objects, Action Classes, Form beans, Transfer Objects Singleton at server side for handling requests and responses from presentation Layer.
- Used Connection Pooling to get JDBC connection and access database procedures.
- Used log4j as a debugging tool and Involved in Java, J2EE coding.
- Involved in preparing Hyperion reports to measure financial performance, analyze profitability, and facilitate effective business decision making.
- Involved in installation and up-gradation of Hyperion Essbase family products.
Environment: Core Java, J2EE1.5/1.6, Struts, Node.js, Rational Rose, Rational Requisite Pro, Hibernate3.0, Spring, CVS, RAD7.0 IDE, Oracle10g, RESTful, log4j, WebSphere6.0, Servlets, JSP, Junit, Hyperion Essbase 11.x/9.x/7.x/6.x, Hyperion Essbase Integration Services, Hyperion Financial Reporting, Hyperion Web Analysis .
Confidential, Irving, TX
Java Developer
Responsibilities:
- Involved in program setup, program profile, fees, card settings Modules.
- Assist in the monthly close process including preparation of JEs, account reconciliations, accrual schedules, and revenue adjustments with the help of Oracle BRM
- Developed Action classes, business classes, helper classes, Hibernate POJO classes.
- Developed spring DAO classes, store proc. classes to connect the DB through spring JDBC
- Developed Action Forms, Form Beans and Java Action classes using Struts framework
- Participated in code reviews and ensured compliance with standards
- Involved in preparing database scripts and deployment process
- Used JDBC API to connect to the database and carry out database operations.
- Involved in design and implementation of web tier using Servlets and JSP.
- Used Apache POI for Excel files reading.
- Used JSP and JSTL Tag Libraries for developing User Interface components.
- Involved in developing UML Diagrams like Use Case, Class, Sequence diagrams.
- Handled the session management to switch from classic application to new wizard and vice versa.
- Involved in using RESTful API, NodeJS and JavaScript in development of the application
Environment: JAVA, J2EE (JSP, JSTL, Servlets), Hibernate, Struts (Validation Framework, Tiles Framework), Spring, Apache JQuery, JavaScript, XMP-RPC, SQL, Tortoise SVN, Maven, VISIO.