Hadoop Developer Resume
MI
SUMMARY:
- Around 8+ years of professional experience this includes Design, Development, Integration, Deployment and Maintenance of quality software applications in big data Hadoop technologies using Java/J2EE Technologies.
- Hands on experience in installing, configuring, monitoring and integration of Hadoop ecosystem components like MapReduce, HDFS, HBase, Pig, Hive, Oozie, Sqoop, Flume, Spark, kafka and Zookeeper.
- Experience working with NoSQL database including Mongo DB, Cassandra and Hbase.
- Experience in developing NoSQL database by using CRUD, Shading, Indexing and Replication.
- Exported the analyzed data to various Databases like Teradata (Sales Data Warehouse), SQL - Server using Sqoop.
- Developed stored procedures and queries using PL/SQL.
- Working with relative ease with different working strategies like Agile, Waterfall and Scrum methodologies.
- Hands-on programming experience in various technologies like JAVA, J2EE, HTML, XML.
- Hands on experience in application development using Java, RDBMS and Linux Shell Scripting.
- Developed enterprise applications using Scala.
- Has extensively worked in Java technologies such as Java Beans, Servlets, and JDBC.
- Hands on experience working with Java project build managers Apache MAVEN and ANT.
- Maintenance/ Implementation of Commercial Software. Extensive work experience with Java/J2EE technologies such as Servlets, JSP, EJB, JDBC, JSF, Struts, spring, SOA, AJAX, XML/XSL, Web Services (REST, SOAP), UML, Design Patterns and XML Schemas.
- Quick Learner with problem solving skills, willing to learn and implement new technologies
- Ability to work effectively in cross-functional team environments, excellent communication and interpersonal skills.
- Understand continuous integration tools and Java IDEs
TECHNICAL SKILLS:
Big Data Technologies: Hadoop, HDFS, Hive, MapReduce, Pig, Sqoop, Flume, Oozie, kafka Spark, Tableau
Programming Languages: C, C++, Java, Scala
Java/J2ee Technologies: Java, Java Beans, J2EE (JSP, Servlets, EJB), JDBC
Databases/ETL: Oracle 11g, SQL Server 2000, MySQL, SQL /PL SQL, Informatica v 8.x
Nosql Databases: HBase and Cassandra
Web Technologies: HTML, JSP, JavaScript, Ajax, XML, PHP, AWS
Servers: Web Sphere, Apache Web Serve, Tomcat Server 7.0
Methodologies: UML, OOP, OOA, OOD and Agile
Version Controls: Tortoise CVS Client, SVN
Mapping Tools: Hibernate 3.0
Operating Systems: LINUX (Centos and Ubuntu), Windows XP, 7, MS DOS, UNIX
Scripting Languages: Perl, Python, Shell scripts
Build Tools: Ant, Maven 2.2.1
IDE/Tools/Utilities: Eclipse Helios, MS Visio, MS Office 2010, Control M, SQL Programmer
PROFESSIONAL EXPERIENCE:
Confidential, MI
Hadoop Developer
Roles & Responsibilities:
- Developed MapReduce and Hive jobs for analyzing large and critical datasets.
- Imported data into HDFS from various SQL databases and files using Sqoop and from streaming systems using Storm into Big Data Lake.
- Collected and aggregated large amounts of log data using Apache Flume and staged data in HDFS for further analysis.
- Developed custom aggregate functions using Spark SQL and performed interactive querying.
- Worked with HBase databases to create tables and store the processed data.
- Wrote Pig scripts to store the data into HBase.
- Created Hive tables, dynamic partitions, buckets for sampling, and worked on them using Hive QL.
- Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
- Involved in creating calculated fields and dashboards in Tableau for visualization of the analyzed data.
- Worked with using different kind of compression techniques to save data and optimize data transfer over network using Snappy.
- Configured deployed and maintained multi-node Dev and Test Kafka Clusters
- Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processed the data with Pig.
- Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, manage and review data backups, manage and review Hadoop log files.
Environment: Hadoop, HDFS, Hive, Pig, Sqoop, Flume, Eclipse, Data Lake, Spark, MapReduce, Cloudera, Tableau, Snappy, Zookeeper, HBase, Scala 2.10/2.11, Windows 7/Vista/ XP, Linux, Unix, NoSQL, MySQL, Shell Scripting, Ubuntu, Teradata.
Confidential, Baltimore, MD
Hadoop Developer
Roles & Responsibilities:
- Primary responsibilities include building scalable distributed data solutions using Hadoop ecosystem.
- Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
- Developed Simple to complex Map/reduce streaming jobs using Python language that are implemented using Hive and Pig.
- Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Handled importing of data from various data sources, performed transformations using Hive, Map Reduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop.
- Analyzed the data by performing Hive queries (Hive QL) and running Pig scripts (Pig Latin) to study customer behavior.
- Tested Apache(TM) Tez, an extensible framework for building high performance batch and interactive data processing applications, on Pig and Hive jobs.
- Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources.
- Continuous monitoring and managing the Hadoop cluster using Cloudera Manager.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
- Used Mahout to understand the machine learning algorithms for an efficient data processing.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Experienced on loading and transforming of large sets of structured, semi structured and unstructured data.
- Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
- Created AWS Route53 to route traffic between different regions.
- Written multiple Map Reduce programs in Java for data extraction, transformation and aggregation from multiple file formats including XML, JSON, CSV and other compressed file formats.
Environment: Hadoop, PIG, Hive, Apache Sqoop, Oozie, HBase, Zoo keeper, Cloudera manager, JDK1.6, HDFS, Map-Reduce, Apache Crunch, Unix, Linux, SQL server, Flume, Spark, lambda, DB2, HBase and Python, Windows 7/XP.
Confidential, Atlanta, GA
Hadoop Developer
Roles & Responsibilities:
- Involved in capacity planning architecture of the Big Data platform.
- Developed Map-Reduce programs in Java for data cleaning and pre-processing.
- Created MapReduce jobs for data transformations and data parsing.
- Created Hive scripts for extracting the summarized information from hive tables.
- Created Hive UDFS to extract data from staging tables.
- Involved in creating Hive tables, loading data and querying data.
- Design and implementation of High Availability feature for Search Engine.
- Volume testing to calculate cluster's throughput.
- Helped the team to increase the Cluster size from 22 to 30 Nodes.
- Solution architecture for proposals in Web technologies, log file analysis of storage equipment data.
- Maintain System Integrity of all sub-components (primarily HDFS, MR).
- Monitor System health, logs and respond accordingly to any warnings or failures.
- Unit testing, Volume testing and Bug fixing.
- Co-ordination with Client and offshore counterparts.
Environment: JDK1.6, CentOS, FLUME, HBase, Maven, Map-Reduce, Hadoop, Hive, Pig, Sqoop, Oozie, Zookeeper, SQL Server, Oracle, Windows 7/Vista, AWS,UNIX, Linux.
Confidential
Java Developer
Roles & Responsibilities:
- Using OOAD Technology classes are designed in UML with the help of Rational Rose tool.
- Created user-friendly GUI interface and Web pages using HTML and DHTML embedded in JSP.
- JavaScript was used for the client side validations.
- Designing and developing generic validator framework for modules and injecting these validators using hibernate framework.
- Creating Hibernate POJOs, Hibernate mapping files for all database tables.
- Developing GUI Screens using JSF (IBM Implementation) and for Ajax functionality.
- Developed and deployed EJB's (Session and Entity) to implement the business logic and to handle various interactions with the database.
- Involved in debugging the application.
- Developed Servlets using JDBC for storing and retrieving user data into and from the SQL database.
- Used Web Logic Application Server to deliver a new class of enterprise applications that enhance business interactions and transactions between a company and its key constituencies.
- Used Web Logic Application Server to deliver high performance and scalability.
- Written Database objects like Triggers, Stored procedures in SQL.
- Interacted with the users and documented the System.
- Used HP QA to manage the defects and issues.
Environment: JSP 2.0, JDBC, HTML, OOAD, Servlets, Web Services, Rational Rose, WSAD 5.0, UML, Java 1.6, EJB 2.0, JSF, QA, Hibernate, AJAX, Windows 7/XP, CVS, XML/XSL.
Confidential
Java Developer
Roles & Responsibilities:
- Worked as software developer for Confidential on developing a supply chain management system.
- The application involved tracking invoices, raw materials and finished products.
- Gathered user requirements and specifications.
- Developed the entire application on Eclipse IDE.
- Developed and programmed the required classes in Java to support the User account module.
- Used HTML, JSP and JavaScript for designing the front end user interface.
- Implemented error checking/validation on the Java Server Pages using JavaScript.
- Developed Servlets to handle the requests, perform server side validation and generate result for user.
- Used JDBC interface to connect to database.
- Used SQL to access data from Microsoft SQL Server database.
- Performed User Acceptance Test.
- Deployed and tested the web application on Web Logic application server.
Environment: JDK 1.4, Servlet 2.3, JSP 1.2, JavaScript, HTML, JDBC 2.1, SQL, MySQL Server, UNIX and BEA Web Logic Application Server.