We provide IT Staff Augmentation Services!

Hadoop Developer Resume

5.00/5 (Submit Your Rating)

Portland, OregoN

PROFESSIONAL SUMMARY:

  • 7+ years of overall IT experience, with specialization in Hadoop andJava development.
  • Experience in installing, configuring, designing, implementing, testing Big data projects using Hadoop Ecosystem like HDFS, Map Reduce, Hive, Pig, Sqoop,Oozie, Flume, HBase, Zookeeper, Cassandra, Chukwa, and Apache Spark.
  • Capable of processing large sets of structured, semi - structured and unstructured data and supporting systems application architecture.
  • Experience working with Hadoop clusters using Cloudera, Horton Works, AWSdistributions.
  • Experience in managing and reviewing Hadoop Log files.
  • Experience with Oozie Workflow Engine in running workflow jobs with actions that run Hadoop Map/Reduce and Pig jobs.
  • Explored Spark, Kafka, Storm along with other open source projects to create a real time analytics framework
  • WroteHive UDFs and Pig UDFs based on requirements.
  • Developed automated scripts using Unix Shell for performing RUNSTATS, REORG, REBIND, COPY, LOAD, BACKUP, IMPORT, EXPORT and other related to database activities. Worked in Multiple Environment in installation and configuration.
  • Successful implementation of Name Node high availability (HA) for major production clusters and designed automatic failover control using Zookeeper and Quorum Journal Nodes.
  • Experience in performing major and minor upgrades of Hadoop clusters in Cloudera and Hortonworks distributions of Hadoop.
  • WroteJDBC, SQL and HQL statements for interacting with persistence layer to performCRUD operations.
  • Extensive knowledge in Agile, Waterfall and TDD Methodologies (SDLC).
  • Proficient in Object Oriented Analysis and Design (OOAD) based on Unified Modeling Language (UML).
  • Expertise in Front-End Development using MVC frameworks such as Spring MVC, Struts, JSF. Extensive experience in Java, J2EE technologies JDBC, JSP, Servlets, EJB, Hibernate.
  • Expertise in Client Side Designing and Validations using HTML, DHTML, XHTML, CSS, Java Script, JQuery, JSP, JSTL, Ajax, AngularJS.
  • Experienced in developing web services with XML based protocols such as SOAP and REST.
  • Experience of working with databases Oracle, SQL/MySQL and DB2.
  • Experience in using various Configuration Management tools like CVS, Clear Case and SVN.

TECHNICAL SKILLS:

Hadoop/Big Data Technologies:: HDFS, Map Reduce, YARN, Pig, Hive, HBase, and MongoDB. Cassandra, Sqoop, Flume, Oozie, Zookeeper.

Languages: J2SE, J2EE, SQL, PLSQL, UML, C/C++, Python.

Web Technologies: HTML, CSS, XHTML, XML, DHTML, JSP, AngularJS, JavaScript, JQuery, AJAX.

Frameworks: Struts, spring, Hibernate JSF, angular JS.

Database: MySQL, Oracle, SQL Server.

Application Servers: Apache Tomcat, Glassfish, Apache TomEE, Google Web Server.

Methodologies: SDLC, OOAD, Agile, Waterfall.

Operating systems: Windows, UNIX, Linux.

IDE Tools: Eclipse and Net Beans.

Version Control Tools: SVN, GIT, CSV.

PROFESSIONAL EXPERIENCE:

Confidential, Portland, Oregon

Hadoop Developer

Environment: s: Hive, HDFS, MapReduce, Sqoop.

Responsibilities:

  • Importing and exporting data into HDFS and Hive using Sqoop .
  • Created reports for the BI team using Sqoop export data into HDFS and Hive .
  • Defining job flows.
  • Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and processing.
  • Worked on developing, installing and configuring Hadoop ecosystem components that moved data from individual servers to HDFS.
  • Managing and reviewing Hadoop log files.
  • Installed and configured MapReduce, HIVE and the HDFS .Implemented CDH3 Hadoop cluster on CentOS. Assisted with performance tuning and monitoring.
  • Running Hadoop streaming jobs to process terabytes of xml format data
  • Loading and transforming large sets of structured, semi structured and unstructured data
  • Managing data coming from different sources.
  • Created HBase table to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Supported code/design analysis, strategy development and project planning.
  • Developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
  • Assisted with data capacity planning and node forecasting.
  • Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.
  • Administrator for Pig, Hive and Hbase installing updates, patches and upgrades.

Confidential, Bridgewater, NJ

Hadoop Developer

Environment: s: HDFS, MapReduce, Hbase, Sqoop, Zookeeper, Oozie, Pig.

Responsibilities:

  • Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
  • Defined Oozie Job flows.
  • Engaged in building skills and working with leading big data technologies like Cassandra, Hadoop, HDFS, MongoDB, and Zookeeper .
  • Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the HDFS and PIG to pre-process the data.
  • Became a major contributor and potential committer of an important open source Apache project.
  • Involved in start to end process of Hadoop cluster setup i.e., installation, configuration and monitoring the Hadoop Cluster.
  • Managed and reviewed Hadoop log files.
  • Responsible for Cluster maintenance, commissioning and decommissioning Data nodes, Cluster Monitoring,
  • Troubleshooting, Manage and review data backups, Manage & review Hadoop log files.
  • Monitoring systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup, and disaster recovery systems and procedures.
  • Configured various property files like core-site.xml, hdfs-site.xml, mapred-site.xml based upon the job requirement.
  • Importing/Exporting data into/from HDFS using Sqoop .
  • Loaded log data directly into HDFS using Flume .
  • Managed and reviewed Hadoop Log files as a part of administration for troubleshooting purposes.
  • Followed standard Back up policies to make sure the high availability of cluster.
  • Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
  • Installed and configured Flume, Hive, Pig, Sqoop and Oozie on the Hadoop cluster.

Confidential, Dallas, TX

Hadoop Developer

Environment: Flume, Hive, Pig, Sqoop and Oozie.

Responsibilities:

  • Used Flume to collect, aggregate, and store the web log data from different sources like web servers, mobile and network devices and pushed to HDFS .
  • Installed and configured Flume, Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
  • Analyzed the web log data using the HiveQL to extract number of unique visitors per day, page- views, visit duration, most purchased product on website.
  • Load log data into HDFS using Flume.
  • Designed and developed scalable custom Hadoop solutions as per dynamic data needs.
  • Involved in analyzing system failures, identifying root causes, and recommended course of actions.
  • Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
  • Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup, and disaster recovery systems and procedures.
  • Participated in development and execution of system and disaster recovery processes.
  • Monitored multiple Hadoop clusters environments using Ganglia and Nagios. Monitored workload, job performance and capacity planning using Cloudera Manager.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Supported technical team members in management and review of Hadoop log files and data backups.
  • Involved in installing and configuring Kerberos for the authentication of users and Hadoop daemons.
  • Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
  • Work with the Tera data analysis team using Big Data technologies to gather the business requirements.
  • Worked with Teradata Appliance team, Horton Works PM and Engineering Team, Aster PM and Engineering team.
  • Assisted in designing, development and architecture of Hadoop and HBase systems.
  • Coordinated with technical teams for installation of Hadoop and third related applications on systems.
  • Formulated procedures for planning and execution of system upgrades for all existing Hadoop clusters.
  • Supported technical team members for automation, installation and configuration tasks.
  • Suggested improvement processes for all process automation scripts and tasks.
  • Integrated Oozie with the rest of the Hadoop stack supporting several types of Hadoop jobs out of the box (like Java MapReduce, Pig, Hive, Sqoop ) as well as system specific jobs (such as Java programs and shell scripts).
  • Managed and reviewed Hadoop log files.

Confidential, East Peoria, IL

Java Developer

Responsibilities:

  • Designed Documents organization.
  • Used JDBC to connect to the information model tier.
  • Understanding client requirements and implementing code based on functionality.
  • Developed the UI panels using JSF, XHTML, CSS, DOJO and JQuery .
  • Used Eclipse IDE for designing, coding and developing applications.
  • Perform complex bug verification, release testing, and beta support for products.
  • At-a- glance dashboard showing voucher/ledger correctness and verification status.
  • Automated the manual process from the existing system.
  • Developed Servlets and JSPs based on MVC pattern using Struts framework and Spring Framework .

Confidential, Nashville, TN

Java Developer

Responsibilities:

  • Involved in Requirements gathering, Analysis, Design, and Development and Support phases of SDLC.
  • Involved in team analysis, design, development, deployment and managed to develop various modules.
  • Developed the UI panels using JSF, XHTML, CSS and JQuery
  • Involved in integration of web focus reporting application.
  • Used JAVA Mail API, and implemented an automated functionality to generate e-mails to the manager whenever a user registers for the application.
  • Responsible for Designing, Coding, Debugging.
  • Used JDBC to connect to the information model tier.
  • Designed and developed User interface with the use of JSF.
  • Developed few stored procedures, triggers and sequences.

Confidential

Data Analyst

Responsibilities:

  • Collecting, manipulating, cleansing & processing data using Excel and SQL .
  • Responsible for loading, extracting and validation of client data.
  • Liaising with end-users and 3rd party suppliers.
  • Analyzing raw data, drawing conclusions & developing recommendations.
  • Writing SQL scripts to manipulate data for data loads and extracts.
  • Developing data analytical databases from complex financial source data.
  • Performing daily system checks.
  • Data entry, data auditing, creating data reports & monitoring all data for accuracy.
  • Designing, developing and implementing new functionality.
  • Monitoring the automated loading processes.
  • Advising on the suitability of methodologies and suggesting improvements.
  • Carrying out specified data processing and statistical techniques.
  • Supplying qualitative and quantitative data to colleagues & clients.
  • Using Informatics & SAS to extract transform & load source data from transaction systems.

We'd love your feedback!