Hadoop Developer Resume
Portland, OregoN
PROFESSIONAL SUMMARY:
- 7+ years of overall IT experience, with specialization in Hadoop andJava development.
- Experience in installing, configuring, designing, implementing, testing Big data projects using Hadoop Ecosystem like HDFS, Map Reduce, Hive, Pig, Sqoop,Oozie, Flume, HBase, Zookeeper, Cassandra, Chukwa, and Apache Spark.
- Capable of processing large sets of structured, semi - structured and unstructured data and supporting systems application architecture.
- Experience working with Hadoop clusters using Cloudera, Horton Works, AWSdistributions.
- Experience in managing and reviewing Hadoop Log files.
- Experience with Oozie Workflow Engine in running workflow jobs with actions that run Hadoop Map/Reduce and Pig jobs.
- Explored Spark, Kafka, Storm along with other open source projects to create a real time analytics framework
- WroteHive UDFs and Pig UDFs based on requirements.
- Developed automated scripts using Unix Shell for performing RUNSTATS, REORG, REBIND, COPY, LOAD, BACKUP, IMPORT, EXPORT and other related to database activities. Worked in Multiple Environment in installation and configuration.
- Successful implementation of Name Node high availability (HA) for major production clusters and designed automatic failover control using Zookeeper and Quorum Journal Nodes.
- Experience in performing major and minor upgrades of Hadoop clusters in Cloudera and Hortonworks distributions of Hadoop.
- WroteJDBC, SQL and HQL statements for interacting with persistence layer to performCRUD operations.
- Extensive knowledge in Agile, Waterfall and TDD Methodologies (SDLC).
- Proficient in Object Oriented Analysis and Design (OOAD) based on Unified Modeling Language (UML).
- Expertise in Front-End Development using MVC frameworks such as Spring MVC, Struts, JSF. Extensive experience in Java, J2EE technologies JDBC, JSP, Servlets, EJB, Hibernate.
- Expertise in Client Side Designing and Validations using HTML, DHTML, XHTML, CSS, Java Script, JQuery, JSP, JSTL, Ajax, AngularJS.
- Experienced in developing web services with XML based protocols such as SOAP and REST.
- Experience of working with databases Oracle, SQL/MySQL and DB2.
- Experience in using various Configuration Management tools like CVS, Clear Case and SVN.
TECHNICAL SKILLS:
Hadoop/Big Data Technologies:: HDFS, Map Reduce, YARN, Pig, Hive, HBase, and MongoDB. Cassandra, Sqoop, Flume, Oozie, Zookeeper.
Languages: J2SE, J2EE, SQL, PLSQL, UML, C/C++, Python.
Web Technologies: HTML, CSS, XHTML, XML, DHTML, JSP, AngularJS, JavaScript, JQuery, AJAX.
Frameworks: Struts, spring, Hibernate JSF, angular JS.
Database: MySQL, Oracle, SQL Server.
Application Servers: Apache Tomcat, Glassfish, Apache TomEE, Google Web Server.
Methodologies: SDLC, OOAD, Agile, Waterfall.
Operating systems: Windows, UNIX, Linux.
IDE Tools: Eclipse and Net Beans.
Version Control Tools: SVN, GIT, CSV.
PROFESSIONAL EXPERIENCE:
Confidential, Portland, Oregon
Hadoop Developer
Environment: s: Hive, HDFS, MapReduce, Sqoop.
Responsibilities:
- Importing and exporting data into HDFS and Hive using Sqoop .
- Created reports for the BI team using Sqoop export data into HDFS and Hive .
- Defining job flows.
- Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and processing.
- Worked on developing, installing and configuring Hadoop ecosystem components that moved data from individual servers to HDFS.
- Managing and reviewing Hadoop log files.
- Installed and configured MapReduce, HIVE and the HDFS .Implemented CDH3 Hadoop cluster on CentOS. Assisted with performance tuning and monitoring.
- Running Hadoop streaming jobs to process terabytes of xml format data
- Loading and transforming large sets of structured, semi structured and unstructured data
- Managing data coming from different sources.
- Created HBase table to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
- Supported code/design analysis, strategy development and project planning.
- Developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
- Assisted with data capacity planning and node forecasting.
- Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.
- Administrator for Pig, Hive and Hbase installing updates, patches and upgrades.
Confidential, Bridgewater, NJ
Hadoop Developer
Environment: s: HDFS, MapReduce, Hbase, Sqoop, Zookeeper, Oozie, Pig.
Responsibilities:
- Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
- Defined Oozie Job flows.
- Engaged in building skills and working with leading big data technologies like Cassandra, Hadoop, HDFS, MongoDB, and Zookeeper .
- Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the HDFS and PIG to pre-process the data.
- Became a major contributor and potential committer of an important open source Apache project.
- Involved in start to end process of Hadoop cluster setup i.e., installation, configuration and monitoring the Hadoop Cluster.
- Managed and reviewed Hadoop log files.
- Responsible for Cluster maintenance, commissioning and decommissioning Data nodes, Cluster Monitoring,
- Troubleshooting, Manage and review data backups, Manage & review Hadoop log files.
- Monitoring systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup, and disaster recovery systems and procedures.
- Configured various property files like core-site.xml, hdfs-site.xml, mapred-site.xml based upon the job requirement.
- Importing/Exporting data into/from HDFS using Sqoop .
- Loaded log data directly into HDFS using Flume .
- Managed and reviewed Hadoop Log files as a part of administration for troubleshooting purposes.
- Followed standard Back up policies to make sure the high availability of cluster.
- Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
- Installed and configured Flume, Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
Confidential, Dallas, TX
Hadoop Developer
Environment: Flume, Hive, Pig, Sqoop and Oozie.
Responsibilities:
- Used Flume to collect, aggregate, and store the web log data from different sources like web servers, mobile and network devices and pushed to HDFS .
- Installed and configured Flume, Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
- Analyzed the web log data using the HiveQL to extract number of unique visitors per day, page- views, visit duration, most purchased product on website.
- Load log data into HDFS using Flume.
- Designed and developed scalable custom Hadoop solutions as per dynamic data needs.
- Involved in analyzing system failures, identifying root causes, and recommended course of actions.
- Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
- Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup, and disaster recovery systems and procedures.
- Participated in development and execution of system and disaster recovery processes.
- Monitored multiple Hadoop clusters environments using Ganglia and Nagios. Monitored workload, job performance and capacity planning using Cloudera Manager.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Supported technical team members in management and review of Hadoop log files and data backups.
- Involved in installing and configuring Kerberos for the authentication of users and Hadoop daemons.
- Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
- Work with the Tera data analysis team using Big Data technologies to gather the business requirements.
- Worked with Teradata Appliance team, Horton Works PM and Engineering Team, Aster PM and Engineering team.
- Assisted in designing, development and architecture of Hadoop and HBase systems.
- Coordinated with technical teams for installation of Hadoop and third related applications on systems.
- Formulated procedures for planning and execution of system upgrades for all existing Hadoop clusters.
- Supported technical team members for automation, installation and configuration tasks.
- Suggested improvement processes for all process automation scripts and tasks.
- Integrated Oozie with the rest of the Hadoop stack supporting several types of Hadoop jobs out of the box (like Java MapReduce, Pig, Hive, Sqoop ) as well as system specific jobs (such as Java programs and shell scripts).
- Managed and reviewed Hadoop log files.
Confidential, East Peoria, IL
Java Developer
Responsibilities:
- Designed Documents organization.
- Used JDBC to connect to the information model tier.
- Understanding client requirements and implementing code based on functionality.
- Developed the UI panels using JSF, XHTML, CSS, DOJO and JQuery .
- Used Eclipse IDE for designing, coding and developing applications.
- Perform complex bug verification, release testing, and beta support for products.
- At-a- glance dashboard showing voucher/ledger correctness and verification status.
- Automated the manual process from the existing system.
- Developed Servlets and JSPs based on MVC pattern using Struts framework and Spring Framework .
Confidential, Nashville, TN
Java Developer
Responsibilities:
- Involved in Requirements gathering, Analysis, Design, and Development and Support phases of SDLC.
- Involved in team analysis, design, development, deployment and managed to develop various modules.
- Developed the UI panels using JSF, XHTML, CSS and JQuery
- Involved in integration of web focus reporting application.
- Used JAVA Mail API, and implemented an automated functionality to generate e-mails to the manager whenever a user registers for the application.
- Responsible for Designing, Coding, Debugging.
- Used JDBC to connect to the information model tier.
- Designed and developed User interface with the use of JSF.
- Developed few stored procedures, triggers and sequences.
Confidential
Data Analyst
Responsibilities:
- Collecting, manipulating, cleansing & processing data using Excel and SQL .
- Responsible for loading, extracting and validation of client data.
- Liaising with end-users and 3rd party suppliers.
- Analyzing raw data, drawing conclusions & developing recommendations.
- Writing SQL scripts to manipulate data for data loads and extracts.
- Developing data analytical databases from complex financial source data.
- Performing daily system checks.
- Data entry, data auditing, creating data reports & monitoring all data for accuracy.
- Designing, developing and implementing new functionality.
- Monitoring the automated loading processes.
- Advising on the suitability of methodologies and suggesting improvements.
- Carrying out specified data processing and statistical techniques.
- Supplying qualitative and quantitative data to colleagues & clients.
- Using Informatics & SAS to extract transform & load source data from transaction systems.