We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

San Diego, CA

PROFESSIONAL SUMMARY:

  • 7+ years of overall IT experience in a variety of industries, which includes hands on experience in Big Data technologies
  • 2+ years of comprehensive experience as a Hadoop Developer
  • Passionate towards working in Big Data and Analytics environment
  • Expertise in writing Hadoop Jobs for analyzing data using Hive and Pig.
  • Experience in writing MapReduce programs using Java.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa
  • In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts
  • Extending Hive and Pig core functionality by writing custom UDFs
  • Experience in analyzing data using Hive QL, Pig Latin, and custom MapReduce programs in Java.
  • Extensive experience with SQL, PL/SQL and database concepts
  • Experience writing job workflow scheduling and coordinator jobs using Oozie.
  • Experience in developing solutions to analyze large data sets efficiently
  • Knowledge of administrative tasks such as installing Hadoop and its ecosystem components such as Hive and Pig.
  • Good knowledge on No SQL database HBASE.
  • Working knowledge on all phases of Software Development Life Cycle (SDLC). Ability to track projects from inception to deployment.
  • Good understanding of XML methodologies (XML,XSL,XSD) including Web Services and SOAP
  • Handled several techno-functional responsibilities including estimates, identifying functional and technical gaps, requirements gathering, designing solutions, development, developing documentation, and production support
  • Development experience using Visual Studio, Visual Studio.Net, Visual Source Safe.
  • An individual with excellent interpersonal and communication skills, strong business acumen, creative problem solving skills, technical competency, team-player spirit, and leadership skills

TECHNICAL SKILLS:

Hadoop/Big Data: HDFS, MapReduce, HBase, Pig, Hive, Sqoop, Oozie

Java & J2EE Technologies: Core Java, Servlets, JSP, JDBC, JNDI, Java Beans

IDE Tools: Eclipse, NetBeans

Programming languages: C, C++, Java, Python, Linux shell scripts, VB.NET

Databases: Oracle 11g/10g/9i, MySQL, DB2, MS-SQL Server

Operating Systems: Windows 95/98/2000/XP/Vista/7, Macintosh, Unix

Web Technologies: HTML, XML, JavaScriptETL Tools: Informatica, Pentaho

Testing: Win Runner, Load Runner, QTP

PROFESSIONAL EXPERIENCE:

Confidential, San Diego, CA

Hadoop Developer

Responsibilities:

  • Mainly involved in creating and running Hadoop jobs to process raw binary data produced by vehicle sensors.
  • Massaging and parsing of the obtained raw binary data.
  • Gathering requirements from the product owner and the data science team.
  • Developed many MapReduce jobs in native Java for pre-processing of the data.
  • Developed Hive scripts to create both Internal and External Hive tables to store the transformed data..
  • Involved in creating and scheduling Oozie workflow scripts to run series of Sqoop imports, Mapreduce Transformation jobs, Hive scripts.
  • Created and maintained technical documentation for all the workflows.
  • Developed DFD (Data Flow Diagrams) on the company’s own Wiki for documentation and knowledge to understand the entire workflow of the project.
  • Worked with business analysts, Data science team and product owners to identify the tasks and obtain new requirements as part of agile scrum methodology.

Environment: MapReduce, Hive, Sqoop, Oozie, Hortonworks, HUE, Ambari, AVRO, Java, Hadoop, HDFS, Pig, and Big Data

Confidential, Irving, TX

Hadoop Developer

Responsibilities:

  • Analyzed large data sets by running Hive queries and Pig scripts
  • Worked with the Data Science team to gather requirements for various data mining projects
  • Involved in creating Hive tables, and loading and analyzing data using hive queries
  • Developed Simple to complex MapReduce Jobs using Hive and Pig
  • Involved in running Hadoop jobs for processing millions of records of text data
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
  • Developed multiple MapReduce jobs in java for data cleaning and preprocessing
  • Involved in loading data from LINUX file system to HDFS
  • Responsible for managing data from multiple sources
  • Extracted files from Couch DB through Sqoop and placed in HDFS and processed.
  • Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Responsible to manage data coming from different sources.
  • Assisted in exporting analyzed data to relational databases using Sqoop
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts

Environment: Hadoop, HDFS, Pig, Hive, MapReduce, Sqoop, LINUX, and Big Data

Confidential, Camarillo, CA

Hadoop Developer

Responsibilities:

  • Worked on analyzing Hadoop cluster using different big data analytic tools including Pig, Hive, and MapReduce
  • Collecting and aggregating large amounts of log data using Apache Flume and staging data in HDFS for further analysis
  • Worked on debugging, performance tuning of Hive & Pig Jobs
  • Created Hbase tables to store various data formats of PII data coming from different portfolios
  • Implemented test scripts to support test driven development and continuous integration
  • Worked on tuning the performance Pig queries
  • Involved in loading data from LINUX file system to HDFS
  • Importing and exporting data into HDFS and Hive using Sqoop
  • Experience working on processing unstructured data using Pig and Hive
  • Supported MapReduce Programs those are running on the cluster
  • Gained experience in managing and reviewing Hadoop log files
  • Involved in scheduling Oozie workflow engine to run multiple Hive and pig jobs
  • Assisted in monitoring Hadoop cluster using tools like Nagios, and Ganglia
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts

Environment: Hadoop, HDFS, Pig, Hive, MapReduce, Sqoop, Oozie, Nagios, Ganglia, LINUX, and Big Data

Confidential, Dallas, TX

Java/J2EE/ Hadoop Developer

Responsibilities:

  • Developed the application using Struts Framework that leverages classical Model View Layer (MVC) architecture UML diagrams like use cases, class diagrams, interaction diagrams, and activity diagrams were used
  • Participated in requirement gathering and converting the requirements into technical specifications
  • Extensively worked on User Interface for few modules using JSPs, JavaScript and Ajax
  • Created Business Logic using Servlets, Session beans and deployed them on Web logic server
  • Wrote complex SQL queries and stored procedures
  • Developed the XML Schema and Web services for the data maintenance and structures
  • Implemented the Web Service client for the login authentication, credit reports and applicant information using Apache Axis 2 Web Service
  • Responsible to manage data coming from different sources.
  • Got good experience with NOSQL database.
  • Involved in loading data from UNIX file system to HDFS.
  • Installed and configured Hive and also written Hive UDFs.
  • Designed the logical and physical data model, generated DDL scripts, and wrote DML scripts for Oracle 9i database
  • UsedHibernateORM framework withSpringframework for data persistence and transaction management
  • Used struts validation framework for form level validation
  • Wrote test cases in JUnit for unit testing of classes
  • Involved in templates and screens in HTML and JavaScript
  • Involved in integrating Web Services using WSDL and UDDI
  • Provided Technical support for production environments resolving the issues, analyzing the defects, providing and implementing the solution defects
  • Built and deployed Java applications into multiple Unix based environments and produced both unit and functional test results along with release notes

Environment: JDK 1.5, J2EE 1.4, Struts 1.3, JSP, Servlets 2.5, WebSphere 6.1, HTML, XML, ANT 1.6, JavaScript, JUnit 3.8

Confidential, Alpharetta GA

Java/J2EE developer

Responsibilities:

  • EJB (Session Beans and Entity Beans) on Web sphere Studio Application Developer.
  • Used different Design patterns, like MVC, EJBs Session facade, Controller Servlets, while implementing the Framework.
  • Front End was built using JSPs, Servlets, JavaScript and HTML.
  • Built Custom Tags for JSPs.
  • Built the report module on reports based from Crystal reports.
  • Integrating data from multiple data sources.
  • Generating schema difference reports for database using toad.
  • Built Prototypes for internationalization.
  • Wrote Stored Procedures in DB2.

Environment: J2EE (JSP’s, Servlets, EJB), HTML, Struts, DB2.

Confidential

Junior .NET Developer

Responsibilities:

  • Preparation of architectural designs for each of the components that were assigned for development.
  • Responsible for designing, developing and testing the Due Date Module using ASP.NET, HTML, CSS, JavaScript and SQL Server 2005 as database
  • Vacation Services Module was designed developed and tested using C#.NET, HTML, CSS, VBScript, JavaScript and SQL Server 2005 as database.
  • Product Report tool was designed and developed using ASP.NET, XHTML, HTML, JavaScript and SQL Server 2005 as database.
  • Development of prototypes using various tools.
  • Most of the internal structure uses XML for generating input/output.
  • Created type libraries from the existing C# dlls and these dlls are consumed in the Middle Tier Architecture.
  • Communicated with so many groups in order to integrate the new enhancement into the regular flow.
  • Gave extensive production support.

Environment: C#.Net, ASP.NET, Microsoft Visual Studio, SQL Server 2005, HTML, CSS, JavaScript, VBScript, XML, Microsoft Visual Source Safe and Windows XP.

We'd love your feedback!