Hadoop Developer Resume
San Diego, CA
PROFESSIONAL SUMMARY:
- 7+ years of overall IT experience in a variety of industries, which includes hands on experience in Big Data technologies
- 2+ years of comprehensive experience as a Hadoop Developer
- Passionate towards working in Big Data and Analytics environment
- Expertise in writing Hadoop Jobs for analyzing data using Hive and Pig.
- Experience in writing MapReduce programs using Java.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa
- In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts
- Extending Hive and Pig core functionality by writing custom UDFs
- Experience in analyzing data using Hive QL, Pig Latin, and custom MapReduce programs in Java.
- Extensive experience with SQL, PL/SQL and database concepts
- Experience writing job workflow scheduling and coordinator jobs using Oozie.
- Experience in developing solutions to analyze large data sets efficiently
- Knowledge of administrative tasks such as installing Hadoop and its ecosystem components such as Hive and Pig.
- Good knowledge on No SQL database HBASE.
- Working knowledge on all phases of Software Development Life Cycle (SDLC). Ability to track projects from inception to deployment.
- Good understanding of XML methodologies (XML,XSL,XSD) including Web Services and SOAP
- Handled several techno-functional responsibilities including estimates, identifying functional and technical gaps, requirements gathering, designing solutions, development, developing documentation, and production support
- Development experience using Visual Studio, Visual Studio.Net, Visual Source Safe.
- An individual with excellent interpersonal and communication skills, strong business acumen, creative problem solving skills, technical competency, team-player spirit, and leadership skills
TECHNICAL SKILLS:
Hadoop/Big Data: HDFS, MapReduce, HBase, Pig, Hive, Sqoop, Oozie
Java & J2EE Technologies: Core Java, Servlets, JSP, JDBC, JNDI, Java Beans
IDE Tools: Eclipse, NetBeans
Programming languages: C, C++, Java, Python, Linux shell scripts, VB.NET
Databases: Oracle 11g/10g/9i, MySQL, DB2, MS-SQL Server
Operating Systems: Windows 95/98/2000/XP/Vista/7, Macintosh, Unix
Web Technologies: HTML, XML, JavaScriptETL Tools: Informatica, Pentaho
Testing: Win Runner, Load Runner, QTP
PROFESSIONAL EXPERIENCE:
Confidential, San Diego, CA
Hadoop Developer
Responsibilities:
- Mainly involved in creating and running Hadoop jobs to process raw binary data produced by vehicle sensors.
- Massaging and parsing of the obtained raw binary data.
- Gathering requirements from the product owner and the data science team.
- Developed many MapReduce jobs in native Java for pre-processing of the data.
- Developed Hive scripts to create both Internal and External Hive tables to store the transformed data..
- Involved in creating and scheduling Oozie workflow scripts to run series of Sqoop imports, Mapreduce Transformation jobs, Hive scripts.
- Created and maintained technical documentation for all the workflows.
- Developed DFD (Data Flow Diagrams) on the company’s own Wiki for documentation and knowledge to understand the entire workflow of the project.
- Worked with business analysts, Data science team and product owners to identify the tasks and obtain new requirements as part of agile scrum methodology.
Environment: MapReduce, Hive, Sqoop, Oozie, Hortonworks, HUE, Ambari, AVRO, Java, Hadoop, HDFS, Pig, and Big Data
Confidential, Irving, TX
Hadoop Developer
Responsibilities:
- Analyzed large data sets by running Hive queries and Pig scripts
- Worked with the Data Science team to gather requirements for various data mining projects
- Involved in creating Hive tables, and loading and analyzing data using hive queries
- Developed Simple to complex MapReduce Jobs using Hive and Pig
- Involved in running Hadoop jobs for processing millions of records of text data
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
- Developed multiple MapReduce jobs in java for data cleaning and preprocessing
- Involved in loading data from LINUX file system to HDFS
- Responsible for managing data from multiple sources
- Extracted files from Couch DB through Sqoop and placed in HDFS and processed.
- Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Responsible to manage data coming from different sources.
- Assisted in exporting analyzed data to relational databases using Sqoop
- Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts
Environment: Hadoop, HDFS, Pig, Hive, MapReduce, Sqoop, LINUX, and Big Data
Confidential, Camarillo, CA
Hadoop Developer
Responsibilities:
- Worked on analyzing Hadoop cluster using different big data analytic tools including Pig, Hive, and MapReduce
- Collecting and aggregating large amounts of log data using Apache Flume and staging data in HDFS for further analysis
- Worked on debugging, performance tuning of Hive & Pig Jobs
- Created Hbase tables to store various data formats of PII data coming from different portfolios
- Implemented test scripts to support test driven development and continuous integration
- Worked on tuning the performance Pig queries
- Involved in loading data from LINUX file system to HDFS
- Importing and exporting data into HDFS and Hive using Sqoop
- Experience working on processing unstructured data using Pig and Hive
- Supported MapReduce Programs those are running on the cluster
- Gained experience in managing and reviewing Hadoop log files
- Involved in scheduling Oozie workflow engine to run multiple Hive and pig jobs
- Assisted in monitoring Hadoop cluster using tools like Nagios, and Ganglia
- Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts
Environment: Hadoop, HDFS, Pig, Hive, MapReduce, Sqoop, Oozie, Nagios, Ganglia, LINUX, and Big Data
Confidential, Dallas, TX
Java/J2EE/ Hadoop Developer
Responsibilities:
- Developed the application using Struts Framework that leverages classical Model View Layer (MVC) architecture UML diagrams like use cases, class diagrams, interaction diagrams, and activity diagrams were used
- Participated in requirement gathering and converting the requirements into technical specifications
- Extensively worked on User Interface for few modules using JSPs, JavaScript and Ajax
- Created Business Logic using Servlets, Session beans and deployed them on Web logic server
- Wrote complex SQL queries and stored procedures
- Developed the XML Schema and Web services for the data maintenance and structures
- Implemented the Web Service client for the login authentication, credit reports and applicant information using Apache Axis 2 Web Service
- Responsible to manage data coming from different sources.
- Got good experience with NOSQL database.
- Involved in loading data from UNIX file system to HDFS.
- Installed and configured Hive and also written Hive UDFs.
- Designed the logical and physical data model, generated DDL scripts, and wrote DML scripts for Oracle 9i database
- UsedHibernateORM framework withSpringframework for data persistence and transaction management
- Used struts validation framework for form level validation
- Wrote test cases in JUnit for unit testing of classes
- Involved in templates and screens in HTML and JavaScript
- Involved in integrating Web Services using WSDL and UDDI
- Provided Technical support for production environments resolving the issues, analyzing the defects, providing and implementing the solution defects
- Built and deployed Java applications into multiple Unix based environments and produced both unit and functional test results along with release notes
Environment: JDK 1.5, J2EE 1.4, Struts 1.3, JSP, Servlets 2.5, WebSphere 6.1, HTML, XML, ANT 1.6, JavaScript, JUnit 3.8
Confidential, Alpharetta GA
Java/J2EE developer
Responsibilities:
- EJB (Session Beans and Entity Beans) on Web sphere Studio Application Developer.
- Used different Design patterns, like MVC, EJBs Session facade, Controller Servlets, while implementing the Framework.
- Front End was built using JSPs, Servlets, JavaScript and HTML.
- Built Custom Tags for JSPs.
- Built the report module on reports based from Crystal reports.
- Integrating data from multiple data sources.
- Generating schema difference reports for database using toad.
- Built Prototypes for internationalization.
- Wrote Stored Procedures in DB2.
Environment: J2EE (JSP’s, Servlets, EJB), HTML, Struts, DB2.
Confidential
Junior .NET Developer
Responsibilities:
- Preparation of architectural designs for each of the components that were assigned for development.
- Responsible for designing, developing and testing the Due Date Module using ASP.NET, HTML, CSS, JavaScript and SQL Server 2005 as database
- Vacation Services Module was designed developed and tested using C#.NET, HTML, CSS, VBScript, JavaScript and SQL Server 2005 as database.
- Product Report tool was designed and developed using ASP.NET, XHTML, HTML, JavaScript and SQL Server 2005 as database.
- Development of prototypes using various tools.
- Most of the internal structure uses XML for generating input/output.
- Created type libraries from the existing C# dlls and these dlls are consumed in the Middle Tier Architecture.
- Communicated with so many groups in order to integrate the new enhancement into the regular flow.
- Gave extensive production support.
Environment: C#.Net, ASP.NET, Microsoft Visual Studio, SQL Server 2005, HTML, CSS, JavaScript, VBScript, XML, Microsoft Visual Source Safe and Windows XP.