Hadoop/big Data Developer Resume Profile
4.50/5 (Submit Your Rating)
Atlanta, GA
Professional Summary:
- Over 8 years of IT experience, with around 3 years of experience in Hadoop and Hadoop Ecosystem.
- Expertise in concepts of end-to-end project planning and implementation from scope management in various environments viz. release based maintenance, custom application development, enterprise wide application deployment, testing support and quality management in adherence to international guidelines and norms
- Cloudera certified Hadoop Developer with hands on experience on major components in Hadoop Ecosystem like Hadoop Map Reduce, HDFS, HIVE, PIG, Hbase, Zookeeper,Oozie and Flume.
- Experience in using Kettleby Pentaho.
- Expertise in setting up processes for Hadoop based application design and implementation.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Experience in managing and reviewing Hadoop log files.
- Experienced in processing Big data on the Apache Hadoop framework using MapReduce programs.
- Excellent understanding and knowledge of NOSQL databases like HBase and Mongo DB.
- Experience in working with Windows, UNIX/LINUX platform with different technologies such as Big Data, SQL, XML, HTML, Core Java, Shell Scripting etc.
- Experience in giving training and guiding new team members in the Project.
- Experience in detailed system design using use case analysis, functional analysis, modelling program with class sequence, activity and state diagrams using UML and rational rose.
- Proficient in Retail, Telecom and Banking Domains.
- Very good experience in customer specification study, requirements gathering, system architectural design and turning the requirements into final product.
- Experience in interacting with customers and working at client locations for real time field testing of products and services.
- Ability to work effectively with associates at all levels within the organization.
- Strong background in mathematics and have very good analytical and problem solving skills.
TechnicalSkills:
- Hadoop/Bigdata :HDFS,MapReduce,Sqoop,Hive,PIG,HBASE,Zookeeper,
- Clusterconfiguration,FLUME,AWS
- Distributions : Cloudera
- JavaTechnologies :CoreJava,JDBC,HTML,JSP,Servlets,Tomcat,JavaScript
- Databases : SQL, NOSQL HBase ,MYSQL,Oracle,PL/SQL.
- ProgrammingLanguages :C, C , C , Java, SQL, Shell, Python
- IDE's Utilities :Eclipse
- WebTechnologies :J2EE, JMS, Web Service
- Protocols :TCP/IP,SSH,HTTPandHTTPS
- Scripting : HTML, JavaScript, CSS, XML and Ajax
- Operating System : Windows, Mac, Linux and Unix
- IDE : Eclipse, Microsoft Visual Studio 2008,2012 , Flex Builder
- Version control : Git, SVN, CVS
- Tools : FileZilla, Putty, PL/SQL Developer, JUnit,
ProfessionalExperience:
Confidential
Role: Hadoop/Big Data Developer
Responsibilities:
- Processed data into HDFS by developing solutions, analyzed the data using MapReduce, Pig, Hive and produce summary results from Hadoop to downstream systems.
- Used Kettle widely in order to import data from various systems/sources like MySQL into HDFS.
- Did various performance optimizations like using distributed cache for small datasets, Partition, Bucketing in hive and Map Side joins.
- Involved in creating Hive tables, and then applied HiveQL on those tables for data validation.
- Moved the data from Hive tables into Mongo collections.
- Used Zookeeper for various types of centralized configurations.
- Involved in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries and Pig scripts
- Managed and reviewed Hadoop log files.
- Tested raw data and executed performance scripts.
- Shared responsibility for administration of Hadoop, Hive and Pig.
Environment:
Hadoop, Pig, Hive, Sqoop, Flume, MapReduce, HDFS, LINUX, Oozie, MongoDB
Confidential
Role: Hadoop/Big Data Developer
Responsibilities:
- Developed multiple MapReduce jobs in java for data cleaning and preprocessing Involved in Importing and exporting data into HDFS and Hive using Kettle.
- Involved in defining job flows.
- Involved in managing and reviewing Hadoop log files.
- Involved in Loading and transforming large sets of structured, semi structured and unstructured data.
- Involved in loading data from UNIX file system to HDFS.
- Involved in creating Hive tables, loading with data and writing hive queries which will run internally in Map reduce way.
Environment:
Java 6 JDK 1.6 , Eclipse, Oracle 10g, Sub Version, Hadoop, Hive, HBase, Linux, HDFS, Hive, Oracle 11g / 10g, PL/SQL, SQL PLUS, Toad 9.6, Windows NT, UNIX Shell Scripting.
Confidential
Role: Hadoop/Big Data Developer
Responsibilities:
- Worked on 3TB of data involving 10Nodes.
- Good understanding and related experience with Hadoop stack-internals, Hive, Pig and Map/Reduce.
- Deep understanding of schedulers, workload management, availability, scalability and distributed data platforms.
- Installed and configured Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and pre-processing.
- Involved in loading data from UNIX file system to HDFS.
- Wrote MapReduce jobs to discover trends in data usage by users.
- Involved in managing and reviewing Hadoop log files.
- Involved in running Hadoop streaming jobs to process terabytes of text data.
- Developed HIVE queries for the analysts.
- Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
- Exported the result set from HIVE to MySQL using Shell scripts.
- Used Git for version control.
- Maintain System integrity of all sub-components primarily HDFS, MR, HBase, and Flume .
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
Environment:
Hadoop, HDFS, Map Reduce, Hive, Pig, Sqoop, Java 1.6, UNIX Shell Scripting.