We provide IT Staff Augmentation Services!

Sr. Hadoop Developer/hive Resume

5.00/5 (Submit Your Rating)

Charlotte, NC

SUMMARY

  • Around 8 years of experience in Business Analytics, Data Modeling and Data warehousing.
  • Hands - on experience with Big Data architecture - Hadoop framework and its components.
  • 4+ years Hadoop development experience with Hive.
  • Extensive experience working with Hive, creating Hive tables, writing custom Hive UDF’s and Hive queries for analysis and reporting.
  • Good knowledge with core components such as MapReduce, YARN, HDFS.
  • In-depth knowledge of Hadoop EcoSystem - Pig, Hive, Spark, Scala, Avro, Parquet, Flume, Sqoop, Crunch, ZooKeeper, NoSQL databases like Cassandra, HBase.
  • Good knowledge of Hadoop Development and various components such as HDFS, Job Tracker, Task Tracker, Data Node, Name Node and Map-Reduce concepts.
  • Experienced working with Tableau software and other BI tools like Cognos.
  • Strong experience in SQL, Python and Shell scripting.
  • Working knowledge in Java development and its environment.
  • Knowledge of data flows, data architecture, ETL and processing of unstructured data.
  • Experience with Data Integration on traditional and Hadoop environments.
  • Experience designing and implementing reporting and visualization for unstructured and structured data sets.
  • Experience with Hadoop Cluster Administration and working with larger data sets.
  • Experienced with Git, GIT hub, Perforce and other source code management tools.
  • Experienced in all phases of Software Development Life Cycle (SDLC)
  • Extensive knowledge in various branching/merging methodologies & release procedures.
  • Excellent communication skills and experienced working with multiple teams gathering business requirements.

TECHNICAL SKILLS

Data Analytics: Hive, Pig, MapReduce, Hadoop 2.0, Apache Spark, Scala.

Databases: Cassandra, HBase,Oracle, MS SQL Server, DB2.

Reporting Technologies: Tableau v8.x, Cognos 10.x

ETL: Datastage 8.x

Languages: Python, C, C++, SQL, PL/SQL, Java, MATLAB.

Web Technologies: Java, Java Script, VBScript, HTML

Version Control SW: Subversion, Perforce, GIT, Jenkins, JIRA

IDE & Tools: Eclipse IDE, Android Studio, Maven

PROFESSIONAL EXPERIENCE

Confidential, Charlotte, NC

Sr. Hadoop Developer/Hive

Responsibilities:

  • Extensively involved in creating Hive tables, loading data and writing Hive queries.
  • Responsibilities included analyzing transformed data using HiveQL and Hive UDF's.
  • Responsible for HIVE performance tuning and optimizing Hadoop Mapreduce jobs.
  • Worked on HBase & Hive scripts to extract, transform & load data into HBase and Hive.
  • Written MapReduce code to process the data from various sources and storing parsed data into HBase and Hive using HBase-Hive Integration.
  • Worked in importing and exporting data into HDFS and Hive from RDBMS using Sqoop.
  • Worked with Oozie workflow engine to run multiple Hive Jobs.
  • Processed data into HDFS using MapReduce, Spark, Pig, Hive, Scala, Hive/Pig UDFs and Sqoop.
  • Involved in development of multiple MapReduce jobs for data cleaning & preprocessing
  • Managed and reviewed Hadoop log files to optimize MapReduce jobs performance.
  • Installed and configured Hive and also written Hive UDFs.
  • Involved in loading data from UNIX file system to HDFS.
  • Performed test run of batch jobs and generated reports.

Environment: Hadoop, Cloudera CDH5, Hive, Pig, HBase, Sqoop, Flume, Spark, Scala.

Confidential, San Diego, CA

Hadoop Developer/Hive Developer

Responsibilities:

  • Extensively worked with Hive creating tables, loading the data and writing hive queries.
  • Involved in developing Hive DDLs to create, alter and drop Hive TABLES.
  • Implemented various Hive UDF's as per business requirements
  • Worked on Hive performance tuning and Hadoop Mapreduce operation optimization.
  • Created partitioned tables in Hive.
  • Performed analysis using Hive on the partitioned and bucketed data to compute various metrics for reporting using Tableau.
  • Importing and exporting of data into HDFS and Hive using Sqoop.
  • Worked with Oozie workflow engine to run multiple Hive Jobs.
  • Developed shell scripts for processing of hive scripts and according to business need.
  • Developed ETL code using different transformations to extract, transform the data from legacy sources and load data into target system.
  • Worked on processing ingested raw data using MapReduce, Apache Pig, Scala and Hive.
  • Worked extensively on Hadoop platform like installing cluster, commissioning & decommissioning of Datanode, capacity planning and slot configuration.
  • Worked with NoSQL databases like Cassandra and HBASE.
  • Gained experience working with Kafka-Storm on MAPR platform for real time analysis.
  • Co-ordinated with Offshore team members to meet tight deadlines.
  • Responsible for implementation and ongoing administration of Cloudera platform.
  • Responsible for configuring and monitoring of a cluster of multiple nodes.
  • Responsible for Installation and configuraton of Hadoop Map reduce, HDFS.

Environment: Hadoop, HDFS, Hive, Pig, Java, Python, Oracle, MS SQL Server.

Confidential, San Diego, CA

Sr. System Administrator

Responsibilities:

  • Performing security management by maintaining roles, privileges, and user profiles. Add/Remove users from many different Security Groups in our domain (Auto groups)
  • Run, Schedule, Enable SQL Jobs. Troubleshoot if the jobs fail, provide log files of jobs to dev teamsLook into the health of the Production Databases.
  • Monitoring Mirroring, Log shipping & Replication. Regular monitoring of SQL Server logs for errors.
  • Schedule backup jobs and object level recovery using maxima tool.
  • Applying DML statements on different production databases as per the customer request. Monitoring of Online and scheduled jobs.

Confidential, Dulles, VA

System Administrator

Responsibilities:

  • Configuring and maintenance various MySQL databases for applications like Drupal.
  • Setup and Configured Oracle Golden Gate software on the client side
  • Have installed Golden Gate for replication between RAC clustered Databases.
  • Experience with Upgrading Oracle Databases Versions from 10g to 11gr2.
  • Have worked with Oracle Data Guard for Disaster Recovery using Standby Databas
  • Experience with installing Drupal Multi Site and Migration of data

We'd love your feedback!