We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

0/5 (Submit Your Rating)

Dallas, TX

SUMMARY:

  • Over 8 years of total IT experience. 5 years in the fields of Big Data, Data Analysis,Business Intelligence in the role of Hadoop Administrator and 3 years hardware and software maintenance as help desk support.
  • Configuration, Administration, Monitoring, Security Management, Performance Tuning, Replication, Backup/Restore and troubleshooting of Hadoop infrastructure (Datastax / Apache) and other Open source / Big Data databases in a highly dynamic environment.
  • Expertise in Hadoop Ecosystem, HDFS Architecture and Cluster concepts such as YARN Management, HDFS,NoSQLtechnology HBASE, MapReduce, HIVE, PIG, FLUME, OOZIE, SQOOP,HBase and ZOOKEEPER.
  • Building Hadoop architecture with Cloudera distribution.
  • Experience in developing software solutions to build out capabilities on a Big Data Platform.
  • Experience in configuring cluster and installing the services, monitoring the cluster by eliminating the compatibility errors.
  • Experience with Virtualization tools like VMware, Puppet and KVM.
  • Experience in writing MapReduce programs with java for Hadoop tasks.
  • Experience with Java programming for MapReduce, Spark and Scalatechnologies.
  • Experience with Cassandra and/or big data databases in a clustered environment.
  • Familiar with Linux operating system and Cassandra architecture
  • Ability to create tools from scratch needed for database administration
  • Familiar with Hortonworks and Cloudera distributions of Hadoop .
  • Installation and configuration procedures, add/bootstrap nodes to clusters, remove/replace nodes in clusters, replication between data centers.
  • Experience with UNIX/Linux including basic commands, shell scripting and system administration/configuration.
  • Experience with Linux OS administration in Ubuntu 14.04 and RedHat 6.5
  • Experience in analyzing, requirements gathering, and writing system functional specifications including use cases for record keeping, under Agile and Scrum SDLC life cycles.
  • Experience in Presenting Project/Product demo for clients/end users with creative presentation and documentation.

TECHNICAL SKILLS:

Skills: Microsoft Office, Microsoft Project, MicrosoftExcel, Java Programing, VMware, KVM, Puppet or OpenStack, Centrify, LDAP and Kerberos

Big Data Ecosystem: SQL, Cloudera manager,Platfora, Apache Drill, HDFS, HBase, Hadoop MapReduce, Zookeeper, Hive, Pig, Sqoop, Flume, Oozie, MySQL,Spark, ScalaSAS, R, Stata and Matlab

Operating System: Windows, UNIX, Linux Ubuntu, Red Hat Enterprise Linux Version 7+

IDE / Testing Tools: Eclipse.

Key Qualities: Organized, team player, quick learner, excellent management and interpersonal skill, motivated and detail oriented.

Others: Windows 7 Windows 8 PC support and troubleshooting. Microsoft server 2003 and 2008 configuration and support, Basic network configuration and troubleshooting, Management of a PC environment, Mobile device installations and support including Blackberry, Android, and iPhone, Hardware support and troubleshooting, Directory account management,Basic web presence management, Familiarity with Microsoft Exchange

PROFESSIONAL EXPERIENCE:

Hadoop Administrator

Confidential - Dallas TX

Responsibilities:

  • Functional and non-functional requirements gathering.
  • Install, Configure and maintain Hadoop on Multi clustered environment on Virtual Systems and worked with MapReduce, Hbase, Hive, Pig, Sqoop, Spark, Scalaand Pig,Latin, Flume, Zookeeper etc.
  • MapReduce code to process the data from social feeds which used to come in various formats like Json, TSV, CSV etc. and to load it into Database
  • Writing MapReduce programs with Java
  • Installed configured, monitored and did maintenance on the Cloudera distribution on Red Hat Enterprise Linux Version 7+
  • Involved in Hadoop cluster task like Adding and Removing Nodes without any effect to running jobs and data.
  • Hands on analyzing Data with HIVE and PIG
  • Work on in deployments and automation task.
  • Worked on evaluation and analysis of Hadoop cluster and different big data analytic tools including Pig, Hbase database and Sqoop.
  • Setup POC Hadoop cluster on Amazon EC2.
  • Importing and exporting data into HDFS and Hive using Sqoop
  • Developing MapReduce Program to transform/process data.
  • Installed and configured Flume, Hive, Pig, Sqoop, HBase on the Hadoop cluster.
  • Installed and configured Hive and also written Hive UDFs.
  • Writing the shell scripts automate the data flow from local file system to hdfs and then to Nosql databases (Hbase, Monbodb) and vice versa.
  • Installation/Configuration/Maintenance of Hadoop clusters for application development using Hadoop tools like Hive, Pig, Hbase, Sqoop and Flumewith Cloudera Management.
  • Used UDF's to implement business logic in Hadoop
  • Configured authentication and authorization with Kerberos Centrify, and LDAP
  • Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
  • Configuration Management System - Puppet 4+
  • Installed Oozie workflow engine to run multiple Hive and Pig jobs
  • Load log data into HDFS using Flume. Worked extensively in creating MapReduce jobs to power data for search and aggregation
  • Carried out Linux OS administration for Ubuntu version 14.04 and Red Hat 6.5
  • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing with Pig
  • Developed PIG Latin scripts to extract the data from the local file system output files to load into HDFS
  • Installed and configured Hadoop MapReduce, HDFS, Developed multiple map reduce jobs in java for data cleaning and preprocessing.

Hadoop Administrator

Confidential

Responsibilities:

  • Worked on analyzing Hadoop clusterusing different big data analytic tools including Pig, Hive, and MapReduce
  • Collecting and aggregating large amounts of log data using Apache Flume and staging data in HDFS for further analysis
  • Worked on debugging, performance tuning of Hive & Pig Jobs.
  • Created Hbase tables to store various data formats of PII data coming from different portfolios.
  • Implemented test scripts to support test driven development and continuous integration
  • Worked on tuning the performance Pig queries
  • Involved in loading data from LINUX file system to HDFS
  • Importing and exporting data into HDFS and Hive using Sqoop
  • Experience working on processing unstructured data using Pig and Hive
  • Supported MapReduce Programs those are running on the cluster
  • Gained experience in managing and reviewing Hadoop log files
  • Involved in schedulingOozie workflow engine to run multiple Hive and pig jobs
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts

Environment: Hadoop, HDFS, Pig, Hive, MapReduce, Sqoop, Oozie,, LINUX

Hadoop Administrator

Confidential

Responsibilities:

  • Installed and configured Hadoop Map Reduce, HDFS and Hadoop cluster.
  • Experienced in installing, configuring and using Hadoop Ecosystem components.
  • Experienced in Importing and exporting data into HDFS and Hive using Sqoop.
  • Knowledge in performance troubleshooting and tuning Hadoop clusters.
  • Participated in development/implementation of Cloudera Hadoop environment.
  • Implemented Partitioning, Dynamic Partitions and Buckets in HIVE for efficient data access.
  • Involved in various NOSQL databases like HBase, Cassandra in implementing and integration.
  • Installed and configured Hive and also written Hive UDFs and Used Map Reduce and Junit for unit testing.
  • Experienced in using Zookeeper and Oozie Operational Services for coordinating the cluster and scheduling workflows.
  • Installed and configured Hive and also written Hive UDFs and used piggy bank a repository of UDF's for PigLatin.
  • Experienced in managing and reviewing Hadoop log files.
  • Worked in installing cluster, commissioning & decommissioning of Datanodes, Namenode recovery, capacity planning, and slots configuration.
  • Supported Map Reduce Programs those are running on the cluster. Involved in loading data from UNIX file system to HDFS.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.

Environment: Hadoop, MapReduce, HDFS, Hive, pig, Impala, Cassandra, PIG, Zookeeper, Sqoop, Teradata

Confidential

Help Desk Support IT

Responsibilities:

  • rSupport all office technology and equipment, including troubleshooting workstations and workstation software.
  • Support video/IP conferencing and audio/visual systems.
  • Assist with telecommunications systems.
  • Support print services.
  • Provide event support for internal and client meetings.
  • Maintain and configured network equipment, including file servers, routers, and switches.
  • Monitor all critical business systems and applications for up-time and availability.
  • Provide excellent customer service, telephone support, and communication.
  • Troubleshoot any problems or issues that arise with Windows desktops/laptops from the network drop to the desktop/laptop prioritizing systems infected with viruses.
  • Respond to alert notification or escalation of an issue from members of the client team, determine the probable cause of the issue and take the appropriate intervention action(s) to restore the Windows desktop/laptop to operational status.
  • Document any troubleshooting and corrective actions taken to resolve Windows desktop/laptop issues in the Help Desk system.
  • Document status of tickets; opened, pending and closed.
  • Add virtual desktop client (CAG- Citrix Access Gateway) to existing office platform.
  • Load and reload, configure, and update anti-virus and virus scan software on desktop/laptop acquired.
  • Create and distribute "clean-up kits" (including instructions and other user documentation) for Windows remote laptops.
  • Load, configure, and update Virtual Private Network (VPN) and RESCUE software on desktop/laptop as required.
  • Respond to any of the service area locations for the purpose of dealing with hardware and/or software problems that cannot be solved through remote tools and that require the physical presence of a technician. (local travel)
  • Image, configure and deploy and install new devices to our customers.
  • Assist Help Desk technicians to provide Tier II support as needed.
  • Provide after hours and weekend support.

We'd love your feedback!