We provide IT Staff Augmentation Services!

Hadoop Administrator Resume

0/5 (Submit Your Rating)

West Lake, TX

SUMMARY:

  • Over 7 years of professional IT experience, which includes experience in Big Data ecosystem, related technologies.
  • 3+ years of experience in HadoopAdministration.
  • Expertise in Big Data technologies and Hadoopecosystem: HDFS, Job Tracker, Task Tracker, NameNode, Data Node and MapReduce.
  • Hands on experience in installing, configuring, and using Hadoopecosystem components like Hadoop MapReduce, HDFS, HBase, Oozie, Hive, Sqoop, Pig, Zookeeper and Flume.
  • Hands on experience with Apache HadoopMap Reduce programming, PIG scripting and Distribute Application and HDFS.
  • Hands on experience in installation, configuration, management and deployment of Big Data solutions and the underlying infrastructure of HadoopCluster using Cloudera, Hortonworks distributions.
  • Excellent understanding of HadoopCluster architecture and monitoring the cluster.
  • Installed, configured and managed Knox gateway and Ranger to provide perimeter security and central security policy administration across the Hadoop cluster.
  • Experience in managing and reviewing Hadooplog files.
  • Experience in NoSQLdatabase Mongodb and Cassandra.
  • Hands on experience importing and exporting data using Sqoop from HDFS to Relational Database Systems.
  • Managing the health of Cluster, resolving performance related issues, coordinating with various parties for Infrastructure support.
  • Update Hadoop1.2.1 to 2.5.2
  • Experience in Talend Big Data Studio 6.0.
  • Push data as delimited files into HDFS using Talend Big Data studio.
  • Expertise in Commissioning and decommissioning the nodes in HadoopCluster using Cloudera Manager Enterprise.
  • Experience on Oracle OBIEE.
  • Setting up HDFS Quotas to enforce the fair share of computing resources.
  • Experience in Rebalance an HDFS Cluster.
  • HadoopCluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
  • Hands on experience in analyzing Log files for Hadoopand eco system services and finding root cause.
  • Expertise in benchmarking, performing backup and disaster recovery of Name Node metadata and important and sensitive data residing on cluster.
  • Rack aware configuration for quick availability and processing of data.
  • Experience in designing and implementing of secure Hadoopcluster using Kerberos.
  • Successfully loaded files to Hive and HDFS from Oracle, SQL Server, MySQL, and Teradata using Sqoop.
  • Loaded streaming log data from various web servers into HDFS using Flume.
  • Created Hive internal and external tables defined with appropriate static and dynamic partitions.
  • Experience in Creating and managing HBase clusters dynamically using Slider and Start & Stop HBase clusters running on Slider.
  • Strong Knowledge on Spark concepts like RDD Operations, Caching and Persistence.
  • Experience in Upgrading Apache Ambary, CDH and HDP Cluster.
  • Extensive knowledge in using job scheduling by Oozie and Centralized Service Zookeeper.
  • Expertise in Collaborating across Multiple technology groups and getting things done.
  • Worked on both traditional Waterfall model and Agile methodology, Sound knowledge of Data warehousing concepts.
  • Handsome experience in Linux admin activities.
  • Supported technical team members for automation, installation and configuration tasks.
  • Experienced in building servers like DHCP, PXE with kick - start, DNS, NFS and used them in building infrastructure in Linux Environment
  • Worked in a 24x7 on-call Production Support Environment.

TECHNICAL SKILLS:

Hadoop: HDFS, YARN, Map-Reduce, Hive, Pig, Sqoop, Oozie, Flume, Zookeeper, Cloudera Manager, Ambari Manager, Hortonworks.

Markup Languages: HTML, CSS and XML

Application Servers: Apache Tomcat Server, Apache HTTP webserver

Operating Systems: (Redhat, CentOS & Ubuntu) Linux, UNIX, Windows

Scripting Languages: Shell Scripting, Python, Chef C, SQL, PLSQL, Java, PHP

Databases: Oracle, MySQL, MS SQL SERVER

PROFESSIONAL EXPERIENCE:

Confidential - West Lake, TX

Hadoop Administrator

Responsibilities:

  • Installed and configured Hortonworks distribution of Hadoop Cluster using Ambari Manager.
  • Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, and backup & DR systems.
  • Involved in analyzing system failures, identifying root-cause and recommendation of remediation actions. Documented issue log with solutions for future s.
  • Worked with systems engineering team for planning new Hadoop environment deployments, expansion of existing Hadoop clusters.
  • Monitored multiple Hadoop clusters environments using Ganglia and Nagios. Monitoring workload, job performance and capacity planning using Cloudera Manager.
  • Worked with application teams to install OS level updates, patches and version upgrades required for Hadoop cluster environments.
  • Installed and configured Hbase, Hive, Pig, Sqoop, Kafka, Spark, Oozie and Storm on the Hadoop cluster.
  • Installation and configuration of Name Node High Availability (NNHA) using Zookeeper.
  • Analyzed web log data using the HiveQL to extract number of unique visitors per day, page views, visit duration, most purchased product on website.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports by our BI team.
  • Experienced in Linux Administration tasks like IP Management (IP Addressing, Ethernet Bonding, Static IP and Subnetting).
  • Principles, generating key tab file for each and every service and managing key tab using key tab tools.
  • Worked on setting up Name Node high availability for major production cluster and designed Automatic failover control using zookeeper and quorum journal nodes.
  • Worked on creation of Unix shell scripts to watch for 'null' files and trigger jobs accordingly and also had good knowledge in Python scripting language.
  • Integrated Oozie with the rest of the Hadoop stack supporting several types of Hadoop jobs out of the box (such as Java like map-reduce, Streaming map-reduce, Hbase, Kafka, Spark, Pig, Hive, Sqoop and Distcp) as well as system specific jobs (such as Java programs and shell scripts).
  • Worked on disaster management for Hadoop cluster.
  • Involved in Installing and configuring Kerberos for the authentication of users and Hadoop daemons.
  • Created a local YUM repository for installing and updating packages.
  • Worked with the Linux administration team to prepare and configure the systems to support Hadoop deployment..
  • Performed various configurations, which includes, networking and IPTable, resolving hostnames, user accounts and file permissions, http, ftp, SSH keyless login

Environment: Hortonworks Hadoop HDP2.3, Hive, Sqoop, Pig, Oozie, Zookeeper, Ganglia, Oozie, Spark, Hbase Ambari Manager, Kerberos, Knox, Kafka, Java, Linux (CentOS/REDHAT).

Confidential - Denver, CO

Hadoop Administration

Responsibilities:

  • Installed and configured CDH5.0.0 cluster, using Cloudera manager.
  • Implemented automatic failover zookeeper and zookeeper failover controller.
  • Developed scripts for benchmarking with Terasort / Teragen.
  • Monitored multiple Hadoop clusters environments using Ganglia and Nagios.
  • Monitored workload, job performance and capacity planning.
  • Managing and reviewing Hadoop log files and debugging failed jobs.
  • Speculative execution of jobs running on the cluster to obtain optimal performance.
  • Tuned the cluster by Commissioning and decommissioning the Data Nodes.
  • Implemented Kerberos Security Authentication protocol for production cluster.
  • Supported cluster maintenance, Backup and recovery for production cluster.
  • Backed up data on regular basis to a remote cluster using distcp.
  • Assisted the BI team by Partitioning and querying the data in Hive.
  • Fine tuning of Hive jobs for better performance.
  • Used Impala to read, write and query the Hadoop data in HDFS or HBASE
  • Created Hbase tables to store various data formats of PII data coming from different portfolios.
  • Automated all the jobs for pulling data from FTP server to load data into Hive tables, using Oozie workflows.
  • Collected and aggregated large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Expertise in Map-Reduce Framework, which include MR daemons, sorting and shuffle phase, task execution.
  • Implemented Fair scheduler and capacity scheduler to allocate fair amount of resources to small jobs.
  • Setting up the machines with Network Control, Static IP, Disabled Firewalls, Swap memory.
  • Good experience in troubleshoot production level issues in the cluster and its functionality

Environment: Cloudera Hadoop HDFS, MAPREDUCE, hbase, hive, PIG, OOZIE, SQOOP, Nagios, Ganglia, CLOUDERA MANAGER.

Confidential - San Jose, CA

Linux Administrator

Responsibilities:

  • Supporting the core Linux environment. This includes the administration, design, documentation and troubleshooting of the core Linux server infrastructure, communications network, and software environment
  • Responsible for monitoring critical usage of resources like CPU, RAM & Hard disks and also monitoring security logs.
  • Developing and maintaining automation systems for the Linux administrative efficiency improvement.
  • Participating in network management and maintenance. Configured local and network file sharing.
  • Within Unix/Linux environment, responsible for provisioning new servers, monitoring, automation, imaging, disaster recovery (planning and testing), scripting backup\recovery of bare metal hardware and virtual machines.
  • Managed load firewalls, balancers, and other networking equipment in a production environment.
  • Maintained data files, directory structure, monitor systems configuration and ensure data integrity.
  • Integrated monitoring, auditing, and alert systems for databases with existing monitoring infrastructure.
  • Responsible for setting up FTP, DHCP, DNS servers and Logical Volume Management.
  • Managed network fine-tuning, upgrades and enhances to optimize Network performance, availability, stability and security.
  • Provided authentication to users for Oracle databases.

Environment: s: Linux Systems (CentOS & Redhat), Oracle, DHCP, DNS, Logical Volume Manager, User Management.

Confidential - Denver, CO

Linux/System Admin/DBA

Responsibilities:

  • The main goal of the Project is to integrate Provider, Hospital and Confidential t data streams in to FACETS application.
  • Installing and maintaining the Red hat and Centos Linux servers.
  • Installed centos using Pre-Execution environment boot and kick-start method on multiple servers.
  • Responsible for performance tuning and troubleshooting Linux servers.
  • Running crontab to back up data.
  • Adding, removing, updating user account information, resetting passwords, etc.
  • Maintaining the SQL server and Authentication to required users for databases.
  • Applied Operating System updates, patches and configuration changes.
  • Used different methodologies to increase the performance and reliability of the IT infrastructure.
  • Responsible for System performance tuning and successfully engineered a virtual private network (VPN).
  • Installing, configuring and maintaining SAN and NAS storage.
  • Performed system administration of UNIX servers by using Operating Systems of RHEL 4.x, 5.x.
  • Installation of patches and other software packages
  • Experience in designing and implementing SAN storage for optimal performance.
  • Setting up the storage to dynamically allocate/de-allocate disk from systems. Configuration of different RAID levels and multi-path.
  • Installation of Oracle Patches and Troubleshooting, Creating and modifying application related objects, Creating Profiles, Users, Roles and maintaining system security.
  • Configuration and maintenance of NIS, NFS, DHCP and DNS Servers on RHEL 5.0.
  • Administrative tasks such as System Startup/shutdown, Backup strategy, Printing, Documentation, User Management, Security, Network management, dumb terminals and devices carried out.
  • Installation and configuration of Oracle 7.x/8.x requisite parameters
  • Performed the administration of the DNS and configuration of the Database Files.

Environment: Red-Hat Enterprise Linux, HP Proliant DL 585, BL 465/485,ML Series, SAN(Netapp), BladeLogic, Veritas Cluster Server 5.0, Windows 2003 server, Shell scripting, Jboss 4.2, VMware Virtual Client 3.5, VMware Infrastructure 3.5.

Confidential

Linux Admin

Responsibilities:

  • Worked on Installation, configuration and upgrading of Oracle server software and related products.
  • Responsible for installation, administration and maintenance of Linux servers.
  • Established and maintain sound backup and recovery policies and procedures.
  • Take care of the Database design and implementation.
  • Implement and maintain database security (create and maintain users and roles, assign privileges).
  • Performed database tuning and performance monitoring.
  • Plan growth and changes (capacity planning).
  • Support pre-production and production support teams in the analysis of critical services and assists with maintenance operations
  • Creating and managing file systems, Creating and managing Logical volumes in Linux and VERITAS Volume manager in Solaris & Linux.
  • Installation of patches and packages using RPM and YUM in Red hat Linux and Suse Linux and also using patchadd and pkgadd in Solaris 10 Operating System.
  • Usage of Service Management Facility for enabling, disabling and also troubleshooting services in maintenance mode.
  • Modified Kernel parameters to improve the server performance in both Solaris and Linux.
  • Providing Statistics on Hard Disk Usage.
  • Configuration of Clusters using VERITAS Cluster Services in Solaris and Linux.
  • Backups of systems and restoration using Net Backup on both Solaris and Linux.
  • Creating Disk Volumes, Installed and Configured RAID controllers RAID 0, RAID 1 and RAID 5.
  • Installation of HBA drivers and Upgrading of HBA cards before the move.
  • Installed, configured and maintained Apache/Tomcat Web Server.
  • Installed and configured NFS, NIS and FTP Servers.
  • Bash Shell Scripting for job automation.
  • Understanding and experience on TCP/IP networking in Solaris and Linux Systems.
  • Experience with VMware ESX Server 3.5.
  • Performance tuning and monitoring of Servers in Solaris, Linux.
  • Supporting Customers on a real time remedy ticketing System for solving online operational issues.
  • Experience with the Change Management tool for reporting and resolving issues.
  • Worked as part of a team and provide 7x24 support when required.
  • Performed general technical trouble shooting on trouble tickets to bring to resolution.
  • Interfaced with Confidential for technical support.
  • Patch Management and Version Control.

Environment: Solaris10/8, RedHat Enterprise Linux 4/5, Windows 2000, Sun SPARC Enterprise T5120/ T5220/ T5240, VERITAS Volume Manager/Cluster, SVM/LVM, VMware ESX Server, NFS, Samba, OpenLDAP, SQUID, Oracle 8i, bash, shell, bugzilla

Confidential

Linux/System Administrator

Responsibilities:

  • Connected to RHEL 5/6 Linux physical and virtual servers, troubleshoot Apache and MySQL issues as well as other common website-hosting issues such as disk space usage, memory usage, CPU usage; determine root cause and recommend possible changes to the customer.
  • Performed patching maintenances on customer servers with complete documentation of changes made
  • Installed and maintained software packages (.rpm) as necessary.
  • Administration of VMware virtual Linux servers.
  • Troubleshoot and resolve Backup issues with Holland MySQL backup service.
  • Resize LVM disk volumes as needed.
  • Created BASH utility scripts as needed to help automate tasks and determine needed information.
  • Implemented proxy server for an extensive ESX virtual server environment.
  • Implemented DHCP server for a development lab.
  • Linux / VMware support on multiple VMware ESXI hosts - diagnose, troubleshoot, optimization, stabilization and migration of various Linux virtual machines running Apache 2.0 webserver, MySQL, backup to "cloud" storage.
  • Migrate virtual Apache hosts between ESXi servers / Linux virtual guests (without using Vmotion) for load balancing.
  • Developed automated scripts for performing regular Linux tasks.

Environment: Solaris 9, Linux (RHEL 4/5), SUSE 11/10, AIX 6.1, Oracle 11/10g, SUN Servers, SUN Fires, Jboss 5.0, WebSphere Application Server 5.x/6.0/6.1, VXVM, LDOM’s, F5 Loadbalancers, LDAP I-planet Directory Server6.0, EMC Storage SAN, Veritas Cluster Server (VCS), VMWare servers, and Apache.

We'd love your feedback!