Hadoop Administrator Resume
West Lake, TX
SUMMARY:
- Over 7 years of professional IT experience, which includes experience in Big Data ecosystem, related technologies.
- 3+ years of experience in HadoopAdministration.
- Expertise in Big Data technologies and Hadoopecosystem: HDFS, Job Tracker, Task Tracker, NameNode, Data Node and MapReduce.
- Hands on experience in installing, configuring, and using Hadoopecosystem components like Hadoop MapReduce, HDFS, HBase, Oozie, Hive, Sqoop, Pig, Zookeeper and Flume.
- Hands on experience with Apache HadoopMap Reduce programming, PIG scripting and Distribute Application and HDFS.
- Hands on experience in installation, configuration, management and deployment of Big Data solutions and the underlying infrastructure of HadoopCluster using Cloudera, Hortonworks distributions.
- Excellent understanding of HadoopCluster architecture and monitoring the cluster.
- Installed, configured and managed Knox gateway and Ranger to provide perimeter security and central security policy administration across the Hadoop cluster.
- Experience in managing and reviewing Hadooplog files.
- Experience in NoSQLdatabase Mongodb and Cassandra.
- Hands on experience importing and exporting data using Sqoop from HDFS to Relational Database Systems.
- Managing the health of Cluster, resolving performance related issues, coordinating with various parties for Infrastructure support.
- Update Hadoop1.2.1 to 2.5.2
- Experience in Talend Big Data Studio 6.0.
- Push data as delimited files into HDFS using Talend Big Data studio.
- Expertise in Commissioning and decommissioning the nodes in HadoopCluster using Cloudera Manager Enterprise.
- Experience on Oracle OBIEE.
- Setting up HDFS Quotas to enforce the fair share of computing resources.
- Experience in Rebalance an HDFS Cluster.
- HadoopCluster capacity planning, performance tuning, cluster Monitoring, Troubleshooting.
- Hands on experience in analyzing Log files for Hadoopand eco system services and finding root cause.
- Expertise in benchmarking, performing backup and disaster recovery of Name Node metadata and important and sensitive data residing on cluster.
- Rack aware configuration for quick availability and processing of data.
- Experience in designing and implementing of secure Hadoopcluster using Kerberos.
- Successfully loaded files to Hive and HDFS from Oracle, SQL Server, MySQL, and Teradata using Sqoop.
- Loaded streaming log data from various web servers into HDFS using Flume.
- Created Hive internal and external tables defined with appropriate static and dynamic partitions.
- Experience in Creating and managing HBase clusters dynamically using Slider and Start & Stop HBase clusters running on Slider.
- Strong Knowledge on Spark concepts like RDD Operations, Caching and Persistence.
- Experience in Upgrading Apache Ambary, CDH and HDP Cluster.
- Extensive knowledge in using job scheduling by Oozie and Centralized Service Zookeeper.
- Expertise in Collaborating across Multiple technology groups and getting things done.
- Worked on both traditional Waterfall model and Agile methodology, Sound knowledge of Data warehousing concepts.
- Handsome experience in Linux admin activities.
- Supported technical team members for automation, installation and configuration tasks.
- Experienced in building servers like DHCP, PXE with kick - start, DNS, NFS and used them in building infrastructure in Linux Environment
- Worked in a 24x7 on-call Production Support Environment.
TECHNICAL SKILLS:
Hadoop: HDFS, YARN, Map-Reduce, Hive, Pig, Sqoop, Oozie, Flume, Zookeeper, Cloudera Manager, Ambari Manager, Hortonworks.
Markup Languages: HTML, CSS and XML
Application Servers: Apache Tomcat Server, Apache HTTP webserver
Operating Systems: (Redhat, CentOS & Ubuntu) Linux, UNIX, Windows
Scripting Languages: Shell Scripting, Python, Chef C, SQL, PLSQL, Java, PHP
Databases: Oracle, MySQL, MS SQL SERVER
PROFESSIONAL EXPERIENCE:
Confidential - West Lake, TX
Hadoop Administrator
Responsibilities:
- Installed and configured Hortonworks distribution of Hadoop Cluster using Ambari Manager.
- Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, and backup & DR systems.
- Involved in analyzing system failures, identifying root-cause and recommendation of remediation actions. Documented issue log with solutions for future s.
- Worked with systems engineering team for planning new Hadoop environment deployments, expansion of existing Hadoop clusters.
- Monitored multiple Hadoop clusters environments using Ganglia and Nagios. Monitoring workload, job performance and capacity planning using Cloudera Manager.
- Worked with application teams to install OS level updates, patches and version upgrades required for Hadoop cluster environments.
- Installed and configured Hbase, Hive, Pig, Sqoop, Kafka, Spark, Oozie and Storm on the Hadoop cluster.
- Installation and configuration of Name Node High Availability (NNHA) using Zookeeper.
- Analyzed web log data using the HiveQL to extract number of unique visitors per day, page views, visit duration, most purchased product on website.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports by our BI team.
- Experienced in Linux Administration tasks like IP Management (IP Addressing, Ethernet Bonding, Static IP and Subnetting).
- Principles, generating key tab file for each and every service and managing key tab using key tab tools.
- Worked on setting up Name Node high availability for major production cluster and designed Automatic failover control using zookeeper and quorum journal nodes.
- Worked on creation of Unix shell scripts to watch for 'null' files and trigger jobs accordingly and also had good knowledge in Python scripting language.
- Integrated Oozie with the rest of the Hadoop stack supporting several types of Hadoop jobs out of the box (such as Java like map-reduce, Streaming map-reduce, Hbase, Kafka, Spark, Pig, Hive, Sqoop and Distcp) as well as system specific jobs (such as Java programs and shell scripts).
- Worked on disaster management for Hadoop cluster.
- Involved in Installing and configuring Kerberos for the authentication of users and Hadoop daemons.
- Created a local YUM repository for installing and updating packages.
- Worked with the Linux administration team to prepare and configure the systems to support Hadoop deployment..
- Performed various configurations, which includes, networking and IPTable, resolving hostnames, user accounts and file permissions, http, ftp, SSH keyless login
Environment: Hortonworks Hadoop HDP2.3, Hive, Sqoop, Pig, Oozie, Zookeeper, Ganglia, Oozie, Spark, Hbase Ambari Manager, Kerberos, Knox, Kafka, Java, Linux (CentOS/REDHAT).
Confidential - Denver, CO
Hadoop Administration
Responsibilities:
- Installed and configured CDH5.0.0 cluster, using Cloudera manager.
- Implemented automatic failover zookeeper and zookeeper failover controller.
- Developed scripts for benchmarking with Terasort / Teragen.
- Monitored multiple Hadoop clusters environments using Ganglia and Nagios.
- Monitored workload, job performance and capacity planning.
- Managing and reviewing Hadoop log files and debugging failed jobs.
- Speculative execution of jobs running on the cluster to obtain optimal performance.
- Tuned the cluster by Commissioning and decommissioning the Data Nodes.
- Implemented Kerberos Security Authentication protocol for production cluster.
- Supported cluster maintenance, Backup and recovery for production cluster.
- Backed up data on regular basis to a remote cluster using distcp.
- Assisted the BI team by Partitioning and querying the data in Hive.
- Fine tuning of Hive jobs for better performance.
- Used Impala to read, write and query the Hadoop data in HDFS or HBASE
- Created Hbase tables to store various data formats of PII data coming from different portfolios.
- Automated all the jobs for pulling data from FTP server to load data into Hive tables, using Oozie workflows.
- Collected and aggregated large amounts of streaming data into HDFS using Flume and defined channel selectors to multiplex data into different sinks.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Expertise in Map-Reduce Framework, which include MR daemons, sorting and shuffle phase, task execution.
- Implemented Fair scheduler and capacity scheduler to allocate fair amount of resources to small jobs.
- Setting up the machines with Network Control, Static IP, Disabled Firewalls, Swap memory.
- Good experience in troubleshoot production level issues in the cluster and its functionality
Environment: Cloudera Hadoop HDFS, MAPREDUCE, hbase, hive, PIG, OOZIE, SQOOP, Nagios, Ganglia, CLOUDERA MANAGER.
Confidential - San Jose, CA
Linux Administrator
Responsibilities:
- Supporting the core Linux environment. This includes the administration, design, documentation and troubleshooting of the core Linux server infrastructure, communications network, and software environment
- Responsible for monitoring critical usage of resources like CPU, RAM & Hard disks and also monitoring security logs.
- Developing and maintaining automation systems for the Linux administrative efficiency improvement.
- Participating in network management and maintenance. Configured local and network file sharing.
- Within Unix/Linux environment, responsible for provisioning new servers, monitoring, automation, imaging, disaster recovery (planning and testing), scripting backup\recovery of bare metal hardware and virtual machines.
- Managed load firewalls, balancers, and other networking equipment in a production environment.
- Maintained data files, directory structure, monitor systems configuration and ensure data integrity.
- Integrated monitoring, auditing, and alert systems for databases with existing monitoring infrastructure.
- Responsible for setting up FTP, DHCP, DNS servers and Logical Volume Management.
- Managed network fine-tuning, upgrades and enhances to optimize Network performance, availability, stability and security.
- Provided authentication to users for Oracle databases.
Environment: s: Linux Systems (CentOS & Redhat), Oracle, DHCP, DNS, Logical Volume Manager, User Management.
Confidential - Denver, CO
Linux/System Admin/DBA
Responsibilities:
- The main goal of the Project is to integrate Provider, Hospital and Confidential t data streams in to FACETS application.
- Installing and maintaining the Red hat and Centos Linux servers.
- Installed centos using Pre-Execution environment boot and kick-start method on multiple servers.
- Responsible for performance tuning and troubleshooting Linux servers.
- Running crontab to back up data.
- Adding, removing, updating user account information, resetting passwords, etc.
- Maintaining the SQL server and Authentication to required users for databases.
- Applied Operating System updates, patches and configuration changes.
- Used different methodologies to increase the performance and reliability of the IT infrastructure.
- Responsible for System performance tuning and successfully engineered a virtual private network (VPN).
- Installing, configuring and maintaining SAN and NAS storage.
- Performed system administration of UNIX servers by using Operating Systems of RHEL 4.x, 5.x.
- Installation of patches and other software packages
- Experience in designing and implementing SAN storage for optimal performance.
- Setting up the storage to dynamically allocate/de-allocate disk from systems. Configuration of different RAID levels and multi-path.
- Installation of Oracle Patches and Troubleshooting, Creating and modifying application related objects, Creating Profiles, Users, Roles and maintaining system security.
- Configuration and maintenance of NIS, NFS, DHCP and DNS Servers on RHEL 5.0.
- Administrative tasks such as System Startup/shutdown, Backup strategy, Printing, Documentation, User Management, Security, Network management, dumb terminals and devices carried out.
- Installation and configuration of Oracle 7.x/8.x requisite parameters
- Performed the administration of the DNS and configuration of the Database Files.
Environment: Red-Hat Enterprise Linux, HP Proliant DL 585, BL 465/485,ML Series, SAN(Netapp), BladeLogic, Veritas Cluster Server 5.0, Windows 2003 server, Shell scripting, Jboss 4.2, VMware Virtual Client 3.5, VMware Infrastructure 3.5.
Confidential
Linux Admin
Responsibilities:
- Worked on Installation, configuration and upgrading of Oracle server software and related products.
- Responsible for installation, administration and maintenance of Linux servers.
- Established and maintain sound backup and recovery policies and procedures.
- Take care of the Database design and implementation.
- Implement and maintain database security (create and maintain users and roles, assign privileges).
- Performed database tuning and performance monitoring.
- Plan growth and changes (capacity planning).
- Support pre-production and production support teams in the analysis of critical services and assists with maintenance operations
- Creating and managing file systems, Creating and managing Logical volumes in Linux and VERITAS Volume manager in Solaris & Linux.
- Installation of patches and packages using RPM and YUM in Red hat Linux and Suse Linux and also using patchadd and pkgadd in Solaris 10 Operating System.
- Usage of Service Management Facility for enabling, disabling and also troubleshooting services in maintenance mode.
- Modified Kernel parameters to improve the server performance in both Solaris and Linux.
- Providing Statistics on Hard Disk Usage.
- Configuration of Clusters using VERITAS Cluster Services in Solaris and Linux.
- Backups of systems and restoration using Net Backup on both Solaris and Linux.
- Creating Disk Volumes, Installed and Configured RAID controllers RAID 0, RAID 1 and RAID 5.
- Installation of HBA drivers and Upgrading of HBA cards before the move.
- Installed, configured and maintained Apache/Tomcat Web Server.
- Installed and configured NFS, NIS and FTP Servers.
- Bash Shell Scripting for job automation.
- Understanding and experience on TCP/IP networking in Solaris and Linux Systems.
- Experience with VMware ESX Server 3.5.
- Performance tuning and monitoring of Servers in Solaris, Linux.
- Supporting Customers on a real time remedy ticketing System for solving online operational issues.
- Experience with the Change Management tool for reporting and resolving issues.
- Worked as part of a team and provide 7x24 support when required.
- Performed general technical trouble shooting on trouble tickets to bring to resolution.
- Interfaced with Confidential for technical support.
- Patch Management and Version Control.
Environment: Solaris10/8, RedHat Enterprise Linux 4/5, Windows 2000, Sun SPARC Enterprise T5120/ T5220/ T5240, VERITAS Volume Manager/Cluster, SVM/LVM, VMware ESX Server, NFS, Samba, OpenLDAP, SQUID, Oracle 8i, bash, shell, bugzilla
Confidential
Linux/System Administrator
Responsibilities:
- Connected to RHEL 5/6 Linux physical and virtual servers, troubleshoot Apache and MySQL issues as well as other common website-hosting issues such as disk space usage, memory usage, CPU usage; determine root cause and recommend possible changes to the customer.
- Performed patching maintenances on customer servers with complete documentation of changes made
- Installed and maintained software packages (.rpm) as necessary.
- Administration of VMware virtual Linux servers.
- Troubleshoot and resolve Backup issues with Holland MySQL backup service.
- Resize LVM disk volumes as needed.
- Created BASH utility scripts as needed to help automate tasks and determine needed information.
- Implemented proxy server for an extensive ESX virtual server environment.
- Implemented DHCP server for a development lab.
- Linux / VMware support on multiple VMware ESXI hosts - diagnose, troubleshoot, optimization, stabilization and migration of various Linux virtual machines running Apache 2.0 webserver, MySQL, backup to "cloud" storage.
- Migrate virtual Apache hosts between ESXi servers / Linux virtual guests (without using Vmotion) for load balancing.
- Developed automated scripts for performing regular Linux tasks.
Environment: Solaris 9, Linux (RHEL 4/5), SUSE 11/10, AIX 6.1, Oracle 11/10g, SUN Servers, SUN Fires, Jboss 5.0, WebSphere Application Server 5.x/6.0/6.1, VXVM, LDOM’s, F5 Loadbalancers, LDAP I-planet Directory Server6.0, EMC Storage SAN, Veritas Cluster Server (VCS), VMWare servers, and Apache.