We provide IT Staff Augmentation Services!

Sr. Hadoop Administrator Resume

4.00/5 (Submit Your Rating)

San Jose, CA

SUMMARY

  • Senior Hadoop big data Administrator passionate about implementation and ongoing administration of Hadoop infrastructure and services to improve Confidential 's life. Open source enthusiast.
  • Over 9+ years of comprehensive IT experience in BigData and Big DataAnalytics, Hadoop, HDFS, Map Reduce, YARN, Hadoop Ecosystem and ShellScripting. Highly capable for processing large sets of Structured, Semi - structured and Unstructured datasets and supporting BigData applications.
  • Experience in configuring, installing and managing MapR, Hortonworks& Cloudera Distributions.
  • Hands on experience in installing, configuring, monitoring and using Hadoop components like Hadoop MapReduce, HDFS, HBase, Hive, Sqoop, Pig, Zookeeper, Hortonworks, Oozie, Apache Spark, Impala.
  • Working experience with large scale Hadoop environments build and support including design, configuration, installation, performance tuning and monitoring.
  • Working knowledge of monitoring tools and frameworks such as Splunk, Influx DB, Prometheus, SysDig, Data Dog, App-Dynamics, New Relic, and Nagios.
  • Experience in setting up automated monitoring and escalation infrastructure for Hadoop Cluster using Ganglia and Nagios.
  • Standardize Splunk forwarder deployment, configuration and maintenance across a variety of Linux platforms. Also worked on Devops tools like Puppet and GIT.
  • Hands on experience on configuring a Hadoop cluster in a professional environment and on Amazon Web Services (AWS) using an EC2 instance.
  • Experience with complete Software Design Lifecycle including design, development, testing and implementation of moderate to advanced complex systems.
  • Hands on experience in installation, configuration, supporting and managing Hadoop Clusters using Apache, Hortonworks,Cloudera and Map Reduce.
  • Extensive experience in installing, configuring and administrating Hadoop cluster for major Hadoop distributions like CDH5 and HDP.
  • Experience in Ranger, Knox configuration to provide the security for Hadoop services (hive, base, hdfs etc.).Experience in administration of Kafka and Flume streaming using Cloudera Distribution.
  • Developed automated scripts using Unix Shell for performing RUNSTATS, REORG, REBIND, COPY, LOAD, BACKUP, IMPORT, EXPORT and other related to database activities.
  • Installed, configured, and maintained several Hadoop clusters which includes HDFS, YARN, Hive, HBase, Knox, Kafka, Oozie, Ranger, Atlas, Infra Solr, Zookeeper, and Nifi in Kerberized environments.
  • Experienced with deployments, maintenance and troubleshooting applications on Microsoft Azure Cloud infrastructure.Excellent knowledge of NOSQL databases like HBase, Cassandra.

TECHNICAL SKILLS

  • BIG Data Ecosystem: HDFS, MapReduce, Spark, Pig, Hive, Hbase, sqoop, zookeeper, Sentry, Ranger, Storm, Kafka, Oozie, flume, Hue, Knox, NiFi
  • BIG Data Security: Kerberos, AD, LDAP, KTS, KMS, Redaction, Sentry, Ranger, Navencrypt, SSL/TLS, Cloudera Navigator
  • No SQL Databases: HBase, Cassandra, MongoDB
  • Programming Languages: Java, Scala, Python, SQL, PL/SQL, Hive - QL, Pig Latin
  • Frameworks: MVC, Struts, Spring, Hibernate
  • Web Technologies: HTML, DHTML, XML, AJAX, WSDL, SOAP
  • Web/Application servers: Apache Tomcat, WebLogic, JBoss
  • Version control: SVN, CVS, GIT

PROFESSIONAL EXPERIENCE

Confidential, San Jose, CA

Sr. Hadoop Administrator

Responsibilities:

  • Experience in architecting, designing, installation, configuration and management of Apache Hadoop Clusters, MapR, and Hortonworks& Cloudera Hadoop Distribution.
  • Responsible for installing, configuring, supporting and managing of Cloudera Hadoop Clusters.
  • Installed Kerberos secured Kafka cluster with no encryption on POC also set up Kafka ACL's
  • Created NoSQL solution for a legacy RDBMS Using Kafka, Spark, SOLR, and HBase indexer for ingestion SOLR and HBase for and real-time querying
  • Experienced in Administration, Installing, Upgrading and Managing distributions of Hadoop clusters with MapR 5.1 on a cluster of 200+ nodes in different environments such as Development, Test and Production (Operational & Analytics) environments.
  • Troubleshooting issues in the execution of MapReduce jobs by inspecting and reviewing log files.
  • Extensively worked on Elastic search querying and indexing to retrieve the documents in high speeds.
  • Installed, configured, and maintained several Hadoop clusters which includes HDFS, YARN, Hive, HBase, Knox, Kafka, Oozie, Ranger, Atlas, Infra Solr, Zookeeper, and Nifi in Kerberized environments.
  • Involved in deploying a Hadoop cluster using Hortonworks Ambari HDP 2.2 integrated with Sitescope for monitoring and Alerting.
  • Install OS and administrated Hadoop stack with CDH5.9 (with YARN) Cloudera Distribution including configuration management, monitoring, debugging, and performance tuning.

Environment: HBase,Hadoop 2.2.4, Hive, Kerberos,Kafka, YARN, Spark, Impala, SOLR, Java Hadoop cluster, HDFS, Ambari, Ganglia, CentOS, RedHat, Windows, MapR, Yarn, Sqoop, Cassandra.

Confidential, Rochester NY

Sr. Hadoop Administrator

Responsibilities:

  • Creating end to end Spark applications using Scala to perform various data cleansing, validation, transformation and summarization activities on user behavioral data.
  • Installed and configured Hadoop multi-node cluster and maintenances by Nagios.
  • Developed custom FTP adaptors to pull the clickstream data from FTP servers to HDFS directly using HDFS File System API.
  • Worked on analyzing Hortonworks Hadoop cluster and different big data analytic tools including Pig, HBase Database and Sqoop.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Experienced on setting up Horton works cluster and installing all the ecosystem components through Ambariand manually from command line.
  • Performed a Major upgrade in production environment from HDP 1.3 to HDP 2.2 and followed standard Back up policies to make sure the high availability of cluster.

Environment: Ambari, Hadoop, Yarn, Spark, Kafka, Hive, Pig, Sqoop, Kerberos, Ranger, NiFi, Oracle, Netezza, Tableau, Python, Log4J,HBase, Scala,Kafka, Zookeeper, GIT, AWS,, Zookeeper, S3, EC2, EMR, JIRA,AWS UNIX Shell Scripting..

Confidential, Plano TX

Sr. Hadoop Administrator

Responsibilities:

  • Worked on setting up cluster for various services such as Hadoop, Spark, HBase, Kafkaetc on Azure HDInsights.
  • Installing and Working on Hadoop clusters for different teams, supported 50+ users to use Hadoop platform and resolve tickets and issues they run into and provide training to users to make Hadoop usability simple and updating them for best practices.
  • Involved in collecting and aggregating large amounts of log data using Apache Flume and staging data in HDFS for further analysis.
  • Upgraded the Hadoop cluster CDH5.8 to CDH 5.9.
  • Worked on Installing cluster, Commissioning & Decommissioning of DataNodes, NameNode Recovery, Capacity Planning, and Slots Configuration.
  • Working on Oracle Big Data SQL. Integrate big data analysis into existing applications
  • Using Oracle Big Data Appliance Hadoop and NoSQL processing and also integrating data inHadoop and NoSQL with data in Oracle Database
  • Involved in performance tuning of various hadoop ecosystem components like YARN, MRv2.
  • Maintains and monitors database security, integrity, and access controls. Provides audit trails to detect potential security violations.
  • Monitored cluster for performance, networking, and data integrity issues.

Environment: MapReduce, Hive 0.13.1, PIG 0.16.0, Sqoop 1.4.6, Spark 2.1, Oozie 4.1.0, Flume, HBase 1.0, Oracle Server X6, SQL Server, Solr, Kerberos and RedHat 6.5.

We'd love your feedback!