We provide IT Staff Augmentation Services!

Splunk And Big Data Developer Resume

5.00/5 (Submit Your Rating)

Phoenix, AZ

SUMMARY:

  • Over 7+ years of IT work experience with 5+ years of relevant experience in the field of Big Data, Hadoop, Hive, Pig, Java, Spark, Sqoop related technologies, 2+ year experience in Oracle Technologies and 8+ months of experience on Java development
  • Seeking a challenging position in Software development industry that needs innovation, creativity and dedication and enable me to continue to work in a demanding and fast paced environment, leveraging my current knowledge and fostering creativity with many learning opportunities.
  • Problem solving capability with main design goals of ER modeling for OLTP and OLAP. And, implemented the best solutions suitable for the business needs. Developed core modules in large cross - platform applications using JAVA, J2EE, JVM.
  • Hands-On Experience in Hadoop/HDFS, MapReduce, Hive, HBase, Pig, Sqoop, Amazon Elastic Map Reduce (EMR), Spark, Cloudera (CDH 3 & 4), Hortonworks sandbox environments
  • Having rich experience on big data, data warehouse, database, and business intelligence experience using the following technologies: Oracle Database (10g, 11g). And, involved in the project activities related to Requirements Gathering, Systems Analysis and Design, Code Generation, Testing, Implementation, Support and Maintenance.
  • Experience in developing and automating application’s using Unix Shell Scripting in the field of Big Data using Map-Reduce Programming for batch processing of jobs on a HDFS cluster, Hive and Pig.
  • Developed real-time Big Data solutions using No-SQL Column Oriented Databases using Hbase, Cassandra, MongoDB, CouchDB which can handle petabytes of data at one time.
  • Working on Spark and Storm environments using Scala programming. Involved in the operations of Spark Streaming, Spark SQL, Scala Programming and Performance Tuning
  • Involved in the activities of RDD Creations and Operations of Data Frames and Datasets for the Use-Case.
  • Worked on HBase Shell, CQL, HBase API and Cassandra Hector API developing ingestion and clustering frameworks with respect to Kafka, Zookeeper, YARN, Spark, Mesos and Kafka.
  • Capturing data from existing databases that provide SQL interfaces using Sqoop and processing stream data using Kafka, Storm, Spark Streaming and Flume.
  • Hands-On experience in setting up Zookeeper for providing High Availability to clusters. Hands on programming with Oozie and having Good knowledge of programming with log data using Apache Flume.
  • Developed Python based API for converting the files to Key-Value pairs for getting the files sourced to the Splunk Forwarder.
  • Developed a fully automated continuous integration system using Git, Jenkins, Splunk, Hunk, Oracle and custom tools developed in Python and Bash.
  • Strong experience in RDBMS using Oracle 10g, SQL Server, PL-SQL programming, schema development, Oracle fine tuning.
  • Active participation in tomcat server and Oracle problems (killing instances, debugging server logs, applications logs).
  • Written SQL queries, stored procedures, modifications to existing database structure as required per addition of new features.
  • Designed and developed Enterprise Eligibility business objects and domain objects with Object Relational Mapping framework such as Hibernate.
  • Experienced in design and development of various web and enterprise applications using J2EE technologies like JSP, Servlets, JSF, EJB, JDBC, Hibernate, iBatis, Spring MVC, Spring Batch, JPA, XML, JAXB, JSTL, JAX-WS, AJAX, ANT and Web Services (SOA, SOAP, REST, WSDL).
  • Experienced in WEB and GUI development using HTML, DHTML, XHTML, CSS, JavaScript, JSP, Angular JS, Backbone JS, JQuery, DOJO, POJO, and AJAX technologies.
  • Working knowledge of Struts3.0, Hibernate 3.0 persistence technology, Spring Framework 2.5, 3.0, 4.0.
  • Hands-on Experience in Spark and Storm environments using Scala programming. Involved in the operations of Spark Streaming, Spark SQL, Scala Programming and Performance Tuning
  • Involved in the activities of RDD Creations and Operations of Data Frames and Datasets for the Use-Case
  • Led team to plan, design, and implement applications and software.
  • Collaborated with business analysts, developers, and technical support teams to define project requirements and specifications. Designed, developed, and managed map-reduce-based applications, integrating with databases, establishing network connectivity, and developing programs.
  • Good knowledge at SOAP/WSDL and REST FUL interfaces in Java. Created and executed both load and functional tests for web services.
  • Assisted project manager in defining project scope, time & effort estimates and deliverable management.
  • Developed Python based API for converting the files to Key-Value pairs for the files getting sourced to the Splunk Forwarder.
  • Developed a fully automated continuous integration system using Git, Jenkins, Splunk, Oracle and custom tools developed in Python and Bash.
  • Developed a proof of concept for using Spark and Kafka to store and process data
  • Capturing data from existing databases that provide SQL interfaces using Sqoop.
  • Importing and exporting data in HDFS and Hive using Sqoop

TECHNICAL SKILLS:

Big Data/ Hadoop: Cloudera CDH 5.1.3, Hortonworks HDP 2.0, MongoDB, Python, shell script, Hadoop, HDFS, MapReduce (MRV1, MRV2 YARN), HBase, Pig, Hive, Sqoop, Flume, ZooKeeper, Oozie, Lucene, Cassandra, CouchDB, MongoDB, Kafka, Scala, R, Kafka

Languages: Java, C, HTML, SQL, PL/SQL, Scala

OS: Windows 8, Windows 7, Windows XP/98, UNIX/LINUX, MAC

Databases: Oracle (SQL / PLSQL), MySQL, NoSQL, Teradata

Web Technologies: HTML, DHTML, XML, WSDL, SOAP, Joomla, Apache Tomcat

Databases: Oracle 9g/10g/11g, SQL Server, MS Access

Build Tools: Ant, Maven

Development Tools: Adobe Photoshop, Adobe Illustrator, Eclipse, Linux/Mac OS environment, MS Visio, Crystal Reports

Business Domains: Distributed Systems, Online advertising, Social media advertising

Data Analytics: Python, R

ETL Tools: Talend/Informatica

PROFESSIONAL EXPERIENCE:

Confidential, Phoenix, AZ

Splunk and Big Data Developer

Job Responsibilities:

  • Participate in business and system requirements sessions
  • Involved in sourcing the data files for different Use-Cases
  • Prepared design documents based on the understanding of Requirement Elicitation to technical specification for the different modules
  • Provided inputs on solution architecture based on solution alternatives, frameworks, products
  • Enhanced Search Query Performance based on Splunk Search Queries
  • Enhanced and provided core design impacting the Splunk framework and components
  • Performance optimizations based on python scripts, shell scripts and CRON schedule
  • Involved in Resolving technical issues during development, deployment, and support
  • Performed testing activities related to performance testing, Unit Testing, Load Testing, Functional Testing, Automated testing for the python scripts developed
  • Requirements elicitation and translation to technical specifications
  • Actively involved in setting up platforms for the WAS, JBOSS, IAAS Servers, EPAAS environments
  • Actively involved in mounting file-systems, software installations, establishing connectivity for the WAS, JBOSS, IAAS Servers to the integration systems related to Databases (Oracle, Mainframe)
  • Actively involved in setting up alerts for the different servers: WAS, JBOSS, IBM servers using the different Alerting tools: Nimsoft, Tivoli, ELK tool
  • Actively involving in monitoring the server’s health using the Splunk Monitoring and Alerting tool, Tivoli Alerting tool
  • Actively involved in performing deployments on the different servers (WAS, JBOSS, IBM) for the different components involved in the eco-system
  • Provided solution road map for future growth, process architecture for logging system on the different servers (WAS, JBOSS, IAAS)
  • Provided inputs on solution architecture based on evaluation/understanding of solution alternatives, frameworks and products
  • Interact with clients to elicit architectural and non-functional requirements like performance, scalability, reliability, availability, maintainability.
  • Anchor proof of concept (POC) development to validate proposed solution and reduce technical risk.
  • Worked using Denodo components for Data Virtualization
  • Hands-on experience in creating Base Views, Derived Views, Joins, Unions, Projection, Selection, Union, Minus, Flatten Views, Interface and associations of data service layers in Denodo.
  • Developed performance utilization charts, optimized and tuned SQL and designed physical databases. Assisted developers with Teradata load utilities and SQL.
  • Converted batch jobs with BULKLOAD utility to TPUMP utility.
  • Created tables, views in Teradata, according to the requirements.
  • Perform performance optimizations on Java/JVM frameworks and UNIX Shell Scripts
  • Prepare estimations, release plan and road map for future releases
  • Design applications based on identified architecture and support implementation of design by resolving complex technical issues faced by the IT project team during infrastructure set-up, development, deployment and support.
  • Engaged multiple teams for sourcing the data files from the databases (Oracle, Mainframe) to the servers involved in the platform
  • Involved in configuring Load Balancer Configuration on the servers
  • Involved in setting up Kafka and Zookeeper Producer-Consumer components for the Big Data Environments
  • Involved in setting up Broker and topic-level Configurations in Kafka
  • Involved in Connection and Stream Configurations in Kafka
  • Involved in the following APIs related to Kafka: Producer API, Consumer API
  • Got Certified as Splunk Certified Power User
  • Performed work as Splunk Developer and Splunk Admin for enhancing the Splunk Search Query Performance
  • Worked as SME for Java, Big Data, and, Splunk technologies.
  • Experienced in developing Enterprise Application integrating across various platforms using Java 1.8
  • Used Java Collection Framework for developing Map-Reduce applications and APIs for NoSQL databases
  • Improved profitability by 30-40% by significantly making improvements in the Big-Data technologies with respect to Splunk.
  • Involved in the activities of dashboard creation, alerts, logging mechanisms, pivot creation, regex for field extractions
  • Involved in the activities of Python and shell-scripting for the Key-Value Pairs creation and masking the PII data fields
  • Increased reliability of data, decreasing errors happening by 50%
  • Improved reporting mechanisms for the Splunk tool to the clients
  • Improved logging activities of the data reporting from manual activity to automatic activity
  • Gained experience in the tool related to Splunk DB Connect and imported the data from the Oracle platform to the Splunk platform.
  • Experienced in Data Modelling in SQL and NoSQL Databases
  • Involved in the operations of Spark Streaming
  • Involved in the operations of Spark SQL
  • Involved in the operations of Scala Programming and Performance Tuning
  • Involved in the activities of RDD Creations and Operations of Data Frames and Datasets for the Use-Case
  • Hands on experience in NOSQL databases like HBase, Cassandra, MongoDB.
  • Experience in designing and developing POCs in Spark using Scala to compare the performance of Spark with Hive and SQL/Oracle.
  • Used Spark-Streaming APIs to perform necessary transformations and actions on the fly for building the common learner data model which gets the data from Kafka in near real time and Persists into Cassandra.
  • Developed Spark scripts by using Scala shell commands as per the requirement.
  • Used Spark API over Cloudera Hadoop YARN to perform analytics on data in Hive.
  • Loaded the data into Spark RDD and do in memory data Computation to generate the Output response.
  • Optimizing of existing algorithms in Hadoop using Spark Context, Spark-SQL, Data Frames and Pair RDD's.
  • Worked on migrating Map Reduce programs into Spark transformations using Spark and Scala

Environment: Apache Hadoop, Apache Hive, Apache Pig, Cloudera (CDH 5), MapR, Ubuntu, HDFS, MapReduce, Amazon Web Services(AWS), Python, Splunk, Supervisor, Monit, Hazelcast, HAProxy, Kafka, Zookeeper, Elastic-Search, Logstash, Kibana (ELK), Servers: JBOSS, WAS, IAAS, E-PAAS, Redhat Enterprise Linux, Talend, Microsoft Azure

Confidential, Bentonville, AR

Big Data Developer

Job Responsibilities:

  • Reviewed in-house and external technology for an internal event management system, analyzing content management system, denodos data virtualization abilities, integrate yammer for social media element of the system and internal text analytics tool managing journals.
  • Worked as a Dev-Ops Engineer.
  • Involved in the activities of Release Planning
  • Involved in the activities of deployments, developments, Change Request Creations, Environment Readiness
  • Performed activities on the development and production clusters
  • Performed monitoring activities in the UNIX environment
  • Implemented Shell Scripts for Environment Readiness in the UNIX environment
  • Documented Design Documents for Big Data Analytics & Reporting
  • Involved in the activities of daily standups and scrum planning
  • Performed Big Data operations in the Horton-Works distribution
  • Dealing with Data Scientists on the R programming queries
  • Well-versed in the concepts of networking with different protocols related to TCP/IP and MQ, SFTP
  • Created/Enhanced Teradata Stored Procedures to generate automated testing SQLs.
  • Experienced in developing Enterprise Application integrating across various platforms using Java 1.8/JVM
  • Performance Management transactional, social data using Couch DB
  • Data Exploration for the Social Analytics using Couch DB
  • Performed and Implemented Decision Science applications using Apache Hive, MongoDB, Couch DB
  • Used Java Collection Framework for developing Map-Reduce applications and APIs for NoSQL databases
  • Worked using Azure Data Lake Store for analyzing the data stored on YARN and HDFS including multiple access methods related to Spark, Hive, HBase
  • Analyzed the different kinds of structured and unstructured data including the processing of files within the data stored in the Data Lake.
  • Strong experience in designing and working with MySQL and Mongo DB
  • Designed and implemented the MongoDB schema
  • Wrote services to store and retrieve user data from the MongoDB for the application on devices
  • Worked on App Engine and Amazon AWS back-ends and in the front-ends as well
  • Migrated Hadoop metadata to Docker container
  • Involved in the activities of Amazon EMR, S3, setting up connectivity using VPC Connection
  • Performed map-reduce operations using Amazon EMR
  • Experienced in Data Modelling in SQL and NoSQL Databases
  • Hands on experience in NOSQL databases like HBase, Cassandra, MongoDB.
  • Worked using the tools of JIRA and Jenkins within the project.
  • Experience in Operational Intelligence using Splunk
  • Good knowledge about Splunk architecture and various components (indexer, forwarder, search heads, deployment server), Heavy and Universal forwarder, License model.

Environment: Apache Hadoop, Apache Hive, Ubuntu, HDFS, MapReduce, Shell Scripting, Python, HBase, Mongo DB, Couch DB, JIRA, Jenkins

Confidential, Beaverton, OR

Big Data Developer

Job Responsibilities:

  • Worked with Data Analytics team for meeting the testing requirements involved with the Hive & Pig scripts for different Use-Cases in Hadoop.
  • Documented Design Documents for Big Data Analytics & Reporting
  • Performed Unit Testing for the python scripts
  • Performed automation testing for the java scripts development involved.
  • Involved in the operations of Cloudera, Hortonworks, MapR environments
  • Performed End-to-End testing for the scripts execution in the big-data clusters
  • Written test results and verified actual test results with the expected results for SQL and HiveQL Queries
  • Analyzed large data sets by running Hive queries and Pig scripts.
  • Worked with the Data Science team to gather requirements for various data mining projects .
  • Involved in creating Hive tables, and loading and analyzing data using Hive queries.
  • Developed Simple to complex MapReduce Jobs using Hive and Pig.
  • Involved in running Hadoop jobs for processing millions of records of text data.
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
  • Developed multiple MapReduce jobs in Java for data cleaning and preprocessing
  • Involved in loading data from LINUX file system to HDFS and then to Amazon S3.
  • Responsible for managing data from multiple sources.
  • Testing of data developed using Hive, Pig, and Impala.
  • Tested the application of various compression formats.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Responsible to manage data coming from different sources of RDBMS systems.
  • Responsible for creating and managing HBase Data Store.
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Script.
  • Handled importing of data from various data sources, performed transformations using Hive, Map Reduce, loaded data into HDFS and Extracted the data from Oracle into HDFS using Sqoop and Oozie Workflows.
  • Imported and Exported the data to/from the relational databases using Sqoop for visualization and to generate reports for the Data Scientist/BI team.
  • Developed Simple to Complex Map/Reduce Jobs.
  • Analyzed the data by performing Hive queries and running Pig scripts to know user behavior.
  • Executed data scripts developed in Python.
  • Developing and testing Hive UDF’s for bringing all the customers information into a structured format.
  • Experienced in Data Modelling in SQL Database and HiveQL
  • Varied the loads of concurrent users using the Cassandra cluster at any given point of time
  • Distributed online analytic platform Kafka, Storm, Cassandra, Elastic search, scaling, fault tolerance
  • Hands on experience in NOSQL databases like HBase, Cassandra, MongoDB.

Environment: Apache Hadoop, Apache Hive, Apache Pig, Cloudera (CDH 5), Ubuntu, Auto-CAD, Sqoop, HDFS, MapReduce, NoSQL, HBase, CouchBase, Oozie, Amazon Web Services(AWS), Spark, Storm, Flume, Python, Shell Scripting

Confidential, MI

Big Data Research Assistant

Job Responsibilities:

  • Worked as a Research Assistant under the supervision of the professors.
  • Provided consulting services, solutions and training around Big Data ecosystem (Hadoop, NoSQL, Cloud).
  • Advised executives on Big Data strategy. Conducted internal Hadoop training for the team. Ran a popular webinar series on Hadoop.
  • Worked with the Big Data Storage Team, that was building the company’s Big Data application. Virtualized Hadoop in Linux environment for providing a safer, scalable analytics sandbox on the application. Developed a HDFS plugin to interface with proprietary file system. Advised file system engineers on Hadoop specific optimizations. Implemented component based tests/ benchmarks.
  • Mentored an intern working on recommendation engine using Hadoop/ Mahout.
  • Built a scalable, cost effective, and fault tolerant dataware-house system on Amazon EC2 Cloud. Developed MapReduce/EMR jobs to analyze the data and provide heuristics and reports. The heuristics were used for improving campaign targeting and efficiency.
  • Worked on multiple virtual machines such as Cloudera and Ubuntu.
  • One of the demonstrations of my work is shown on
  • Implemented Map-Reduce Programming on Classical and YARN MapReduce daemons
  • Worked on Big Data using Hive-Data Warehousing tool. And, developed a website to coordinate with the Big Data.
  • Developed 2D and 3D designs using Auto-CAD.

Environment: CloudEra (CDH 3,4,5), Apache Hadoop, Linux, HDFS, Hive, Pig, Sqoop, Flume, Zookeeper, HBase, Oozie, Flume, HortonWorks, MongoDB, Java, Map Reduce, Amazon EC2 infrastructure, Amazon Elastic Map Reduce (EMR), MySQL, shell scripts.

Confidential

Oracle Database Administrator

Responsibilities:

  • Created Oracle databases for departments in data warehouse and production environments.
  • Successfully developed and implemented database for company’s client’s details such as phone number, clients and client’s phone numbers using SQL Programming.
  • Implemented backup and recovery procedures for databases involved in IT departments.
  • Involved with remodeling and interface development for migrating from MS Access to MS SQL Server and from MS SQL Server to Oracle Servers.
  • Databases created using MS Access were designed for improving performance, multiple-user access, and increasing security and stability.
  • Involved in all phases of database development from needs assessment to Quality Assurance/ Quality Control, design and support.
  • Enhancing database performance using performance restores, implementing recovery procedures, handling performance tuning, and conducting regular system backups. Ensuring that the technical and functional designs were meeting the specified requirements of the business.
  • Managing projects from business requirements to solutions delivery and support.
  • Delivering solution to the problem on time, on specification and on budget.
  • Developed and tested storage strategies, and disaster-recovery plan for large manufacturing company’s database system, delivering solution that guaranteed recovery performance and high availability.
  • Enhanced end-users understanding of the databases by implementing workshops, delivering group and individual training sessions, and creating user-friendly training materials.
  • Developed reports using Crystal reports with T-SQL, MS Excel, and Access.
  • Involved with most of the stuff related to business intelligence, business objects in Oracle technology.
  • Responsible for the development of database, design, performance tuning, implementation and maintenance.
  • The databases created involved mostly importing and exporting: installing and configuring database software for Oracle Application development environment.
  • Handled database server tuning, application tuning, maintaining documentation, and database security.
  • Have developed Oracle Forms, Oracle tables, Oracle dashboards using Oracle 11g.
  • Did Course in SAP - BI, SAP - BO, SAP - BW.

Environment: Oracle 12c, 11g, 10g & Oracle 6i, SAP - BW/BI/BO, Oracle BO/BI, PL/SQL, MS Access, DB2

Confidential

Java Developer

Responsibilities:

  • Developed a website using Joomla CMS System.
  • Developed and maintained corporate internet websites.
  • Designed and created online marketing collateral, as well as online applications, and blogs.
  • The website was created by taking into consideration user accounts involved in the company, day-to-day transactions involved in the company, publishing and generating product specifications.
  • Static Content migrated to Content Management System.
  • Some of the technologies majorly involved with include PHP, JavaScript, HTML, CSS and MySQL for developing online applications.
  • Steps for tracking event-activities involved with the company were done using the help of company’s CEO. System encompassed all preparatory steps, generated timely reports to responsible persons, maintaining status of the various activities involved with the event, and providing reminder and overview emails.
  • Improved the website appearance with involvement of departmental representatives.
  • Developed secure web application and back-end database to automate confidential survey process for legal department of Major Corporation.
  • Improved Company’s product design’s using Adobe Illustrator, Adobe Photoshop.

Environment: Windows, Joomla CMS, Drupal CMS, PHP, JSP, Java

We'd love your feedback!