We provide IT Staff Augmentation Services!

Spark Developer Resume

4.00/5 (Submit Your Rating)

Cherry Hill, NJ

SUMMARY:

  • 6 years senior developer with 3 years as a Hadoop /Spark Developer in Big Data domain and 3 years as a Java developer.
  • Experience d in analyzing the different types of data flow from source to Hadoop Clusters.
  • Good at with Big Data Hadoop core and Eco - System components ( HDFS , MapRedue , Hive , Yarn, Sqoop , Oozie , Hbase and Pig ).
  • Hands on Hortonworks Distribution System of Hadoop .
  • Experience d in manipulating the streaming data to clusters through Kafka and Spark-Streaming .
  • Extensive e xperience in partitioning Data according to the business requirements using Hive .
  • Working with data importing and exporting in Sqoop .
  • Good knowledge in data extraction, transformation and loading between Hive , HDFS and RDBMS.
  • Good work with data transformation from HDFS , HIVE , HBase , and RDBMS .
  • Experience d in creating UDF 's, UDAF 's for Hive .
  • Acumen in choosing an efficient ecosystem in Hadoop and providing the best solutions to Big Data problems.
  • Extensive Knowledge in Spark, Spark - Streaming, SparkSQL and Scala .
  • Proficient in working with NoSQL database like Cassandra and HBase .
  • Prolific in generating the splendid and informative dashboards for Business Intelligence teams.
  • Experience d in using Front-end technologies like JavaScript, HTML, CSS, HTML5 and CSS3.
  • Experienced in Test such as JUnit
  • Proficient in relational databases like Oracle , SQL , My SQL and SQL Server .
  • Used to work in Agile environment .
  • Comfortable in Linux, UNIX environments. And Expertise in Unix-Shell .
  • Extensive experience with Front-end and Back-end IDE 's like Eclipse , Visual Studio.
  • Proficient communication skills with an ability to lead a team & keep them motivated.
  • Highly motivated and versatile team player with the ability to work independently & adapt quickly to new emerging technologies.

TECHNICAL SKILLS:

Hadoop Ecosystem Framework: MapReduce, Hive, Pig, Hadoop, Spark HBase, Sqoop, Spark, HDFS, KafkaDatabases Methodologies: Oracle, MySQL, SQL Server, Agile Scrum Hbase, Cassandra, MongoDBLanguages Web Technologies: Java, Scala, C++, SQL, HiveQL, HTML 5, JSON, Pig Latin, JavaScript, Shell-Scripting, XML, CSS 3, Angular.JSSystems Other: Windows, Linux, UNIX Eclipse, Maven, JUnit, Git

PROFESSIONAL EXPERIENCE:

Confidential, Cherry Hill, NJ

Spark Developer

Responsibilities:

  • Designed real-time data importing method as producer to pull the data from the sources to Kafka clusters.
  • Realized collecting incoming data in real-time and processing them within few seconds at the same time using Spark-Streaming and SparkSQL.
  • Aggregated and stored the data result into HDFS and Cassandra .
  • Accomplished different goals of the management team by implement algorithms and business logics of data process in Spark related technologies.
  • Involved in retrieving the consumption records from Cassandra to spark.
  • Processed the data which extracted from Cassandra utilizes scala programming in Spark Framework.
  • Implemented machine learning model which is based on RDBMS records with specific API to help detect the fraud actions.
  • Outputted a file which included all the fraud actions and then save it to Cassandra.
  • Strong problem solving experience on t est , i dentify, address, debug and then resolve technical issue s that affect the integrity of the application .
  • Expertized in improving the performance of application by analytical skills.
  • Worked in the agile Environments, e ffectively communicated at all levels of an organization in Management and Technical roles.

Environment: Hadoop, Spark, Apache Kafka, RDBMS, Hive, HDFS, Cassandra, Scala, Parquet, Linux.

Confidential, NY C, NY

Hadoop /Spark Developer

Responsibilities:

  • Developed a data pipeline using Kafka , Spark , RDBMS , HDFS , Scala , MapReduce , Hive and Sqoop to ingest, transform and analyzing customer behavioral data.
  • Imported the data from RDBMS into HDFS using Sqoop .
  • Designed dynamically storage logic of data and improved performance with HDFS and Hive.
  • Designed and Developed MapReduce in Java to process large volume of data.
  • Used Kafka to collect streaming data from customers' consumption records. Configured Kafka to read and write messages from external data sources.
  • Realized process of real time streaming data using Spark Streaming and Kafka , storage data into HDFS .
  • Exported the analyzed data from HDFS to Hive using Sqoop to further visualize and generate reports for the BI team.
  • Utilized Spark Core , Spark Streaming and Spark SQL API for faster processing of data instead of using MapReduce in Java .
  • Save the final result back into HDFS and then imported to Hive .

Environment: Kafka, Spark, Scala, Hadoop, RDBMS, Hive, Sqoop, Java, HDFS, Linux, HBase

Confidential, NY

BigData Developer

Responsibilities:

  • Implemented Pig scripts to clean and accomplish data preprocessing.
  • Involved in importing and exporting data using Sqoop from HDFS to RDBMS and vice-versa.
  • In order to improve performance, utilize Hive scripts to reorganize data.
  • M igrat ed data from the sources such as HDFS and Hive in to Map Reduce .
  • Developed Map Reduce programs in Java to process data.
  • Wrote various Queries on data according to business requirement.
  • Effectively communicated at all levels of an organization ’s Management and Techni cians .
  • Working closely with BI team.

Environment: Oracle, Linux, Hadoo p, Hiv e, P ig, Sqoop, HBase, Java, JSON, Hortonworks

Confidential

Web Developer

Responsibilities:

  • Developed database schema and SQL queries for querying, inserting and managing database
  • Developed pages using HTML5, CSS 3 and JavaScript as part of the presentation layer
  • Hibernate framework is used in persistence layer for mapping an object-oriented domain model to database.
  • Used Maven scripts to fetch, build, and deploy application to development environment
  • Used JUnit for functional and unit testing code.

Environment: JavaScript, MySQL, JUnit, CSS 3, HTML 5, RDBMS, Hibernate

Confidential

Java Developer

Responsibilities:

  • Used Apache Tomcat as the web server of the house reservation records.
  • Implemented JDBC in Java to connect Apache Tomcat and Oracle.
  • Implemented JDBC to connect User interface and Apache Tomcat.
  • Created and developed database with Oracle, participated in manipulation of data transaction and improve the performance of data
  • Developed of application using Spring . MVC to handle the data communication from User interface to Server then to Database and vice-versa.
  • Involved in exchange of data between the Client, Server and Database in Hibernate.

Environment: Eclipse, JDBC, Spring.MVC, JavaScript, JUnit, Java 6, Servlets, CSS3, HTML 5, Oracle.

We'd love your feedback!