Spark Developer Resume
Cherry Hill, NJ
SUMMARY:
- 6 years senior developer with 3 years as a Hadoop /Spark Developer in Big Data domain and 3 years as a Java developer.
- Experience d in analyzing the different types of data flow from source to Hadoop Clusters.
- Good at with Big Data Hadoop core and Eco - System components ( HDFS , MapRedue , Hive , Yarn, Sqoop , Oozie , Hbase and Pig ).
- Hands on Hortonworks Distribution System of Hadoop .
- Experience d in manipulating the streaming data to clusters through Kafka and Spark-Streaming .
- Extensive e xperience in partitioning Data according to the business requirements using Hive .
- Working with data importing and exporting in Sqoop .
- Good knowledge in data extraction, transformation and loading between Hive , HDFS and RDBMS.
- Good work with data transformation from HDFS , HIVE , HBase , and RDBMS .
- Experience d in creating UDF 's, UDAF 's for Hive .
- Acumen in choosing an efficient ecosystem in Hadoop and providing the best solutions to Big Data problems.
- Extensive Knowledge in Spark, Spark - Streaming, SparkSQL and Scala .
- Proficient in working with NoSQL database like Cassandra and HBase .
- Prolific in generating the splendid and informative dashboards for Business Intelligence teams.
- Experience d in using Front-end technologies like JavaScript, HTML, CSS, HTML5 and CSS3.
- Experienced in Test such as JUnit
- Proficient in relational databases like Oracle , SQL , My SQL and SQL Server .
- Used to work in Agile environment .
- Comfortable in Linux, UNIX environments. And Expertise in Unix-Shell .
- Extensive experience with Front-end and Back-end IDE 's like Eclipse , Visual Studio.
- Proficient communication skills with an ability to lead a team & keep them motivated.
- Highly motivated and versatile team player with the ability to work independently & adapt quickly to new emerging technologies.
TECHNICAL SKILLS:
Hadoop Ecosystem Framework: MapReduce, Hive, Pig, Hadoop, Spark HBase, Sqoop, Spark, HDFS, KafkaDatabases Methodologies: Oracle, MySQL, SQL Server, Agile Scrum Hbase, Cassandra, MongoDBLanguages Web Technologies: Java, Scala, C++, SQL, HiveQL, HTML 5, JSON, Pig Latin, JavaScript, Shell-Scripting, XML, CSS 3, Angular.JSSystems Other: Windows, Linux, UNIX Eclipse, Maven, JUnit, Git
PROFESSIONAL EXPERIENCE:
Confidential, Cherry Hill, NJ
Spark Developer
Responsibilities:
- Designed real-time data importing method as producer to pull the data from the sources to Kafka clusters.
- Realized collecting incoming data in real-time and processing them within few seconds at the same time using Spark-Streaming and SparkSQL.
- Aggregated and stored the data result into HDFS and Cassandra .
- Accomplished different goals of the management team by implement algorithms and business logics of data process in Spark related technologies.
- Involved in retrieving the consumption records from Cassandra to spark.
- Processed the data which extracted from Cassandra utilizes scala programming in Spark Framework.
- Implemented machine learning model which is based on RDBMS records with specific API to help detect the fraud actions.
- Outputted a file which included all the fraud actions and then save it to Cassandra.
- Strong problem solving experience on t est , i dentify, address, debug and then resolve technical issue s that affect the integrity of the application .
- Expertized in improving the performance of application by analytical skills.
- Worked in the agile Environments, e ffectively communicated at all levels of an organization in Management and Technical roles.
Environment: Hadoop, Spark, Apache Kafka, RDBMS, Hive, HDFS, Cassandra, Scala, Parquet, Linux.
Confidential, NY C, NY
Hadoop /Spark Developer
Responsibilities:
- Developed a data pipeline using Kafka , Spark , RDBMS , HDFS , Scala , MapReduce , Hive and Sqoop to ingest, transform and analyzing customer behavioral data.
- Imported the data from RDBMS into HDFS using Sqoop .
- Designed dynamically storage logic of data and improved performance with HDFS and Hive.
- Designed and Developed MapReduce in Java to process large volume of data.
- Used Kafka to collect streaming data from customers' consumption records. Configured Kafka to read and write messages from external data sources.
- Realized process of real time streaming data using Spark Streaming and Kafka , storage data into HDFS .
- Exported the analyzed data from HDFS to Hive using Sqoop to further visualize and generate reports for the BI team.
- Utilized Spark Core , Spark Streaming and Spark SQL API for faster processing of data instead of using MapReduce in Java .
- Save the final result back into HDFS and then imported to Hive .
Environment: Kafka, Spark, Scala, Hadoop, RDBMS, Hive, Sqoop, Java, HDFS, Linux, HBase
Confidential, NY
BigData Developer
Responsibilities:
- Implemented Pig scripts to clean and accomplish data preprocessing.
- Involved in importing and exporting data using Sqoop from HDFS to RDBMS and vice-versa.
- In order to improve performance, utilize Hive scripts to reorganize data.
- M igrat ed data from the sources such as HDFS and Hive in to Map Reduce .
- Developed Map Reduce programs in Java to process data.
- Wrote various Queries on data according to business requirement.
- Effectively communicated at all levels of an organization ’s Management and Techni cians .
- Working closely with BI team.
Environment: Oracle, Linux, Hadoo p, Hiv e, P ig, Sqoop, HBase, Java, JSON, Hortonworks
Confidential
Web Developer
Responsibilities:
- Developed database schema and SQL queries for querying, inserting and managing database
- Developed pages using HTML5, CSS 3 and JavaScript as part of the presentation layer
- Hibernate framework is used in persistence layer for mapping an object-oriented domain model to database.
- Used Maven scripts to fetch, build, and deploy application to development environment
- Used JUnit for functional and unit testing code.
Environment: JavaScript, MySQL, JUnit, CSS 3, HTML 5, RDBMS, Hibernate
Confidential
Java Developer
Responsibilities:
- Used Apache Tomcat as the web server of the house reservation records.
- Implemented JDBC in Java to connect Apache Tomcat and Oracle.
- Implemented JDBC to connect User interface and Apache Tomcat.
- Created and developed database with Oracle, participated in manipulation of data transaction and improve the performance of data
- Developed of application using Spring . MVC to handle the data communication from User interface to Server then to Database and vice-versa.
- Involved in exchange of data between the Client, Server and Database in Hibernate.
Environment: Eclipse, JDBC, Spring.MVC, JavaScript, JUnit, Java 6, Servlets, CSS3, HTML 5, Oracle.