We provide IT Staff Augmentation Services!

Sr. Big Data Developer Resume

4.00/5 (Submit Your Rating)

Charlotte, NC

SUMMARY

  • Over 8 years of experience in Information Technology which includes experience in Big data,HADOOP Ecosystem,Core Java/J2EE and strong in Design, Software processes, Requirement gathering, Analysis and development of software applications
  • Excellent Hands on Experience in developing Hadoop Architecture in Windows and Linux platforms.
  • Experience in building bigdata solutions using Lambda Architecture using Cloudera distribution of Hadoop, MapReduce,Cascading,HIVE and Sqoop.
  • Strong development experience in Java, Maven, Jenkins, Servlets, Spring Boot, JUnit, Log4j.
  • Experienced in J2EE Design Pattern.
  • Experience implementing Data Pipelines using Lambda Architecture and Kappa Architecture.
  • Excellent working experience onBig Data Integrationand Analytics based onHadoop,Spark,Kafka frameworks.
  • Hands on experience working on NoSQL databases including HBase,MongoDB.
  • Hands on experience in writingAd - hoc Queriesfor moving data from HDFS to HIVEand analyzing the data using HIVEQL.
  • Good understanding on Cloud Based technologies like AWS.
  • Good knowledge on Bit Bucket and GitHub Enterprise.
  • Knowledge on Docker to create Containers using Dockerfile and orchestrate using Docker Compose and Kubernetes.
  • Expertise inAmazon Web ServicesincludingElastic Cloud Compute (EC2) and Dynamo DB.
  • Good understanding and experience with Software Development methodologies likeAgileandWaterfall.
  • Experienced in importing and exporting data usingSqoopfromHDFS (Hive &HBase)to Relational Database Systems(Oracle, MySQL)and vice-versa.
  • Experienced in developing and designingWeb Services (Restful Web services).
  • Expertise in variousJava/J2EEtechnologies like Servlets,Spring Boot.

PROFESSIONAL EXPERIENCE

Confidential, Charlotte, NC

Sr. Big Data Developer

Responsibilities:

  • Developed Batch Processing data pipelines using Spark.
  • Developed Stream Processing data pipelines using Spark Streaming and Kafka.
  • Used Spark to perform in-memory processingon data in Hive.
  • Implemented Data Ingestion, Data Processing frameworks for building Scalable Data Platform solutions.
  • Developed a plumbing services to gather metrics for monitoring and bench marking the Data Pipeline.
  • Explored with theSparkimproving the performance and optimization of the existing algorithms inHadoopusingSparkContext,Spark -SQL, Data Frame,Pair RDD's,Spark YARN.
  • Handled importing of data from various data sources, performed transformations usingHive, MapReduce, loaded data intoHDFSand Extracted the data fromSQL into HDFS using Sqoop.
  • Deployed application toAWSand monitored the load balancing of differentEC2 instances
  • Created pipeline scripts for build and deployment automation for Java based project by using Jenkins.
  • Deployed Spring Boot based micro services Docker container using Amazon EC2 container services.
  • Developed a POC for project migration from on premHadoopMapR system to Snowflake
  • DevelopedKafkaproducer and consumers using the Kafka Java API.
  • Successfully translated the SQL scripts to Spark jobs.
  • ImplementedAWS EC2, Key Pairs, Security Groups, Auto Scaling, ELB, SQS, and SNS using AWS API and exposed as the Restful Web services.
  • Involved in convertingMapReduceprograms intoSpark transformationsusingSpark RDD's on Scala.
  • Implemented usingSCALA and SQLfor faster testing and processing of data. Real time streaming the data using withKafka.
  • Developed and designed automation framework using Shell scripting.
  • DevelopedHive Scripts, UNIX Shell scripts,programming for allETLloading processes and converting the files into parquet in theHadoop File System.
  • UtilizedAgile Scrum Methodologyto help manage and organize a team of 4 developers with regular code review sessions
  • Worked on Version control systems like Subversion, GIT, CVS.

Confidential, Cherry Hill, NJ

Java/ Big Data Developer

Responsibilities:

  • Proactively monitored systems and services, architecture design and implementation ofHadoopdeployment, configuration management, backup, and disaster recovery systems and procedures
  • ConfiguredSparkstreaming to receive real time data from theKafkaand store the stream data toHDFSusingScale.
  • Installed and configuredHadoop, MapReduce, HDFS (Hadoop Distributed File System),developed multipleMapReducejobs injavafor data cleaning and processing.
  • Worked on implementingSparkusingScalaandSpark SQLfor faster analyzing and processing of data.
  • UsedJAVA, J2EEapplication development skills with Object Oriented Analysis and extensively involved throughoutSoftware Development Life Cycle (SDLC)
  • Containerized applications using Docker and orchestrated using Kubernetes.
  • Handled in Importing and exporting data intoHDFSandHiveusingSQOOPandKafka
  • Involved in creatingHive tables, loading the data and writinghive queries, which will run internally in map reduce.
  • AppliedMapReduceframework jobs in java for data processing by installing and configuringHadoop, HDFS.
  • Worked on reading and writing multiple data formats like JSON, ORC, Parquet on HDFS using Spark.
  • WrittenHivejobs to parse the logs and structure them in tabular format to facilitate effective querying on the log data.
  • Created UDFs in Hive to implement custom funtions.
  • Involved in developingShell scriptsto easy execution of all other scripts(Hive, and MapReduce)and move the data files within and outside of HDFS.
  • Involved in convertingHive/SQLqueries into Spark transformations using SparkRDDs andScala.
  • Worked withNoSQLdatabases likeHBasein creating tables to load large sets of semi structured data.
  • GeneratedJava APIsfor retrieval and analysis onNo-SQLdatabase such as HBase.
  • CreatedETLjobs to generate and distribute reports fromMySQLdatabase using Pentaho Data Integration.
  • Worked on loading data fromUNIXfile system toHDFS
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Continuous integration using Jenkins for nightly builds and send automatic emails to the team.
  • Used Jenkins plugins for code coverage and also to run all the test before generating war file.

Confidential, Bloomington, IL

Hadoop Developer

Responsibilities:

  • Involved in writing Client side Scripts usingJava Scriptsand Server Side scripts usingJava Beansand usedservletsfor handling the business.
  • DevelopedScalaprograms with Spark for data inHadoop ecosystem.
  • Used Service RESTWeb Services(JAX-RS)for integration with other systems.
  • ImplementedAWS EC2, Key Pairs, Security Groups, Auto Scaling, ELB, SQS, and SNS using AWS APIand exposed as the Restful Web services.
  • Data modeledHBase tablesto load large sets of structured, semi-structured and unstructured data coming fromUNIX,NoSQLand a variety of data sources.
  • Solved performance issues in Hive scripts with understanding of Joins, Group and aggregation and translate to MapReduce jobs.
  • DevelopedUDFsinJavaas and when necessary to use in HIVEqueries.
  • Coordinated with various stakeholders such as the End Client, DBA Teams, Testing Team and Business Analysts.
  • Involved in gathering requirements and developing a project plan.
  • Involved in understanding requirements, functional specifications, designing documentations and testing strategies.
  • Involved in UI designing, Coding, Database Handling.
  • Involved inUnit TestingandBug Fixing.
  • Worked over the entireSoftware Development Life Cycle (SDLC)as a part of a team as well as independently.
  • Written SQL queriesto query the database and providing data extracts to users as per request

Confidential

Java Developer

Responsibilities:

  • Develop Web tier usingSpring MVCFramework.
  • Perform database operations on the consumer portal usingSpringJdbc template.
  • Implementeddesign patternsin Scala for the application.
  • Setting upinfrastructureImplementing Configuring ExternalizingHTTPDmod jkmod rewrite.mod proxy JNDI SSL etc.
  • Involved in converting Hive/SQL queries into Spark transformations using Spark RDD, Scala.
  • ImplementedRestfulservices in Spring.
  • Serialize and de-serialize objects usingPlayJson library.
  • Developing traits and case classes etc in scala.
  • Develop quality code adhering to Scala coding Standards and best practices.
  • Writing complex Sql queries.
  • Develop GUI usingJQueryJsonandJava script.
  • Unit testing Integration testing and bug fixing.
  • Followed Agile methodology (Stand up meetings, retrospective meetings, sprint development and Pair programming).
  • Developed application code using Eclipse IDE and configured with Maven, Glassfish server and JUnit.
  • Developed Use Case Diagrams, Sequence Diagrams and Class Diagrams using Rational Rose.
  • Developed the controller servlet to handle the requests and responses.
  • Developed JSP pages with MVC architecture using Spring MVC, Servlets and Simple tags.
  • Configured Maven dependencies for application building processes.
  • Used Spring Dependency Injection to set up dependencies between the objects.
  • Optimized the source code and queries to improve performance using Hibernate.
  • Assisted other team members with various technical issues includingJavaScript, CSS, JSPand Server related issues.

We'd love your feedback!