Sr. Big Data Developer Resume
4.00/5 (Submit Your Rating)
Charlotte, NC
SUMMARY
- Over 8 years of experience in Information Technology which includes experience in Big data,HADOOP Ecosystem,Core Java/J2EE and strong in Design, Software processes, Requirement gathering, Analysis and development of software applications
- Excellent Hands on Experience in developing Hadoop Architecture in Windows and Linux platforms.
- Experience in building bigdata solutions using Lambda Architecture using Cloudera distribution of Hadoop, MapReduce,Cascading,HIVE and Sqoop.
- Strong development experience in Java, Maven, Jenkins, Servlets, Spring Boot, JUnit, Log4j.
- Experienced in J2EE Design Pattern.
- Experience implementing Data Pipelines using Lambda Architecture and Kappa Architecture.
- Excellent working experience onBig Data Integrationand Analytics based onHadoop,Spark,Kafka frameworks.
- Hands on experience working on NoSQL databases including HBase,MongoDB.
- Hands on experience in writingAd - hoc Queriesfor moving data from HDFS to HIVEand analyzing the data using HIVEQL.
- Good understanding on Cloud Based technologies like AWS.
- Good knowledge on Bit Bucket and GitHub Enterprise.
- Knowledge on Docker to create Containers using Dockerfile and orchestrate using Docker Compose and Kubernetes.
- Expertise inAmazon Web ServicesincludingElastic Cloud Compute (EC2) and Dynamo DB.
- Good understanding and experience with Software Development methodologies likeAgileandWaterfall.
- Experienced in importing and exporting data usingSqoopfromHDFS (Hive &HBase)to Relational Database Systems(Oracle, MySQL)and vice-versa.
- Experienced in developing and designingWeb Services (Restful Web services).
- Expertise in variousJava/J2EEtechnologies like Servlets,Spring Boot.
PROFESSIONAL EXPERIENCE
Confidential, Charlotte, NC
Sr. Big Data Developer
Responsibilities:
- Developed Batch Processing data pipelines using Spark.
- Developed Stream Processing data pipelines using Spark Streaming and Kafka.
- Used Spark to perform in-memory processingon data in Hive.
- Implemented Data Ingestion, Data Processing frameworks for building Scalable Data Platform solutions.
- Developed a plumbing services to gather metrics for monitoring and bench marking the Data Pipeline.
- Explored with theSparkimproving the performance and optimization of the existing algorithms inHadoopusingSparkContext,Spark -SQL, Data Frame,Pair RDD's,Spark YARN.
- Handled importing of data from various data sources, performed transformations usingHive, MapReduce, loaded data intoHDFSand Extracted the data fromSQL into HDFS using Sqoop.
- Deployed application toAWSand monitored the load balancing of differentEC2 instances
- Created pipeline scripts for build and deployment automation for Java based project by using Jenkins.
- Deployed Spring Boot based micro services Docker container using Amazon EC2 container services.
- Developed a POC for project migration from on premHadoopMapR system to Snowflake
- DevelopedKafkaproducer and consumers using the Kafka Java API.
- Successfully translated the SQL scripts to Spark jobs.
- ImplementedAWS EC2, Key Pairs, Security Groups, Auto Scaling, ELB, SQS, and SNS using AWS API and exposed as the Restful Web services.
- Involved in convertingMapReduceprograms intoSpark transformationsusingSpark RDD's on Scala.
- Implemented usingSCALA and SQLfor faster testing and processing of data. Real time streaming the data using withKafka.
- Developed and designed automation framework using Shell scripting.
- DevelopedHive Scripts, UNIX Shell scripts,programming for allETLloading processes and converting the files into parquet in theHadoop File System.
- UtilizedAgile Scrum Methodologyto help manage and organize a team of 4 developers with regular code review sessions
- Worked on Version control systems like Subversion, GIT, CVS.
Confidential, Cherry Hill, NJ
Java/ Big Data Developer
Responsibilities:
- Proactively monitored systems and services, architecture design and implementation ofHadoopdeployment, configuration management, backup, and disaster recovery systems and procedures
- ConfiguredSparkstreaming to receive real time data from theKafkaand store the stream data toHDFSusingScale.
- Installed and configuredHadoop, MapReduce, HDFS (Hadoop Distributed File System),developed multipleMapReducejobs injavafor data cleaning and processing.
- Worked on implementingSparkusingScalaandSpark SQLfor faster analyzing and processing of data.
- UsedJAVA, J2EEapplication development skills with Object Oriented Analysis and extensively involved throughoutSoftware Development Life Cycle (SDLC)
- Containerized applications using Docker and orchestrated using Kubernetes.
- Handled in Importing and exporting data intoHDFSandHiveusingSQOOPandKafka
- Involved in creatingHive tables, loading the data and writinghive queries, which will run internally in map reduce.
- AppliedMapReduceframework jobs in java for data processing by installing and configuringHadoop, HDFS.
- Worked on reading and writing multiple data formats like JSON, ORC, Parquet on HDFS using Spark.
- WrittenHivejobs to parse the logs and structure them in tabular format to facilitate effective querying on the log data.
- Created UDFs in Hive to implement custom funtions.
- Involved in developingShell scriptsto easy execution of all other scripts(Hive, and MapReduce)and move the data files within and outside of HDFS.
- Involved in convertingHive/SQLqueries into Spark transformations using SparkRDDs andScala.
- Worked withNoSQLdatabases likeHBasein creating tables to load large sets of semi structured data.
- GeneratedJava APIsfor retrieval and analysis onNo-SQLdatabase such as HBase.
- CreatedETLjobs to generate and distribute reports fromMySQLdatabase using Pentaho Data Integration.
- Worked on loading data fromUNIXfile system toHDFS
- Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
- Continuous integration using Jenkins for nightly builds and send automatic emails to the team.
- Used Jenkins plugins for code coverage and also to run all the test before generating war file.
Confidential, Bloomington, IL
Hadoop Developer
Responsibilities:
- Involved in writing Client side Scripts usingJava Scriptsand Server Side scripts usingJava Beansand usedservletsfor handling the business.
- DevelopedScalaprograms with Spark for data inHadoop ecosystem.
- Used Service RESTWeb Services(JAX-RS)for integration with other systems.
- ImplementedAWS EC2, Key Pairs, Security Groups, Auto Scaling, ELB, SQS, and SNS using AWS APIand exposed as the Restful Web services.
- Data modeledHBase tablesto load large sets of structured, semi-structured and unstructured data coming fromUNIX,NoSQLand a variety of data sources.
- Solved performance issues in Hive scripts with understanding of Joins, Group and aggregation and translate to MapReduce jobs.
- DevelopedUDFsinJavaas and when necessary to use in HIVEqueries.
- Coordinated with various stakeholders such as the End Client, DBA Teams, Testing Team and Business Analysts.
- Involved in gathering requirements and developing a project plan.
- Involved in understanding requirements, functional specifications, designing documentations and testing strategies.
- Involved in UI designing, Coding, Database Handling.
- Involved inUnit TestingandBug Fixing.
- Worked over the entireSoftware Development Life Cycle (SDLC)as a part of a team as well as independently.
- Written SQL queriesto query the database and providing data extracts to users as per request
Confidential
Java Developer
Responsibilities:
- Develop Web tier usingSpring MVCFramework.
- Perform database operations on the consumer portal usingSpringJdbc template.
- Implementeddesign patternsin Scala for the application.
- Setting upinfrastructureImplementing Configuring ExternalizingHTTPDmod jkmod rewrite.mod proxy JNDI SSL etc.
- Involved in converting Hive/SQL queries into Spark transformations using Spark RDD, Scala.
- ImplementedRestfulservices in Spring.
- Serialize and de-serialize objects usingPlayJson library.
- Developing traits and case classes etc in scala.
- Develop quality code adhering to Scala coding Standards and best practices.
- Writing complex Sql queries.
- Develop GUI usingJQueryJsonandJava script.
- Unit testing Integration testing and bug fixing.
- Followed Agile methodology (Stand up meetings, retrospective meetings, sprint development and Pair programming).
- Developed application code using Eclipse IDE and configured with Maven, Glassfish server and JUnit.
- Developed Use Case Diagrams, Sequence Diagrams and Class Diagrams using Rational Rose.
- Developed the controller servlet to handle the requests and responses.
- Developed JSP pages with MVC architecture using Spring MVC, Servlets and Simple tags.
- Configured Maven dependencies for application building processes.
- Used Spring Dependency Injection to set up dependencies between the objects.
- Optimized the source code and queries to improve performance using Hibernate.
- Assisted other team members with various technical issues includingJavaScript, CSS, JSPand Server related issues.