We provide IT Staff Augmentation Services!

Cassandra Developer/dba Resume

3.00/5 (Submit Your Rating)

Philadelphia, PA

SUMMARY:

  • More than Seven Years of total IT experience which includes experience in various Big Data technologies, Data Analytics and Java technologies.
  • Datastax certified Apache Cassandra Developer .
  • Experience in building Cassandra cluster .
  • Excellent CQL and data modeling knowledge from conceptual to physical data models including modern Chebotko diagrams.
  • Excellent Knowledge in the Architecture of Cassandra .
  • Having Knowledge on Datacenter, Datacenter racks, Cassandra Ring and nodes.
  • Proficient in Compaction strategies in Cassandra system.
  • Having knowledge on Spark and Scala Programming .
  • Tested the application and the cluster with different consistency levels to check for the writes and reads performance with respective to Consistency Level.
  • Hands on experience on Cassandra and oracle database architecture with better understanding and write processes including Mem - table, SSTable and Commit Log.
  • Excellent knowledge on CQL (Cassandra Query Language), for retrieving the data present in Cassandra cluster by running queries in CQL.
  • Involved in designing various stages of migrating data from RDBMS to Cassandra.
  • Gained Hands on experience in analyzing the Cassandra data from flat files using Spark .
  • Basic knowledge on Apache Solr .
  • Excellent Understanding of Distributed Systems and Parallel Processing architecture.
  • Creates Linux Virtual Machines using VMware Virtual Center.
  • Having Hands on experience in installing, configuring and maintaining Linux.
  • Can handle commissioning and decommissioning nodes along with monitoring of Cassandra Cluster.
  • Experience on Cassandra node tool to manage Cassandra cluster.
  • Proficient in managing and monitoring the cluster using Datastax OpsCenter.
  • Experienced in Hadoop architecture, Hadoop Distributed File System .
  • Excellent understanding on technologies such as Hive, PIG, Sqoop, Oozie, Hbase.
  • Hands on experience with Importing/Exporting data using Hadoop Data Management tool Sqoop.
  • Experienced using Hive and PIG for data analytics .
  • Expertise in Application Development using Java/J2EE.
  • Proficient in Web services, WCF (SOA), REST, and WEB-API using Java.
  • RDBMS experience includes SQL-Server, Access database connectivity using ADO.NET/ADO, ODBC and programming using SQL, T-SQL, Stored Procedures etc.
  • Experience in developing complex queries, Stored Procedures, Functions, Views, and Triggers using SQL Server.
  • Background in a disciplined software development life cycle (SDLC) process and excellent analytical, programming and problem solving skills.
  • Ability to take and give directions, instructions, provide assistance through completion of tasks and to work

TECHNICAL SKILLS:

Languages: C, C#, Java, J2EE (Servlets, JSP, JDBC, Java Beans), Python, SQL.

Web services: WCF, RESTful, Web Services

NoSQL Databases: Cassandra, Hbase.

Relational Databases: MySQL, Oracle10g.

Hadoop Eco System: HDFS, Map Reduce, YARN, Pig, Hive, Sqoop, Oozie, Zookeeper.

Cassandra: Cassandra, DataStax OpsCenter and DevCenter, nodetool, Spark on Cassandra and OpenStack.

Enterprise: J2EE, Spring Framework and Hibernate.

Databases: SQL Server 12/10/08, Oracle 9I/10g/11g/12c.

Operating Systems: Windows 8/7/XP/Vista, Linux.

Web technologies: XML, XAML, HTML, CSS, XMLHTTP, Java script.

Software: Eclipse, Visual Studio, NetBeans, Pycharm.

Tools: GIT, Maven, OpsCenter, DevCenter.

PROFESSIONAL EXPERIENCE:

Confidential,Philadelphia, PA

Cassandra Developer /DBA

Responsibilities:
  • Excellent Knowledge and understanding on Cassandra Architecture.
  • Involved in the process of Conceptual and Physical Data Modeling techniques.
  • Created data models in CQL for customer data.
  • Involved in Hardware installation and capacity planning for cluster setup.
  • Involved in the hardware decisions like CPU, RAM and disk types and quantities.
  • Used the Spark - Cassandra Connector to load data to and from Cassandra.
  • Prepared chebotko diagrams for logical and physical models during the phase of data modeling.
  • Ran many performance tests using the Cassandra-stress tool in order to measure and improve the read and write performance of the cluster.
  • Modified Cassandra.YAML files to set the configuration properties like cluster name, node addresses, seed provider, replication factors, memTable size and flush times etc.
  • Used the DataStax OpsCenter for maintenance operations and keyspace and table management.
  • Loaded and transformed large sets of structured, semi structured and unstructured data in various formats like text, zip, XML, CSV, YAML and JSON.
  • Created data-models for customer data using the Cassandra Query Language.
  • Involved in moving the SSTables data on to the live cluster.
  • Created User defined types to store specialized data structures in Cassandra.
  • Implemented advanced procedures like text analytics and processing using the in-memory computing capabilities like Spark.
  • Used the Java API to load the data into Cassandra Clusters.
  • Participated in NoSQL database integration and implementation.
  • Enhanced and optimized product Spark code to aggregate, group and run data mining tasks using the Spark framework.
  • Implemented a distributed messaging queue to integrate with Cassandra using Apache Kafka
  • Tuned and recorded performance of Cassandra clusters by altering the JVM parameters. Changed garbage collection cycles to place them in tune with backups/compactions so as to mitigate disk contention.
  • Good hands on experience with Solron effective and extensive search on the Cassandra Database cluster built on DataStax using the dynamic fields and Faceting.
  • Queried and analyzed data from Datastax Cassandra for quick searching, sorting and grouping.

Environment: s: Datastax 4.8, Cassandra 2.2, DevCenter, Cqlsh, Apache Hadoop 2.2.0, OpsCenter, Shell Scripting, Maven, Eclipse, Solr, Apache Kafka, Spark.

Confidential,Atlanta,GA

Cassandra Developer

Responsibilities:
  • Installed and configured Cassandra DSE multi-node, multi-data center cluster.
  • Involved in Data modelling design for various use cases
  • Requirement analysis, capacity planning, architecture, configuration.
  • Involved in Cassandra data modeling going through the phases of creating Conceptual model, Application flow, Logical model, Physical Optimization, Final Physical model.
  • Integrated Cassandra with Application Servers.
  • Familiar with all the internal tools of Cassandra.
  • Experience in working with CCM (Cassandra Cluster Manager).
  • Designed and implemented a strategy to upgrade the DataStax enterprise cluster.
  • Worked with Cassandra Query Language (CQL) to execute queries on the data persisting in theCassandra cluster.
  • Worked on tuning Bloom filters and configured compaction strategy based on the use case.
  • Performed Design, implementation and integration of client APIs.
  • Bulk loading of the data into Cassandra cluster using Java API's.
  • Maintained and Monitored the cluster using OpsCenter, Node tool.
  • Performed Benchmarking and Load testing using Cassandra-stress tool.
  • Performance Tuning, JVM Tuning, Disk Tuning to improve the overall performance of the cluster.
  • Fix Bugs and Troubleshoot operational issues as they occur.
  • Documented and demonstrated on various ways to securely connect to the cluster
  • Designed and implemented a strategy to securely move production data to Development for testing purposes using stable loader.
  • Gained Knowledge on Solr.
  • Worked on major and minor upgrades of cluster, Knowledge on applying updates and maintenance patches for the existing clusters

Environment: s: Cassandra 2.2, Java, Python, BASH Scripting, Apache Hadoop 2.2.0, Apache Kafka, MapReduce, Spark, OpenStack, JMeter, OpsCenter, DevCenter, Linux, XML.

Confidential,Milwaukee,WI

Hadoop/ Cassandra Developer

Responsibilities:

  • Developed Cassandra data model to match the business requirements.
  • Involved in Administration of Cassandra cluster along with Hadoop, Pig and Hive.
  • Extracted Data from RDMS into Cassandra Cluster using Sqoop.
  • Involved in commission and decommission of the nodes in the existing Cassandra cluster.
  • Designed, automated the process of installation and configuration of secure DataStax Enterprise Cassandra using Chef recipes.
  • Good Command on CQL to run queries on the data present in Cassandra Cluster with multi DC’s in 8 nodes each.
  • Experience in working with node tool to manage and monitor the clusters.
  • Familiar with all the internal tools of Cassandra.
  • Experience in working with CCM (Cassandra Cluster Manager).
  • Created Hive queries that analyze the Data and Provide Insights to the Analysts.
  • Used Oozie to manage Pig and Hive Jobs.
  • Involved in Load Testing and Analyzing Bottlenecks using Cassandra-stress tool.
  • Evaluated business requirements and prepared detailed specifications that follow project guidelines required to develop the application.
  • Involved in loading data from LINUX file system to HDFS.
  • Importing and exporting data into HDFS and HBase using Sqoop.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in loading data from edge node to HDFS using shell scripting.
  • Used Sqoop to import the data from RDBMS to Hadoop Distributed File System (HDFS) and later analyzed the imported data using Hadoop Components

Environment: CDH 5, Cassandra 2.0, Datastax 4.0, HDFS, MapReduce, Shell Scripting, Sqoop, Git, Maven, Apache Pig.

Confidential,Chesapeake, VA

Hadoop Developer

Responsibilities:

  • Worked on analyzing, writing Hadoop MapReduce jobs using Java API, Pig Latin and Hive.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in loading data from edge node to HDFS using shell scripting.
  • Used Sqoop to import the data from RDBMS to Hadoop Distributed File System (HDFS) and later analyzed the imported data using Hadoop Components
  • Created HBase tables to store variable data formats of data coming from different portfolios.
  • Implemented a script to transmit information from Oracle to HBase using Sqoop.
  • Implemented best income logic using Pig scripts and UDFs.
  • Implemented test scripts to support test driven development and continuous integration.
  • Worked on tuning the performance using Apache Pig queries.
  • Involved in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries and Pig scripts.
  • Managed and reviewed Hadoop log files.
  • Assisted application teams in installing Hadoop updates, operating system, patches and version upgrades when required
  • Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, manage and review data backups, manage and review Hadoop log files.
  • Used Oozie workflow engine to run multiple Hive and pig jobs automatically.
  • Shared responsibility for administration of Hadoop, Hive and Pig.

Environment: Hadoop, HDFS, Hive, Apache Pig, Sqoop, HBase, Shell Scripting, Ubuntu, Linux Red Hat, Zookeeper.

Confidential

Java Developer

Responsibilities:

  • Involved in the process of analysis, design, and development of the application.
  • Involved in the implementation of design using vital phases of the Software development life cycle (SDLC) that includes Development, Testing, Implementation and Maintenance Support in WATER FALL methodology.
  • Developed user interface using JSP and Java Script to simplify the complexities of the application.
  • Developed the application using Struts Framework that leverages classical Model View Controller (MVC) architecture.
  • Developed Struts Action Forms, Action classes and performed action mapping using Struts.
  • Performed data validation in Struts Form beans and Action Classes.
  • Used Form Based Authentication.
  • Involved in Documentation and Use case design using UML modeling include development of Class diagrams, Sequence diagrams, and Use case Transaction diagrams.
  • Participated in design and code reviews.
  • Developed ANT scripts and developed builds using Apache ANT.
  • Developed User Interface using AJAX in JSP and also performed client-side validation
  • Used Hibernate framework for Entity Relational Mapping
  • Used MySQL database for data persistence.
  • Expertise on entire Intranet and Internet applications using Web/Application Servers like Tomcat and WebSphere.
  • Involved in unit testing, integration testing.
  • Developed JUnit test cases for all the developed modules. Used SVN as version control

Environment: Java, J2EE, JSP, Struts, JNDI, HTML, XML, UML, DOM, SAX, ANT, AJAX, JavaScript, Rational Rose, Eclipse, Apache Tomcat, MySQL, LOG4J, SVN.

We'd love your feedback!