Cassandra Developer/dba Resume
Philadelphia, PA
SUMMARY:
- More than Seven Years of total IT experience which includes experience in various Big Data technologies, Data Analytics and Java technologies.
- Datastax certified Apache Cassandra Developer .
- Experience in building Cassandra cluster .
- Excellent CQL and data modeling knowledge from conceptual to physical data models including modern Chebotko diagrams.
- Excellent Knowledge in the Architecture of Cassandra .
- Having Knowledge on Datacenter, Datacenter racks, Cassandra Ring and nodes.
- Proficient in Compaction strategies in Cassandra system.
- Having knowledge on Spark and Scala Programming .
- Tested the application and the cluster with different consistency levels to check for the writes and reads performance with respective to Consistency Level.
- Hands on experience on Cassandra and oracle database architecture with better understanding and write processes including Mem - table, SSTable and Commit Log.
- Excellent knowledge on CQL (Cassandra Query Language), for retrieving the data present in Cassandra cluster by running queries in CQL.
- Involved in designing various stages of migrating data from RDBMS to Cassandra.
- Gained Hands on experience in analyzing the Cassandra data from flat files using Spark .
- Basic knowledge on Apache Solr .
- Excellent Understanding of Distributed Systems and Parallel Processing architecture.
- Creates Linux Virtual Machines using VMware Virtual Center.
- Having Hands on experience in installing, configuring and maintaining Linux.
- Can handle commissioning and decommissioning nodes along with monitoring of Cassandra Cluster.
- Experience on Cassandra node tool to manage Cassandra cluster.
- Proficient in managing and monitoring the cluster using Datastax OpsCenter.
- Experienced in Hadoop architecture, Hadoop Distributed File System .
- Excellent understanding on technologies such as Hive, PIG, Sqoop, Oozie, Hbase.
- Hands on experience with Importing/Exporting data using Hadoop Data Management tool Sqoop.
- Experienced using Hive and PIG for data analytics .
- Expertise in Application Development using Java/J2EE.
- Proficient in Web services, WCF (SOA), REST, and WEB-API using Java.
- RDBMS experience includes SQL-Server, Access database connectivity using ADO.NET/ADO, ODBC and programming using SQL, T-SQL, Stored Procedures etc.
- Experience in developing complex queries, Stored Procedures, Functions, Views, and Triggers using SQL Server.
- Background in a disciplined software development life cycle (SDLC) process and excellent analytical, programming and problem solving skills.
- Ability to take and give directions, instructions, provide assistance through completion of tasks and to work
TECHNICAL SKILLS:
Languages: C, C#, Java, J2EE (Servlets, JSP, JDBC, Java Beans), Python, SQL.
Web services: WCF, RESTful, Web Services
NoSQL Databases: Cassandra, Hbase.
Relational Databases: MySQL, Oracle10g.
Hadoop Eco System: HDFS, Map Reduce, YARN, Pig, Hive, Sqoop, Oozie, Zookeeper.
Cassandra: Cassandra, DataStax OpsCenter and DevCenter, nodetool, Spark on Cassandra and OpenStack.
Enterprise: J2EE, Spring Framework and Hibernate.
Databases: SQL Server 12/10/08, Oracle 9I/10g/11g/12c.
Operating Systems: Windows 8/7/XP/Vista, Linux.
Web technologies: XML, XAML, HTML, CSS, XMLHTTP, Java script.
Software: Eclipse, Visual Studio, NetBeans, Pycharm.
Tools: GIT, Maven, OpsCenter, DevCenter.
PROFESSIONAL EXPERIENCE:
Confidential,Philadelphia, PA
Cassandra Developer /DBA
Responsibilities:- Excellent Knowledge and understanding on Cassandra Architecture.
- Involved in the process of Conceptual and Physical Data Modeling techniques.
- Created data models in CQL for customer data.
- Involved in Hardware installation and capacity planning for cluster setup.
- Involved in the hardware decisions like CPU, RAM and disk types and quantities.
- Used the Spark - Cassandra Connector to load data to and from Cassandra.
- Prepared chebotko diagrams for logical and physical models during the phase of data modeling.
- Ran many performance tests using the Cassandra-stress tool in order to measure and improve the read and write performance of the cluster.
- Modified Cassandra.YAML files to set the configuration properties like cluster name, node addresses, seed provider, replication factors, memTable size and flush times etc.
- Used the DataStax OpsCenter for maintenance operations and keyspace and table management.
- Loaded and transformed large sets of structured, semi structured and unstructured data in various formats like text, zip, XML, CSV, YAML and JSON.
- Created data-models for customer data using the Cassandra Query Language.
- Involved in moving the SSTables data on to the live cluster.
- Created User defined types to store specialized data structures in Cassandra.
- Implemented advanced procedures like text analytics and processing using the in-memory computing capabilities like Spark.
- Used the Java API to load the data into Cassandra Clusters.
- Participated in NoSQL database integration and implementation.
- Enhanced and optimized product Spark code to aggregate, group and run data mining tasks using the Spark framework.
- Implemented a distributed messaging queue to integrate with Cassandra using Apache Kafka
- Tuned and recorded performance of Cassandra clusters by altering the JVM parameters. Changed garbage collection cycles to place them in tune with backups/compactions so as to mitigate disk contention.
- Good hands on experience with Solron effective and extensive search on the Cassandra Database cluster built on DataStax using the dynamic fields and Faceting.
- Queried and analyzed data from Datastax Cassandra for quick searching, sorting and grouping.
Environment: s: Datastax 4.8, Cassandra 2.2, DevCenter, Cqlsh, Apache Hadoop 2.2.0, OpsCenter, Shell Scripting, Maven, Eclipse, Solr, Apache Kafka, Spark.
Confidential,Atlanta,GA
Cassandra Developer
Responsibilities:- Installed and configured Cassandra DSE multi-node, multi-data center cluster.
- Involved in Data modelling design for various use cases
- Requirement analysis, capacity planning, architecture, configuration.
- Involved in Cassandra data modeling going through the phases of creating Conceptual model, Application flow, Logical model, Physical Optimization, Final Physical model.
- Integrated Cassandra with Application Servers.
- Familiar with all the internal tools of Cassandra.
- Experience in working with CCM (Cassandra Cluster Manager).
- Designed and implemented a strategy to upgrade the DataStax enterprise cluster.
- Worked with Cassandra Query Language (CQL) to execute queries on the data persisting in theCassandra cluster.
- Worked on tuning Bloom filters and configured compaction strategy based on the use case.
- Performed Design, implementation and integration of client APIs.
- Bulk loading of the data into Cassandra cluster using Java API's.
- Maintained and Monitored the cluster using OpsCenter, Node tool.
- Performed Benchmarking and Load testing using Cassandra-stress tool.
- Performance Tuning, JVM Tuning, Disk Tuning to improve the overall performance of the cluster.
- Fix Bugs and Troubleshoot operational issues as they occur.
- Documented and demonstrated on various ways to securely connect to the cluster
- Designed and implemented a strategy to securely move production data to Development for testing purposes using stable loader.
- Gained Knowledge on Solr.
- Worked on major and minor upgrades of cluster, Knowledge on applying updates and maintenance patches for the existing clusters
Environment: s: Cassandra 2.2, Java, Python, BASH Scripting, Apache Hadoop 2.2.0, Apache Kafka, MapReduce, Spark, OpenStack, JMeter, OpsCenter, DevCenter, Linux, XML.
Confidential,Milwaukee,WI
Hadoop/ Cassandra Developer
Responsibilities:
- Developed Cassandra data model to match the business requirements.
- Involved in Administration of Cassandra cluster along with Hadoop, Pig and Hive.
- Extracted Data from RDMS into Cassandra Cluster using Sqoop.
- Involved in commission and decommission of the nodes in the existing Cassandra cluster.
- Designed, automated the process of installation and configuration of secure DataStax Enterprise Cassandra using Chef recipes.
- Good Command on CQL to run queries on the data present in Cassandra Cluster with multi DC’s in 8 nodes each.
- Experience in working with node tool to manage and monitor the clusters.
- Familiar with all the internal tools of Cassandra.
- Experience in working with CCM (Cassandra Cluster Manager).
- Created Hive queries that analyze the Data and Provide Insights to the Analysts.
- Used Oozie to manage Pig and Hive Jobs.
- Involved in Load Testing and Analyzing Bottlenecks using Cassandra-stress tool.
- Evaluated business requirements and prepared detailed specifications that follow project guidelines required to develop the application.
- Involved in loading data from LINUX file system to HDFS.
- Importing and exporting data into HDFS and HBase using Sqoop.
- Responsible for building scalable distributed data solutions using Hadoop.
- Involved in loading data from edge node to HDFS using shell scripting.
- Used Sqoop to import the data from RDBMS to Hadoop Distributed File System (HDFS) and later analyzed the imported data using Hadoop Components
Environment: CDH 5, Cassandra 2.0, Datastax 4.0, HDFS, MapReduce, Shell Scripting, Sqoop, Git, Maven, Apache Pig.
Confidential,Chesapeake, VA
Hadoop Developer
Responsibilities:
- Worked on analyzing, writing Hadoop MapReduce jobs using Java API, Pig Latin and Hive.
- Responsible for building scalable distributed data solutions using Hadoop.
- Involved in loading data from edge node to HDFS using shell scripting.
- Used Sqoop to import the data from RDBMS to Hadoop Distributed File System (HDFS) and later analyzed the imported data using Hadoop Components
- Created HBase tables to store variable data formats of data coming from different portfolios.
- Implemented a script to transmit information from Oracle to HBase using Sqoop.
- Implemented best income logic using Pig scripts and UDFs.
- Implemented test scripts to support test driven development and continuous integration.
- Worked on tuning the performance using Apache Pig queries.
- Involved in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries and Pig scripts.
- Managed and reviewed Hadoop log files.
- Assisted application teams in installing Hadoop updates, operating system, patches and version upgrades when required
- Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting, manage and review data backups, manage and review Hadoop log files.
- Used Oozie workflow engine to run multiple Hive and pig jobs automatically.
- Shared responsibility for administration of Hadoop, Hive and Pig.
Environment: Hadoop, HDFS, Hive, Apache Pig, Sqoop, HBase, Shell Scripting, Ubuntu, Linux Red Hat, Zookeeper.
Confidential
Java Developer
Responsibilities:
- Involved in the process of analysis, design, and development of the application.
- Involved in the implementation of design using vital phases of the Software development life cycle (SDLC) that includes Development, Testing, Implementation and Maintenance Support in WATER FALL methodology.
- Developed user interface using JSP and Java Script to simplify the complexities of the application.
- Developed the application using Struts Framework that leverages classical Model View Controller (MVC) architecture.
- Developed Struts Action Forms, Action classes and performed action mapping using Struts.
- Performed data validation in Struts Form beans and Action Classes.
- Used Form Based Authentication.
- Involved in Documentation and Use case design using UML modeling include development of Class diagrams, Sequence diagrams, and Use case Transaction diagrams.
- Participated in design and code reviews.
- Developed ANT scripts and developed builds using Apache ANT.
- Developed User Interface using AJAX in JSP and also performed client-side validation
- Used Hibernate framework for Entity Relational Mapping
- Used MySQL database for data persistence.
- Expertise on entire Intranet and Internet applications using Web/Application Servers like Tomcat and WebSphere.
- Involved in unit testing, integration testing.
- Developed JUnit test cases for all the developed modules. Used SVN as version control
Environment: Java, J2EE, JSP, Struts, JNDI, HTML, XML, UML, DOM, SAX, ANT, AJAX, JavaScript, Rational Rose, Eclipse, Apache Tomcat, MySQL, LOG4J, SVN.