We provide IT Staff Augmentation Services!

Senior Nosql Developer Resume

3.00/5 (Submit Your Rating)

CAREER OBJECTIVE:

A versatile engineer having 10+ years’ experience with strong debugging skills offers a solid understanding, use of algorithms and analytics as well as proven technical leadership and project management skills. Leverages comprehensive technical expertise for software engineering, data management, and developing innovative business solutions for corporate customers.

PROFESSIONAL SUMMARY:

  • Highly skilled software development professional bringing more than 10 years in design, development and integration.
  • Certified Power Center Data Integration 9.x: Developer Specialist by Informatica University
  • Certified Teradata Basics (TEO - 121) by Teradata Cooperation.
  • Involved in 5 Implementation Projects from scratch and one Internship.
  • Experience in creating semantic layer on HDFS using Map Reduce Programming model, Pig and Hive ecosystems.
  • Experience in creating Ingestion modules to Ingest data into HDFS from Heterogenous data sources.
  • Experience in creating XQUERY Transformation modules using them while ingestion of documents into Marklogic.
  • Worked on Java API operations with Marklogic server.
  • Experienced in modelling XML data using Resource Description Framework, Ingesting with Embedded Triples and querying them using SPARQL in Marklogic.
  • Worked on XML document modelling to SQL modelling and Performing Analytics.
  • Expertise in XML Technologies XML, XSLT, XSD, XSD validation & JSON and Worked on parsers like SAX and DOM.
  • Strong knowledge on XQUERY, XPATH and XSLTs.
  • Worked on implementing CURD operations using Marklogic NoSQL REST API Services
  • Worked on Integrating Messages Queues systems with Marklogic database using Java API
  • Experienced with Roxy Framework for Deploying the modules to Marklogic Database.
  • Worked of post process of Large volume of XML data in multithreading way within Marklogic using CoRB.
  • Experienced in automating the Ingestion using MLCP and CURL library.
  • Extensively worked on creating REST API’s in Marklogic which gives UI Elements such as filters (Range filters, Date and String filters), dropdown lists, search fields, pagination.
  • Experience in creating Cluster Health Reporting and forecasting for Marklogic Distributed NoSQL database.
  • Involved in Business user support, generating ad hoc analytical reports and point in queries.
  • Experience in playing Senior developer role for Data warehousing and Bigdata projects.
  • Extensively worked on ETL Systems, Source Systems Integrations, Design, Development, Debugging, Maintenance and End Reporting.
  • Well versed in OLAP and OLTP systems.
  • Strong knowledge in ER concept, Facts, Dimensions and Slowly changing dimensions and Dimensional modelling.
  • Extensively worked on Scripting (Both Shell and Batch) used to integrate Source systems with the applications.
  • Experience in writing, testing and implementation of PL/SQL Procedures, Functions and Load utility scripts (BTEQ, MLOAD, FLOAD).
  • Contributed to logical and physical data modeling for the new data warehouse.
  • Experienced in developing Informatica web services to load the data into Master Data Management (EBX)
  • Developed Informatica Code Review Tool (Namely ICR) reviews Informatica code against a set of widely accepted guidelines which can be pre-configured and change over dynamically. This tool is an xml file dependent tool written in Java.

TECHNICAL SKILLS:

Programming Languages: Java, Python, R Programming, XQuery, Batch and Shell Scripting.

Big Data Technologies: Hadoop MapReduce, Pig, Hive, Sqoop, Mark logic (NoSQL)

Databases & Tools: Informatica, Teradata, SQL Server, PL/SQL, SPARQL and Tableau.

Other Utilities: Control-M, Erwin, IBM Clear Case, Apache Subversion.

PROFESSIONAL EXPERIENCE:

Confidential

Senior NoSQL Developer

Responsibilities:

  • Designing the API responses inline to user stories.
  • Developing XQuery transformations modules used for ingestions.
  • Creating API on Egress layer to expose the data to front end portal.
  • Developing XQuery modules which provides information on UI elements (Filters, search, dropdown lists and pagination).
  • Transforming huge volume of documents in a multithreading within Marklogic using
  • CoRB tool.
  • Working with users and understanding the requirements.
  • Roxy framework for project building using XQuery.
  • Code version management in IBM Clear Case.
  • Creating Re-usable library modules and REST Endpoints modules.
  • Creating and maintaining HTTP app servers.
  • Creating indexes and rebuilding up new data loads during off production hours.
  • Monitoring the data loads and cluster health management in regular intervals.
  • Running data loads and Data transformations.
  • Creating ingestion pipelines (using MLCP, CURL and Shell Scripting) according to business needs.
  • Go live support and user platform support.

Confidential

Big Data Developer

Responsibilities:

  • Developing XQuery transformation modules using while ingesting data via ML content pump util.
  • Modelling data per Resource Description Framework Triples and Loading embedded triples using MLCP.
  • Creating Report using SPARQL by querying embedded triples.
  • Integration of Message Queues to Marklogic using Java API operations to load xml’s in canonical fashion.
  • XML Document modelling to SQL data modelling and creating SQL views for reporting.
  • Roxy framework for code migrations and code unit testing
  • Using ODBC drivers along with XDBC app service accessing the views from Tableau connecting to Marklogic.
  • Map Reduce, Pig and Hive QL scripts to create semantic layer on HFDS and access the layer using Java API and load into data into Marklogic incrementally.
  • Creating Pig Load & Store, Filter UDF’s within the semantic layer.
  • Creating Schemas and Views using HiveQL.

Confidential

Data Analyst

Responsibilities:

  • Converting the financial analyst’s intelligence into rules and document them for the use of designing classification algorithm.
  • Implementing the algorithms using Python programming.
  • Application of Text retrieval using OCR for scanned set of images which are embedded in PDF documents.
  • Term list creation for each set of document type and calculating the percent of words that appear for each document type.
  • Empirical study on evaluating suitable classification models. Building Training data, Cross validation data and Test Data.
  • Exploration of IBM Watson Analytics API calls via Python API.

Confidential

ETL Engineer

Responsibilities:

  • Logical to Physical data model maintenance using Sparx System Data modeler tool and generating scripts and executing on database.
  • Understanding, designing & implementation of ETL mappings.
  • Enhancement and developing of product backlogs by revising Informatica code mappings, Java routines, UNIX Shell Scripts, PL/SQL Procedures and semantic layer views.
  • Using Informatica web services load the data into Master data management (EBX).
  • Implementing load balance control mechanisms on MDM by making use of EBX tool back end tables read strategy.
  • FTP & SFTP scripts used to pull data files from the various externals system and transfer of encrypted files from Zurich to NY instances using batch and shell scripting.
  • Building & maintenance of semantic layer (data virtualization views) created on composite.
  • Developing Reusable PL/SQL Procedures for both Business rules and data validations.
  • Scheduling Jobs in Control-m either in windows or UNIX, L3 support activities.
  • Code Management using Subversion tool and Deployments.
  • Conducting scrum calls and managing the sprint releases.
  • Effort logging in JIRA and creating Burn down charts.

Confidential

ETL Engineer

Responsibilities:

  • Requirement analysis and design ETL pipeline design.
  • Conducting meeting with users and clarifying the business rules.
  • Developing ETL mappings in Informatica Tool and creating reusable routines.
  • Developing reusable maplets for change data capture (CDC) and Slowly changing dimensions.
  • Creating Drivers for TPUMP, FLOAD and MLOAD at informatica server level and creating data load pipelines using them.
  • Data validation SQL scripts between sources to Targets.
  • Creation of BTEQ (Teradata Load utility) & UNIX scripts for data transfers between layers within Information factory (i.e. Production).
  • Maintenance of disaster recover database.
  • Implementation of semantic layer using DB Views and expose the data for reporting.
  • Integrate security measures in sematic layer.
  • Integrate load utility (BTEQ, MLOAD & FLOAD) scripts using Shell scripting.
  • Created ad-hoc queries for Auditing of the data.
  • Metadata creation and maintenance.
  • Issue Tracking and clarification log maintenance.
  • Supporting QA and creating regression suites.

We'd love your feedback!