Data Modeler / Data Engineer Resume
NY
SUMMARY
- Over 7+ years of IT experience inDataModeling, Data Engineering, Designing, Data architecture, anddataanalysis.
- Excellent experience in writing SQL queries to validatedatamovement between different layers in data warehouse environment.
- Expertise in AWS Resources like EC2, S3, EBS, VPC, ELB, SNS, RDS, IAM, Route 53, Auto scaling.
- Extensively worked on GCP google cloud platform for the Data acquisition process.
- Experienced in Conceptual, Logical visionary and Physical Modeling for Online Transaction Processing (OLTP), Online Analytical Processing (OLAP) andDataWarehousing and developed TEMPeffective and efficient solutions and ensuring client deliverables within committed timelines.
- Experienced insnowflake and its analytics tools of business intelligence.
- Experienced in Performance tuning on oracle databases by leveraging explain plans, and tuningSQL queries.Cloud Formation, Cloud Watch, Security Groups.
- Involved in implementing the Referential Integrity by using Primary and Foreign key relationships.
- Experienced with both DDL and DMLincluding Joins, Functions, Indexes, Views, Constraints, Primary Keys and Foreign Keys.
- Excellent in creating various artifacts for projects which include specification documents,data mapping anddataanalysis documents.Proficient in Normalization (1NF/2NF/3NF) /De - normalization techniques in relational/dimensional database environments and Design Patterns, Cube design, BI Strategy and Design.
- Experience in Developing Teradata 3NF and Semantic Data Models, worked with Teradata utilities likeBTEQ, Fast Export, Fast Load, Multi Load, TPTto export and load data to/from different source systems including flat files and includes developing Semantic Layer Views along with Indexes, Collect Stats, Compressions, Performance
- Experienced in trouble shooting SQL queries, ETL jobs, Datawarehouse, Data mart, data store models.
- In depth noledge of architecture of Yarn and Spark.
- Experience in designing star schema, Snowflake schema for Data Warehouse, ODS architecture.
- Experience in SQL queries,PL/SQLPackages, SQLPlus, Stored Procedures, Functions, Triggers andPerformance Analysis, Creating Partitions,Indexes, aggregating tables when required.
- Have in depth noledge in creating business requirement documents (BRD), in process documentation and client requirement document.
- Experience in backend programming including schema and table design, Oracle PL/SQL stored packages/procedures/Functions, Triggers, Views, and different Indexes, performance tuning, DBA activities.
- Excellent analytical, inter-personal and communication skills with a strong technical background.
- me is well versed to deliver and administer highly comprehensive Master Data Management designs to consolidate data across a wide range of business entities.
- Reviewed and approved hundreds of data modeling/database design documents. Helped increase the efficiency of data modeling procedures. Developed a statistical tool to optimize the design of large database tables and achieve data storage savings more than 70%.
- Experience with Data Analysis, mapping source and target systems for data migration efforts and resolving issues relating to data migration.
- In depth noledge of data forecasting techniques, visualization of data in Microsoft excel, Minitabnormalization of data, regression analysis.
- Sound with the principles of lean six sigma lean manufacturing, manufacturing analysis, decision making under uncertainty, in depth noledge of statistics supply chain management, simulation techniques.
TECHNICAL SKILLS
Operating Systems: Unix, Linux(Ubuntu, CentOS), Mac OS, OpenSUSE, Windows 2003/2008/2012/ XP/7/8/9X/NT/Vista
Hadoop Ecosystem/Distributions: HDFS, MapReduce, Yarn, Oozie, Zookeeper, Job Tracker, Task Tracker, Name Node, Data Node, Cloudera, Horton works
Big Data Ecosystem: Hadoop, Spark, MapReduce, YARN, Hive, SparkSQL, Impala, Pig, SqoopHBase, Flume, Oozie, Zookeeper, Avro, Parquet, Maven, Snappy, Hue
Data Ingestion: Sqoop, Flume, NiFi, Kafka
Cloud Computing Tools: Snowflake, SnowSQL, AWS, Databricks, GCP, Azure data lake services, Amazon EC2
NoSQL Databases: HBase, Cassandra, MongoDB, CouchDB, Apache, Hadoop HBase
Programming Languages: Python( Jupyter Notebook, PyCharm IDE), R, Java, C, Scala, SQL, PL/SQL, VBScript and Shell Scripts, XML, HTML, Visual Basic 6.0, Fox Pro, SAS
Frameworks: MVC, Struts, Spring, Hibernate
Web Technologies: HTML 5, CSS 3, XML, JavaScript, Maven Spring 4, Spring MVC, JSP, Angular JS, Ajax, jQuery, XSP, WSDL, JSON
Scripting Languages: Bash, Pearl, Python, R Language
Databases: Snowflake Cloud DB, Oracle, MySQL, Teradata 12/14, DB2 10.5, MS Access, SQL Server 2000/2005/2008/2012 , PostgreSQL 9.3, Sybase ASE 11.9.2, Netezza, AmazonRDS
SQL Server Tools: SQL Server Management studio, Enterprise Manager, Query Analyzer, Profiler, Export and Import (DTS)
IDE: IntelliJ, Eclipse, Visual Studio, IDLE
Web Services: Restful, SOAP, O9iAS, Oracle Form Server, Weblogic 8.1/10.3, Web Sphere MQ 6.0
Packages and Tools: MS-Office, TOAD, SQL Developer & Navigator, Share point portal server, Visual Source Safe, SVN, TFS, BTEQ
Methodologies: Agile, Scrum, Iterative Development, Waterfall Model, UML, Design Patterns, UML
ETL/ Data: Tensor flow, Data API, PySpark, Pervasive Cosmos Business Integrator ( Data Junction Tool), CTRL-M, Data Stage, Informatica Power Center 9.6.1/9.5/8.6.1/8.1/7.1 , Talend, Pentaho, Microsoft SSIS, Data Stage 7.5, Ab Initio
OLAP Tools: MS Analysis Services, Business Objects & Crystal Reports 9, MS SQL Analysis Manager, DB2 OLAP, Cognos Powerplay
Warehousing and Modelling/ Architect Tools: Erwin 7.3&9.5 ( Dimensional Data Modelling, Relational DM, Star Schema, Snow-flake, Fact and Dimensional Tables, Physical and Logical DM, Canonical Modelling), Visio 6.0, ER/Studio, Rational System Architect, IBM Infosphere DA, MS Visio Professional, DTM, DTS 2000, SSIS, SSAS
Reporting / BI Tools: MS Excel, Tableau, Tableau Server and Reader, Power BI, QlikView, SAP Business objects, Crystal Reports, SSRS, Splunk
Utilities/ Tools: Bugzilla, QuickTestPro 9.2, Selenium, Quality Center, Test Link, TWS, Documentum, Tortoise SVN, Putty, WIN SCP, Log4J, Junit,, GIT, Jasper Reports, Jenkins, Eclipse, TOMCAT, NetBeans, SVN, SOAPUI
PROFESSIONAL EXPERIENCE
Confidential, NY
Data Modeler / Data Engineer
Responsibilities:
- Involved inDatamappingspecifications to create and execute detailed system test plans. Thedatamapping specifies whatdatawill be extracted from an internaldatawarehouse, transformed, andsent to an external entity.
- Have been working with AWS cloud services (VPC, EC2, S3, RDS, Redshift, Data Pipeline, EMR, DynamoDB, Workspaces, Lambda, Kinesis, RDS, SNS, SQS).
- Documentedlogical,physical, relational, and dimensionaldatamodels. Designed theDataMartsin dimensionaldatamodeling using star and snowflake schemas.
- Refined CCPA/ CPRA technical workflow requirements for improvements of end-to-end request to response.
- Worked on GCP for the purpose of data migration from Oracle database to GCP.
- Worked on snowflakeFor dynamic data migration
- Developed documents and testedCCPA/ CPRA access request data collection processes Develop data deletion request notification processes with verification receipt validation tracking.
- Prepared documentation for all entities, attributes,datarelationships, primary and foreign keystructures, allowed values, codes, business rules, and glossary evolve and change during theproject.
- Tested the ETL process for both before data validation and after data validation process. Tested themessages published by ETL tool and data loaded into various databases.
- Coordinated withDBAondatabase build and tablenormalizationsandde-normalizations.
- Identified the entities and relationship between the entities to developConceptual Model usingERWIN.
- Involved inSQL Development,Unit TestingandPerformance Tuningand to ensure testing issues areresolved based on using defect reports.
- Created Logical Model from the conceptual model.
- Administered differentDatamappingactivities from Source systems.
- Indulged indata modelreviews with internaldata architect, business analysts,and business userswith explanation of the data model to make sure it is in-line with business requirements.
- Involved withData Profilingactivities for new sources before creating new subject areas indata warehouse.
- Extensively worked onData Governance, me.e.,Metadata management, Master data Management, DataQuality, and Data Security.
- Redefined many attributes and relationships in theirverse engineeredmodel and cleansedunwanted tables/columns onTeradatadatabase as part of data analysis responsibilities.
- Performed complexdata analysisin support of ad-hoc and standing customer requests.
- Delivereddata solutionsin report/presentation format according to customer specifications andtimelines.
- UsedReverse Engineeringapproach to redefine entities, relationships, and attributes in the datamodel as per new specifications inErwinafter analyzing the database systems currently in use.
- Enforced referential integrity in theOLTP data modelfor consistent relationship between tables andefficient database design.
- Created the test environment for Staging area, loading the Staging area withdatafrom multiple sources.
Confidential, Rhode Island
Data Modeler
Responsibilities:
- Worked with Data Vault Methodology Developed normalized Logical and Physical databasemodels.
- Owned and managed all changes to the data models. Created data models, solution designs anddata architecture documentation for complex information systems.
- TEMPEffectively articulated reasoning for data model design decisions and strategically incorporatedTeam member feedback to produce the highest quality data models.
- Worked with reverse engineering Data Model from Database instance and Scripts.
- Proficiency in SQL acrossseveral dialects (we commonly write MySQL, PostgreSQL, Redshift, SQL Server, and Oracle).
- Exposure on AZURE cloud platform.
- Implemented Data Vault Modeling Concept solved the problem of dealing with change in theenvironment by separating the business keys and the associations between those business keys,from the descriptive attributes of those keys using HUB, LINKS tables and Satellites.
- Developed mapping spreadsheets for (ETL) team with source to target data mapping withphysical naming standards, datatypes, volumetric, domain definitions, and corporate meta-datadefinitions.
- Distributed database Design, Data modeling, Development and Support in Data tax Cassandradistribution.
- Data modeling, Design, implement, and deploy high-performance, custom applications at scaleon Hadoop/Spark.
- Worked on AWS Redshift and RDS for implementing models and data on RDS and RedshiftAble to guide / partner with VP / Directors for architecting solutions for the big data Organization.
- Established and maintained comprehensive data model documentation including detaileddescriptions of business entities, attributes, and data relationships.
- Perform profiling, troubleshooting of existing solutions.Provide capacity planning, performance tuning and monitoring.
- Develop and implement data cleansing, data security, data profiling and data monitoringprocesses.
- Cassandra products strengths and weakness to produce efficient schema designs dat servesTEMPeffective and high-performance queries.
- Maintain and work with our data pipeline dat transfers and processes several terabytes of datausing Spark, Scala, Python, Apache Kafka, Pig/Hive & amp Impala.
- Apply data analysis, data mining and data engineering to present data clearly.
- Ensure high-quality data and understand how data is generated out experimental design and howthese experiments can produce actionable, trustworthy conclusions.
- Full life cycle of Data Lake, Data Warehouse with Big data technologieslikeSpark, Hadoop, Cassandra.
Environment: used: AWS Redshift, RDS, Big Data, JDBC, Cassandra, NOSQL, Spark, Scala, Python, Hadoop,MySQL, PostgreSQL, SQL Server.
Confidential, NJ
Data Modeler
Responsibilities:
- Created Mapping using Informatica Cloud for the salesforce project.
- Conducted one-on-one sessions with business users to gather the requirements.
- Analyzed database requirements in detail with the project stakeholders by conducting Joint Requirements Development sessions.
- Identified, formulated, and documented detailed business rules and Use Cases based on requirementsanalysis.
- DevelopedData Mapping, Data Governance, and TransformationandCleansingrules for the Master Data.
- Management Architecture involving OLTP, ODS and OLAP.
- Created Logical and Physical Data Models by usingErwinbased on requirements analysis.
- Used Forward Engineering to create a Physical Data Model withDDLdat best suits the requirements from the Logical Data Model.
- Experience working with Big Data eco-system including tools such as HDFS, Nifi, Spark streaming, Map.
- UsedIDAfor reverse engineering to connect to existing database and ODS to create graphical.
- Representation in the form of Entity Relationships Diagram (ERD)
- Worked with Database Administrators, Business Analysts and Content Developers to conduct designreviews and validate the developed models.
- Facilitated development, testing and maintenance of quality guidelines and procedures along with necessary documentation.
- Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch datafrom legacy Oracle andSQL Serverdatabase systems.
Environment: - Salesforce, Informatica Power Designer, Informatica Cloud, HDFS, Nifi, Spark streaming, Map, Reduce, Sqoop, HBase, Hive, Impala, Oozie.
Confidential
PL/SQL developer
Responsibilities:
- Involved in identifying the process flow, the workflow, and data flow of the core systems.Worked extensively on user end.
- Requirements gathering and gap analysis.
- Involved in full development cycle of Planning, Analysis, Design, Development, Testing and Implementation.
- Testing all forms, PL/SQL code for logic correction.
- UsedSQL Server SSIStoolto build high performance data integration solution.includingextraction, Transformation.
- Load packagesfordata warehousing.Extracted data from theXMLfile andloaded it intothedatabase.
- Involved inbuilding, debugging, and runningforms. DevelopedPL/SQL triggersandmaster tablesforautomaticCreationof primary keys.
- Involved in Data analysis for data conversion included data mapping from source to target databaseschemas.
- Specification and writing data extract scripts/programming of data conversion, in test andproduction environments.
- Handled errors usingException handlingextensively for the ease of debugging and displaying the error messagesin the application.
- Developed AdvancePL/SQLpackages, procedures, triggers, functions, IndexesandCollectionsto implement.
- Business logicusingSQLNavigator. Generated server-sidePL/SQLscripts fordata manipulationandvalidation materialized views for remote instance.
- Designed and developedOracle forms&reportsgenerating up to 60 reports.
- Involved inData loadingandExtractingfunctions usingSQL*Loader.
- Performed Database Administration of all database objects includingtables, clusters, indexes, views, Sequences packages andprocedures.
- Designed and developed all thetables, viewsfor the system in Oracle.