Data Engineer Resume
SUMMARY
- Over 9+ years of diverse IT experience working as Data Engineer & other roles in Confidential ecosystem
- SQL server DB/DW
- Business Intelligence Development (SSIS, SSAS, Azure ADF)
- Analytics using Power BI & SSRS
- Azure Databricks, Spark SQL
- Passionate Data analytics and Business Intelligence professional with experience in data - driven business transformation, analytics in supply chain, finance, energy, and healthcare industries
- Extensive Domain knowledge of oil and gas industry, worked on curating data related P&ID, Functional Location and Equipment Data, Protection Layers (SIS, SIF & SIL), Work Management and Maintenance Data(SAP & Maximo).
- Extensive Domain Knowledge of Supply Chain management and Insurance from projects worked on.
- Proficiency in translating business problems into data problems to give more insights on data.
- Experience in designing and developing data models (semantic layer) for analytics.
- Proficiency in analyzing large and complex datasets to come up with and present answers to several business questions using variety of tools and technologies like T-SQL, Spark SQL, Databricks.
- Expertise in building end to end ETL workflow using Confidential stack (ADF/SSIS & SSAS).
- Designed, Built, Deployed and Presented BI Solutions in close cross-functional collaboration with various departments like Sales, Marketing, Product Management, Production Development
- Lead Business Data Analysis, Reporting, Data Quality, Test coverage, Data Validation and Presentation to various business stakeholders and executives
- Integration and in-bound / out-bound data flow automation of various enterprise source systems using ETL & Data Warehousing concepts
- Curate data sourced out of Lake in to Databricks in different phases environments and perform Delta strong data engineering experience in Spark and Azure Databricks, running notebooks using ADF
- Working Knowledge of Delta Lake
- Create and maintain Spark jobs for data transformation and aggregation
- Monitor and maintain CI/CD pipelines in Azure environment for data load from Lake to DBX and from DBX to SQL DW.
- Maintained Batch Logging, Error Logging with Event Handlers and Configuring Connection Managers using SSIS/ADF
- Created data models and database objects
- Developed and implemented Historical and Incremental Loads using Databricks & Delta Lake run using ADF pipelines.
- Created sanity checks and alert for monitoring data quality
- Developed interactive dashboards and reports using Power BI for day-to-day business decision making and strategic planning needs
- Design, develop and maintain MS Power BI reports and dashboards and publish to targeted user groups
- Implemented row-level security using Power BI Desktop for different groups of stakeholders
- Developed and deployed OLAP cubes for various dashboards. Developed data models, dimensions, measures, and calculated measures (SSAS)
- Experience in using Data Analysis Expression (DAX) to create complex calculated columns and measures
- Creation of workflows with Data flow Task, Script Task, Execute SQL Task and Containers in SSIS/ADF.
- Used Transformations like Derived Column, Conditional Split, Aggregates, Sort, Data Conversion, Merge join and Union all. Did error handling while moving the data in SSIS/ADF(Dataflows).
- Implemented Performance Tuning at various levels.
- Created Configurations, Checkpoints and Breakpoints in packages for better reusability of SSIS Packages
- Experience in creating jobs, triggers and Scheduling SSIS Packages.
- Extensive Knowledge on creating Dashboards by using Reports and KPI’s.
- Excellent T- SQL programming skills in writing complex queries to create and update database objects.
- Extensive knowledge in writing MDX and DAX query.
- Created new and redesigned existing multidimensional cubes in tabular model using Azure analysis server.
- Having good experience in Indexes, Joins, Views, Stored procedures, Functions, Constraints.
- Experience in creating PowerPivot data models and integrate them with SharePoint and automate refresh.
- Created POC in PowerApps for automating a data entry process using SharePoint list.
- Experience working in Agile and Sprint methodologies.
TECHNICAL SKILLS
ETL Tools: T-SQL, SQL Server Integration Services (SSIS), Azure Data Factory
Visualization / Reporting Tools: Power BI, Tableau, SQL Server Reporting Services (SSRS)
Programming Languages: T-SQL, Python (Beginner), DAX
Analytics Platform: Databricks, Delta Lake and DW
Database Tools: Azure SQL, SQL Server Management Studio,MS SQL Server, MS Access
SQL Server Tools: Azure SQL Datawarehouse, SQL Server Business Intelligence Development studio, Enterprise Manager, SQL Profiler, Query Analyzer, Import & Export (DTS), SSAS
Other Tools: MS Office Suite ( Confidential Word, Power Point, MS Visio, MS Excel (VBA))
Operating Systems: Windows 10, Windows XP, Windows 7, Windows 2008, 2003R2, Windows Server 2008
PROFESSIONAL EXPERIENCE
Confidential
Data Engineer
Responsibilities:
- Implementation of end-to-end data solution on Azure using Azure Databricks, ADF, DW and PowerBI.
- Designed a robust data modelling environment using Databricks on Azure, enabling consumers to easily operate highly descriptive Notebooks in a fully governed environment
- Migration large data sets to Databricks (Spark), create and administer cluster, load data, configure data pipelines, loading data from ADLS Gen2 to Databricks using ADF pipelines
- Extensive hands-on experience of writing notebooks in data bricks using python/Spark SQL for complex data aggregations, transformations, schema operations. Good familiarity with Databricks delta and data frames concepts
- Extensive hands-on experience in designing and implementing scalable ETL pipelines to process variety of data types (structured, unstructured), file formats (json, csv, text delimited…)
- Collaborate with analysts to prepare complex data sets that can be used to solve data needs
- Built pipelines to take data from various telemetry streams and data sources to craft a unified data model for analytics and reporting.
- Creating Temporary views and loading curating data in destination tables.
- Databricks job configuration, Refactoring of ETL Databricks notebooks
- Created various pipelines to load the data from Azure data lake into Staging SQLDB and followed by to Azure SQL DB.
- Knowledge in retrieve, analyze and present data using Azure Data Explorer/Kusto
- Experience in Lift and shift existing SSIS packages using ADF.
- Created pipelines to load data from Lake to Databricks and Databricks to Azure SQL DB.
- Created Databricks Notebooks to streamline data and curate data for various business use cases.
- Designed various azure data factory pipelines to pull data various data sources and load the data into Azure SQL database.
- Triggers have been created for the pipelines to run on a day to day basis.
- Various tabular model cubes are created on top of the azure SQL database which are consumed for various reporting needs.
- Migrated the data from existing on prem database into Azure SQL.
- Worked on HugeData Transfersfrom & to SQL Server Databases using utilities / tools likeDTS, SSIS, BULK INSERTetc and used configuration files and variables for production deployment.
- Created SSIS packages to transform source data into dimension and fact tables.
- Developed complex Stored Procedures and views, ingested them in to SSIS packages. implemented slowly changing dimensions while transforming data in SSIS.
- Designed and loaded data into tabular model cubes to connect power bi reports.
- Created pipelines using ADF to run SQL scripts.
- Created database tables and stored procedures as required for reporting and ETL needs.
- Created/modified the existing SQL views as needed.
- Created Power BI reports and dashboards as per the business requirement using different data sources.
- Extensively used DAX to create complex calculated measures and columns in power bi and cubes.
- Implemented row level security within power bi to create user specific view of the data within report based on the role.
Environment: MS SQL Server 2012/2014, SSIS 2012/2014, SSAS, ADF, Databricks and ADL
Confidential
Application Development Team Lead
Responsibilities:
- Created Power BI reports and dashboards as per the business requirement.
- Designed SQL tables and stored procedures required for reports and ETL needs.
- Actively involved in the monthly publishing cycle which involves creating tables and loading the data from multiple sources to SQL server destination using SSIS, refreshing the tabular models and creating the power pivot reports.
- Created SQL jobs to schedule packages and stored procedures.
- Deployed the power BI reports to server and scheduled the automatic data refresh
- Modified the existing SQL views and created new views for various business needs
- Expert in creating SSRS reports connecting to SSAS Cubes (MDX) and SQL queries.
- Interactive data visualization products focused on business intelligence such as Power BI.
- Developed confidential proprietary analytical tools and reports with Confidential Excel, and Power Pivot, and Power Point.
- Created stored procedures to build the Fact tables in the data mart for Multi-Dimensional analysis using (SSAS) and produced ad-hoc, standard and super user reports using SSRS.
- Designed Dimensional Modeling using SSAS packages for End-User. Created Hierarchies and defined Dimension Relationships.
- Extensive Knowledge in SSAS storage and partitions, and Aggregations, calculation of queries with MDX, developing reports using MDX and SQL.
- Data management and data processing solutions using SSIS, this includes Error Handling and Slowly Changing Dimensions.
- Build Tabular, line charts, Graph charts visuals for Marketing KPI's metrics based on fields brought in and transform using Power Query functions and DAX calculated Measures.
- Captured the long running jobs and worked on fine tuning the store procedures and tables to improve the performance.
- Actively participated in the daily stand-up meetings and biweekly sprint planning sessions.
Environment: MS SQL Server 2012/2014, SSIS 2012/2014, SSAS, Power BI, DB2
Confidential
Database Administrator
Responsibilities:
- Creating and troubleshooting SQL Agent Job Failures, Log Shipping & Replication failures.
- Worked on resolving SQL Performance Alerts and Database/Disk Space/Capacity Issues.
- Troubleshooting Deadlock/Locking/Blocking, Service Outages (SQL Agent; Server; etc.)
- Creating Logins, Users and Roles in context of SQL Server security.
- Design and Build ETL (SSIS) to get data from different sources, with incremental loads
- Involved in scheduling, creating snapshots and subscriptions for the reports using SSRS
- Creating complex SQL Queries, Stored Procedures, Triggers, Views, Cursors, Joins, Constraints, DDL, DML and User Defined Functions to implement the business logic and also created clustered and non-clustered indexes.
- Optimized the performance of queries with modifications in TSQL queries, removed unnecessary columns, and eliminated redundant and inconsistent data.
- Created efficient, robust SSIS packages for data movement and loading with error checking and well documented code.
- Created SSIS packages to validate, extract, transform and load data to data warehouse databases.
- Extensively worked with SSIS tool suite, designed and created mappings using various SSIS transformations like OLEDB Command, Conditional Split, Lookup, Aggregator, Multicast and Derived Column.
- Wrote stored procedures and User Defined Scalar Functions (UDFs) to be used in the SSIS packages and S QL scripts .
- Created stored procedures to build the Fact tables in the data mart for Multi-Dimensional analysis using (SSAS) and produced ad-hoc, standard and super user reports using SSRS.
- Designed Dimensional Modeling using SSAS packages for End-User. Created Hierarchies and defined Dimension Relationships.
- Extensive Knowledge in SSAS storage and partitions, and Aggregations, calculation of queries with MDX, developing reports using MDX and SQL.
- Data management and data processing solutions using SSIS, this includes Error Handling and Slowly Changing Dimensions.
Environment: MS SQL Server 2008 R2, SSIS 2008 R2, SSAS 2008 R2.
Confidential
Software Engineer
Responsibilities:
- Responsible for designing database, creating tables, stored procedures, user defined functions and implementing joins, index using T-SQL.
- Responsible in monitoring and troubleshooting the production support and ensure that the SLA is met and no business impact.
- Responsible in creating power pivot reports as per the business requirement
- Designed dynamic SSIS Packages to transfer data crossing multiple platforms, validated data during transferring and archived data files for multiple DBMS.
- Designed SSIS packages to migrate the data from SQL Server to the staging area.
- Filtered bad data from the legacy system utilizing derived columns, lookups, fuzzy lookups, and conditional splits.
- Incorporated error redirection during ETL loads in SSIS Packages.
- Monitored and configured SSIS packages for dynamic execution; deployed packages to the testing environment.
- Generated various Drill-through Reports, Parameterized Reports, Matrix and Tabular Reports.
- Responsible for publishing and deploying the reports in Web Server and Report Server respectively.
- Optimizing the long running reports, packages and scripts.
- Responsible for deploying the objects (Scripts, packages, reports, web apps) into production environment on a weekly basis
- Monitoring the log shipping, mirroring, and replication jobs.
- Hands on Experience in installing, configuring, monitoring and trouble-shooting SQL Server 2005 and 2008.
Environment: MS SQL SERVER 2005, SSRS 2005, SSAS 2005.