Azure Data Engineer Resume
SUMMARY
- 6+ Experience in Design and Development of Enterprise ETL methodology and solutions for supporting Data Migration, data transformations, data processing and business reporting using Azure, Teradata and SQL Server.
- 2+ years of experience in Azure Cloud, Azure Data Factory, Azure Data Lake Storage, Azure Synapse Analytics, Azure HDInsight Bigdata Technologies (ApacheHadoop and ApacheSpark) and Data bricks.
- Experience in designing pipelines, Dataflows, Linked services and Trigger in Azure Data Factory and Azure Synapse Analytics.
- Experience working with different data storage options including Azure SQL Database, Azure SQL Data warehouse, Azure BLOB storage, ADLS Gen 1 and Gen 2.
- Developed scrips in Pyspark notebooks in reading data from different source systems into Databricks and Synapse Analytics.
- Experience working in reading data from different source systems and processing the Files and creating the files in Parquet format and Delta format.
- Good knowledge in Snowflake Data warehouse on Azure.
- Well - versed in creating Stored procedures using T-SQL, SQL Server Indexes, Triggers, and constraints in Azure SQL DB.
- In-depth understanding of Spark Architecture including Spark Core, Spark SQL, Data Frames, and Spark Streaming.
- Much experience as Teradata Developer with knowledge in Data stage, Informatica and exposure to design, development, testing, maintenance, and customer support environments on heterogeneous platforms.
- Good Experience in developing modules to extract, process & transfer the customer data using Teradata utilities.
- Expertise in creatingtables, views, macros, stored procedures, functions and join indexes in Teradata database.
- Strong experience in writing applications using Python using different libraries likePandas,NumPy, SciPy, Matplotlib, etc.
- Created Pipelines in Azure Data Factory ADF using Linked Services/Datasets/Pipeline/to Extract, Transform and load data from different sources like Azure SQL, Blob storage, Azure SQL data warehouse.
- Much experience in performing Data Modelling by designing Conceptual, Logical data models and translating them to Physical data models for high-volume datasets from various sources likeOracle, Teradata, Vertica, and SQL Server by using the Erwin tool.
- Also, design and develop relational databases for collecting and storing data and build and design data input and data collection mechanisms.
- Well-versed with Relational and Dimensional Modeling techniques likeStar, Snowflake Schema, Fact, and Dimensional Tables.
- Good knowledge in database like RDBMS, MySQL, Oracle, Microsoft SQL Server, Teradata, Snowflakes, Azure DB
- Hands-on experience across all stages of Software Development Life Cycle (SDLC) including business requirement analysis, data mapping, build, unit testing, systems integration, and user acceptance testing.
- Conducted overview meetings and code walk-through meetings and creating technical inspections document.
- Ability to connect and communicate with people at all levels.
- Self-motivated, hardworking, possess strong analytical and problem-solving skills and result-oriented with the spirit of teamwork and effective communication and interpersonal skills. Eager to learn, able to adapt quickly, well organized, and very reliable.
TECHNICAL SKILLS
Azure Cloud Platform: ADFv2, BLOB Storage, ADLS, Azure SQL DB, SQL server, Data bricks, Mapping Dataflow (MDF), CosmosDB, App Services, Logic Apps, Event Grid, Service Bus, Azure DevOps, GIT Repository Management, Azure Synapses
Teradata Tools and Utilities: FatsLoad, FastExport, MultiLoad, Tpump, TPT, Teradata SQL Assistant, BTEQ
Modelling & DA Specs Tools: CA Erwin Data Modeler, MS Visio, Gromit for DA Specs
ETL Tools: Informatica Power Center 9.1/8.6/8.5/8.1/7 ), DataStage, Abnitio
Programming Languages: Java, Python, T-SQL, LINUX Shell Scripting, AZURE PowerShell
Databases: RDBMS, MySQL, Oracle, Microsoft SQL Server, Teradata, Snowflakes, Azure DB
IDE and Tools: Eclipse, Tableau, IntelliJ, R Studio, SSMS, Maven, SBT, MS-Project, GitHub, MIcrosoft Visual Studio
Scheduler Tools: Tivoli workload scheduler, Autosys Scheduler, Control-M
Methodologies: WaterFall, Agile/Scrum, SDLC
PROFESSIONAL EXPERIENCE
Confidential
Azure Data Engineer
Environment: Azure SQL DB, Azure SQL Warehouse, Azure DevOps, Azure Cloud, Azure Data Factory (ADF v2), Azure functions Apps, Azure Data Lake, BLOB Storage, SQL server, Windows remote desktop, Data bricks, Python, Erwin Data Modelling Tool, Microsoft SQL server, Azure Data Lake, BLOB Storage.
Responsibilities:
- Performed analysis of complex business issues and provided recommendations for possible solutions. Writing SQL queries.
- Extract Transform and Load data from Sources Systems to Azure Data Storage services using a combination of Azure Data Factory, T-SQL, Spark SQL and U-SQL Azure Data Lake Analytics . Data Ingestion to one or more Azure Services - (Azure Data Lake, Azure Storage, Azure SQL, Azure DW) and processing the data in InAzure Databricks.
- Involved in understanding the Requirements of the End Users/Business Analysts and Developed Strategies for ETL processes.
- Created numerous pipelines in Azure using Azure Data Factory v2 to get the data from different source databases Informix
- Sybase etc by using different Azure Activities like Move &Transform, Copy, filter, for each, Databricks etc.
- Created several Databricks Spark jobs with Pyspark to perform several tables to table operations
- Creating pipelines, data flows and complex data transformations and manipulations using ADF and PySpark with Databricks.
- Improved performance by optimizing computing time to process the streaming data and saved cost to company by optimizing the cluster run time.
- Created Build and Release for multiple projects (modules) in production environment using Azure DevOps.
- Conducting code reviews for team members to ensure proper test coverage and consistent code standards.
- Responsible for documenting the process and cleanup of unwanted data.
- Hands on experience in working on Spark SQL queries, Data frames, and import data from Data sources, perform transformations; perform read/write operations.
- Worked in Agile development environment in sprint cycles of two weeks by dividing and organizing tasks.
- Experience in working withRestfulAPIs.
- Responsible for modifying the code, debugging, and testing the code before deploying on the production cluster.
- Experience in creating Data Lake Analytics accountand creating Date Lake Analytics Job.
- Helping team member to resolve any technical issue, Troubleshooting, Project Risk & Issue identification and management
- Addressing resource issue, Monthly one on one, Weekly meeting.
Confidential
Azure Data Engineer
Environment: Azure Cloud, SQL server, Teradata Utilities, Windows remote desktop, UNIX Shell Scripting, AZURE PowerShell, Data bricks, Python, Erwin Data Modelling Tool, Azure Data Factory (ADF v2), Azure functions Apps, Azure DataLake, BLOB Storage.
Responsibilities:
- Attended requirement calls and worked with Business Analyst and Solution Architects to understand the requirements from clients.
- Analyzed the data flow from different sources to target to provide the corresponding design Architecture in Azure environment.
- Take initiative and ownership to provide business solutions on time.
- Created High level technical design documents and Application design document as per the requirements and delivered clear, well-communicated and complete design documents.
- Created DA specs and Mapping Data flow and provided the details to developer along with HLDs.
- Created all the tables needed in Azure SQL server by using the Erwin Data Modelling tool according to the standards of Caesars.
- Created Application Interface Document for the downstream to create new interface to transfer and receive the files through Azure Data Share.
- Created Linked service to land the data from Caesars SFTP location to Azure Datalake.
- Created numerous pipelines in Azure using ADFv2 to get the data from different source databases Informix, Sybase etc.
- By using different Azure Activities like Move & Transform, Copy, filter, for each, Databricks etc..
- Created several Databricks jobs with Python PSQL to perform several tables to table operations.
- Extensively used SQL Server Import and Export Data tool.
- Created database users, logins and permissions to setup.
- Working with complex SQL, Stored Procedures, Triggers, and packages in large databases from various servers.
- Experience in creatingData LakeAnalytics accountand creatingDate Lake Analytics Jobin Azure Portal usingU-SQLScript.
- Helping team member to resolve any technical issue, Troubleshooting, Project Risk & Issue identification, and management
Confidential
ETL Developer
Environment: Teradata v15.x, Vertica v7.x Informatica 8.6.1/Informatica 9.6.1, Tivoli, Oracle 11g, Linux, Erwin Data Modelling
Responsibilities:
- Developing/implementing the code for new requirements from Business and the downstream systems as per Design specification document (DA Specs).
- This is to enhance Confidential and Confidential applications to bring new functionality/to enhance the existing functionality.
- Involved in creating mapping specification documents based on design documents
- Worked on Teradata BTEQ, Fast Export and TPT scripts for extract and loading the data to target tables
- Worked on developing mappings using various transformations like Lookup, Filter, Expression, Aggregator, Joiner etc. for implementing the ETLs
- Worked on performance tuning Teradata SQLs for various applications to improve the SLAs
- Conducting code review sessions with Offshore/Onshore team to make sure that Code is meeting the functionalities as well as ATT Code and Quality standards.
- Preparing Development package which has all the information about Software Configuration Management (SCM) data, Deployment guide and Code review artifacts.
- Supporting the project postproduction deployment during warranty (Red hat) period
- Analyzing and fixing the day-to-day data issues reported on Confidential Uverse and Confidential data by the Business users and down streams.
- Communicating the critical information related to the project activities across all the application teams
Confidential
SQL Server Developer
Environment: SQLServer 2016, SSIS, SSAS, SSRS, Microsoft Visual Studio 2012, Teradata, Informatica, T-SQL, MS Access, MS Excel, Putty, WinSCP, Outlook
Responsibilities:
- Created complex Freeform’s at Folder and Property level for data entry and maintenance.
- Undertake other software development project related tasks which may be reasonably expected.
- Analyzing the given documents, understanding and implementing it.
- Implemented the changes according to the change request document.
- Written many queries by using different sql objects, functions and keywords.
- Worked with various scripts (BTEQ, Fastload and Mload) by creating and manipulating them.
- Implemented performance tuning wherever required by creating required indexes, Eliminating spool space issues etc.
- Participate in peer-reviews, including, but not limited to design and development (coding).
- Participate and present status updates in each of the Stand-Up meetings on a regular basis.
- Solve complex problems in a time bound manner, document issues and fixes for easy system maintenance
- Mainly did error handling in ET, UV and Worktables.
- Analyzed the errors in error tables and reported to the client for restart ability.
- Did data reconciliation across source systems.
- Creating Unit test cases and implemented unit test results.
- Performance tuning, SQL query enhancements, code enhancements to achieve performance targets using Explain Plans.
- Involved in writing Fast load, Multi load and scripts to make loading, Updating.
- Answered Ad-hoc request of the client by using Functions, Analytical functions etc.