Data Analyst Resume
SUMMARY
- Around 5 years of professional experience as a Data Analyst with excellent understanding of analysing and documenting Business Requirements, Functional Specifications, Business Process Flow, Business Process Mapping, data extraction, manipulation, visualization, validation techniques and Data Modelling.
- Experience in all phases of Software Development Life Cycle (SDLC), Waterfall and AGILE methodologies. Performed various system - related analytical activities within all the phases.
- Ability to gather and document Business Requirements, experienced in writing Use Cases. Proficiency in SDLC life cycle, understands the workflow concept, ability to gather and document.
- Expertise in various phases like understanding the User Requirements, Analysis/Design, Testing, Project Management, & Product Development along with end to end product delivery across the SDLC & STLC phases.
- Experienced in requirements gathering and analysis using various elicitation techniques like interviews, surveys, JAD sessions, observation, prototyping and brainstorming.
- Experienced in various diagramming techniques like wireframes, Process maps, flowcharts, functional demonstration, context diagrams and BPM modelling techniques.
- Worked with AWS Cloud platform and its features which includes EC2, VPC, RDS, EBS, S3, CloudWatch, Cloud Trail, CloudFormation and Autoscaling etc.
- High expertise in tracing requirements throughout the development process and verifying adherence to Requirement Traceability Matrix (RTM).
- Expertise in performing GAP analysis, Requirements Traceability Matrix (RTM), Data mapping and Data Modelling.
- Have good working knowledge and experience on Microsoft Word, Excel, Visio, PowerPoint and HP Quality Center.
- Experience in conducting GAP analysis, SWOT analysis, Cost benefit analysis and ROI analysis.
- Used AWS command line client and management console to interact with AWS resources and APIs.
- Hands-on experience with MS Visio and use case diagrams for creating data flow/process flow diagrams such as flowcharts, activity charts, sequence diagrams as per UML.
- Knowledge of Data Warehousing concepts and Extract Transform and Load (ETL) processes.
- Experience in Data-Modelling, Data Warehousing, Schemas, Data Marts and Extract/Transform/Load (ETL).
- Key strengths include analysis using BI tools, COSMOS with Scope Scripts, Reporting using Power BI, ETL using SSIS, requirement gathering, RDBMS concepts, SQL Queries and preparation of design documents.
- Experience in data analysis, data visualization using Tableau/Microsoft Power BI, model building with machine learning algorithms for prediction and forecasting using data (historical or time series with regression techniques), using statistical/econometric tools like SAS and R, data mining using python, SQL, Hadoop, Spark, Hive, etc.
- Reviewed basic SQL queries and edited inner, left, and right joins in Tableau Desktop by connecting live/dynamic and static datasets.
- Good BPM understanding.
- Strong skills in SQL, data warehouse, data exploration, data extraction, data validation, reporting and excel.
- Used tools like Tableau and Microsoft Excel for data analysis and generating data reports
- Developed visualizations using Tableau for better understanding of data, performed data cleaning, normalization, data transformation
- Extensively used SQL for accessing and manipulating database systems.
- Adept in RDBMS such as Oracle, MS SQL Server and MS Access &also skilled at writing SQL queries and Stored Procedures.
- User Acceptance Testing (UAT) and Manual testing (Functionality Testing) of UI and Web Applications. Extensively worked in creating Test Procedures, Test plans, Test cases and reviewing them for quality assurance.
- Involved in System Integration Testing (SIT), Regression Testing, GUI Testing, Performance Testing & User Acceptance Testing (UAT).
TECHNICAL SKILLS
Programming Languages: C, Visual Basic, and C++, VB 6.0, SQL, Hadoop (Hive, Pig), Python, R.
Scripting Languages: MS-DOS, Bash, Korn.
ETL tools: Confidential Power center, SSIS, AB Initio.
Data modelling: Sybase Power Designer / IBM Data Architect.
Frameworks: Struts, Spring, Hibernate, Spring MVC, Spring Web Flow, Spring IOC, Spring AOP, Groovy.
Application/Web Servers: JBoss, Glassfish 2.1, Web Logic, Web Sphere, Apache Tomcat Server.
MS-Office Package: Microsoft Office (Windows, Word, Excel, PowerPoint, Visio, Project).
Visualization tools: Tableau Desktop, Python, Pandas, NumPy, Datorama.
ETL Tools / Tracking tool: Confidential, SSIS, SSAS, SSRS / JIRA.
Database Development: T-SQL and PL/SQL, Microsoft Hyper-V Servers.
Databases: Teradata R12 R13 R14.10, MS SQL Server, DB2, Netezza.
Testing and defect tracking: HP/Mercury (Quality Center, Win Runner, Quick Test Professional, Performance Center, Requisite, MS Visio & Visual Source Safe.
Operating Systems: Windows, UNIX, Sun Solaris.
PROFESSIONAL EXPERIENCE
Confidential
Data Analyst
Responsibilities:
- Analysed the requirements and segregated them into high level and low-level Use Cases, activity diagrams using Rational Rose according to UML methodology thus defining the Data Process Models
- Built out the data and reporting infrastructure from the ground up using Tableau and SQL to provide real-time insights
- Was responsible for indexing of the tables in that data warehouse. Used senior level SQL query skills (Oracle and TSQL) in analysing and validating SSIS ETL database data warehouse processes
- Converted Business Requirements to the Functional Specification and Conducted JAD Sessions to develop an architectural solution that the application meets the business requirements, resolve open issues, and change requests
- Using Python, SQL, and Excel developed and owned the reporting for a nationwide retention program, saving ~100 hours of labour each month.
- Used tools like Tableau and Microsoft Excel for data analysis and generating data reports
- Analysed the data warehouse project database requirements from the users in terms of the dimensions they want to measure and the facts for which the dimensions need to be analysed
- Responsible for physical/logical data modelling; metadata documentation; user documentation; and production specs
- Identify business rules for data migration and perform data administration through data models and metadata
- Used Toad Data Analysts to connect to Oracle DB2 for data analysis
- Create and maintain requirements, test cases and defects in Mercury Quality Center
- Develop / Auto deploy content using AWS (Amazon Web Services), GIT/Bitbucket, Maven, Jenkins
- Develop custom tools/scripts/packaging solutions for AEM using Java/Unix
- Develop integration solutions between AEM, AWS (Lambda, S3, API Gateway and Cloud Formation) and Spredfast (Social) Platforms
- Developed detailed ERDs and DFDs using various modelling tools and developed databases based on the system model following the techniques in SDLC (software development life cycle)
- Gathered, Analysed and Translated systems requirements into technical specifications utilizing UML and RUP methodology. Responsible for creating different sessions and workflows to load the data to Data Warehouse using Confidential Workflow Manager
- Worked on AWS Elastic Beanstalk to deploy, monitor, and scale an application
- Create new EC2 instance in AWS, allocate volumes and giving Provisionals using IAM
- Analysing the existing reports, reporting system. Worked on exporting data using Teradata
- Utilized Agile/ SCRUM and PMI methodologies to monitor steer and develop project objectives
- Implemented transformation component of DataStage to integrate the data and to implement the data to implement the business logic. Created Test Cases and scenarios for Unit, Regression, Data Integration as well as Back end and System testing
- Strong interpersonal and communication skills within all levels of the organization and familiarity of regulatory mandates and internal controls
- Proposed solutions for reporting needs and developed prototypes using SQL and Business Objects that address these needs. Developed data conversion strategy for data migration from legacy systems to Technology product modules
- Familiarity with reporting and Business intelligence, Toad Data tools such as Crystal Reports
- Develop and maintain sales reporting using MS Excel queries, SQL in Teradata, and MS Access
- Extensive experience in testing and implanting Extraction, Transformation and Loading of data from multiple sources into Data warehouse using Confidential
- Using Crontab in UNIX and Scheduler in Windows and Jenkins for Scheduling
- Responsible to design, develop and test the software (Confidential, PL SQL, UNIX shell scripts) to maintain the data marts (Load data, Analyse using OLAP tools). Authored functional requirements documents (FRD) by interacting with development team for improving client's legacy system
- Extensive Data Warehousing experience using Confidential as ETL tool on various data bases like Oracle, SQL Server, Teradata, MS Access
- Experience working with Microsoft Outlook to set up meetings, manage calendar etc.
- Experience working with Peer for maintaining Weekly Status and Planning
Environment: Python, SQL, Tableau, Power BI, Microsoft Outlook, MS Access, MS Excel, MS Word, MS Project, MS Visio, TOAD, JAD, Teradata Test Track, Peer, Mainframe, OLAP, ERWIN, ER Studio, Clear Quest, Clear Case, Java, DB2 Database.
Confidential, Charlotte, NC
Data Analyst
Responsibilities:
- Worked with BI team in gathering the report requirements and Sqoop to export data into HDFS and Hive
- Received, cleaned, and prepped data using Python, SQL, and Excel to help data scientists build marketing mix models that resulted in a lift in ROI of 6 basis points
- Worked with client to understand business needs and translate those business needs into actionable reports in Tableau saving 16 hours of manual work each week.
- Used stored procedures, triggers, and views to provide structured data to business units by combining millions of rows of data from 19 disparate data sources
- Involved in the below phases of Analytics using R, Python and Jupyter notebook a. Data collection and treatment:
- Analysed existing internal data and external data, worked on entry errors, classification errors and defined criteria for missing values
- Used cluster analysis for identifying customer segments, Decision trees used for profitable and non-profitable customers, Market Basket Analysis used for customer purchasing behaviour and part/product association
- Developed multiple Map Reduce jobs in Java for data cleaning and pre-processing
- Assisted with data capacity planning and node forecasting
- Installed, Configured and managed Flume Infrastructure
- Administrator for Pig, Hive and HBase installing updates patches and upgrades
- Worked closely with the claims processing team to obtain patterns in filing of fraudulent claims
- Worked on performing major upgrade of cluster from CDH3u6 to CDH4.4.0
- Developed Map Reduce programs to extract and transform the data sets and results were exported back to RDBMS using Sqoop
- Have good exposure to GIT, Jenkins, JIRA
- Patterns were observed in fraudulent claims using text mining in R and Hive
- Exported the data required information to RDBMS using Sqoop to make the data available for the claims processing team to assist in processing a claim based on the data
- Developed Map Reduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW
- Adept in statistical programming languages like Rand Python including Big Data technologies like Hadoop and Hive
- Experience working as Data Engineer, Big Data Spark Developer, Front End Developer and Research Assistant
- Created tables in Hive and loaded the structured (resulted from Map Reduce jobs) data
- Using HiveQL developed many queries and extracted the required information
- Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics
- Was responsible for importing the data (mostly log files) from various sources into HDFS using Flume
- Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data
- Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems
- Managed and reviewed Hadoop log files
- Tested raw data and executed performance scripts
Environment: HDFS, PIG, HIVE, Map Reduce, Linux, HBase, Flume, Sqoop, R, VMware, Eclipse, Cloudera, Python, Excel, SQL, Power BI.