Software Developer Resume
Bellevue, WA
SUMMARY:
- 8 + years of experience in all phases of software development, software architecture.
- Multiple years of Big Data Experience that includes Hadoop, HDFS, Hive, Pig, Oozie, Sqoop, HBase, HCatalog, Flume, Cassandra, Sqoop and any other related technologies.
- Expertise in deriving big data patterns out of big data problems and designing solutions.
- Experience in Data Visualization tools like Datameer, Tibco Spotfire and Tableau.
- Proficient in Hive QL and SQL scripting and Query optimizations.
- Experience in developing Hive Generic UDFs, UDAFs, UDTFs.
- In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, YARN, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce concepts.
- Database Experience includes Oracle, PostgresSQL, MySQL, Microsoft SQL server.
- Proficient in Data warehousing concepts and Methodologies, Data Modeling Concepts: Dimensional Data Modeling, Fact Table, Dimension Table, OLTP, OLAP, Star schema, Snow Flake schema.
- Extensively used Slowly Changing Dimension (SCD) technique for incremental updating of the targets.
- Experience in Data modelling tools like SAP power designer and Erwin.
- Proficient in Data Modelling techniques like Physical Modeling, Logical Modeling, Conceptual Modeling, Dimensional Modeling (Star Schema, Snow - Flake, FACT, Dimensions), Entities, Attributes, Cardinality, ER Diagrams.
- Experience with clickstream data includes Omniture.
- Development involvement in Web Front end Technologies., viz HTML, CSS, Tomcat, Apache.
- Proficient in C and scripting languages (Perl, shell).
- Expertise in ETL tools Informtica Power Center and IBM Datastage.
- Extensive Experience in Oracle Data warehouse/Informatica/Siebel Analytics/OBIEE, ETL Processes and Database Deliverables in various Projects.
- Experienced in OS environments Unix, Linux, and Win32.
- Experience in scheduler tools like DAC, CONTROLM, TIDAL.
- Extensively used Slowly Changing Dimension (SCD) technique for incremental updating of the targets.
- Extensively worked with Parameter files, Pipeline Partitioning, Pushdown Optimization, and Informatica Scheduler.
- Strong experience in Error handling, Performance tuning and Optimizing- the source, mappings, transformations, sessions, and the targets, and expertise using Informatica Debugger.
- Solid experience in writing SQL queries, PL/SQL, Stored Procedures, and optimization.
- Experience in Unix Shell Scripting for scheduling/running ETL jobs in production environment.
- Hands on experience in DWH concepts, Data Modeling and Data Profiling.
- Familiar with Project Management and SCM tools including Git, CVS, SVN, Rational Suite, MS Project, Visio, Omni Graffle.
- Familiar with wide range of development Methodologies Waterfall, Agile, Scrum, XP.
- Demonstrated proven Problem-solving abilities.
TECHNICAL SKILLS:
Hadoop Technologies and Distributions: Apache Hadoop, Horton works Data Platform (HDP)
Hadoop Ecosystem: HDFS, Map: Reduce, Hive, Pig, Sqoop, Oozie, Flume, Zookeeper
NoSql Databases: Cassandra, HBase
Programming: Java, PL/SQL
RDBMS: ORACLE, MySQL, SQL Server,Teradata
Operating Systems: Linux (RedHat, CentOS), Windows XP/7/8
Web Servers: Apache Tomcat
ETL: Informatica,Datameer
Cluster Management Tools: Ambari
BI Tools: Tableau,OBIEE,Spotfire
PROFESSIONAL EXPERIENCE:
Confidential, Bellevue, WA
Software DeveloperResponsibilities:
- Worked on Data Movement Frame work to land the data from Different sources systems to HDFS.
- Created Hive tables and written the Hive scripts to extract the data.
- Developed Generic UDF, UDAFs for data in hive windowing functions.
- Created Pig Transformations to transform and load the data to Data Lake.
- Implemented SCD1 and SCD2 loading strategies on HDFS and HBase.
- Designed Hbase schema with row distribution by date and consumer
- Implemented a very intelligent ingestion framework from various data sources. Both bulk loads and event loads exists
- Created Oozie workflows and Scoop jobs to import the data from different RDBMS sources to HDFS Data Lake.
- Created Dispatch scripts to dispatch the data from HDFS to Teradata data.
- Created Logical and Physical data models for Hive using SAP Power designer.
Environment: Hive, Pig, Hadoop, Hbase, Oozie, Sqoop, SAP Power designer, Informatica v9.1.5. Informatica PowerCenter, Windows 7, Teradata, UNIX B2B, PL/SQL, XML,SQL Server 2008,Shell Scripting
Confidential, New Jersey
Software DeveloperResponsibilities:
- Extracted the data from raw files and separated for processing.
- Actively involved in design analysis, coding and strategy development.
- Trouble-shoot and triage production issues.
- Presentedknowledgesharingsessionsongeneralarchitecturetoother team members.
- Coded Custom Mapreduce depending on the requirements.
- Developed HIVE and PIG UDFs.
- Log analysis and ETL processing.
- Implemented daily workflow for extraction, processing and analysis of data with Oozie.
- Gatheredperformancemetricsonmonthlyprojectsandreportedto upper management
Environment: Hadoop,Hive,Pig, Python, Tableu, Informatica v9.0.1. Informatica PowerCenter, Windows 7, Oracle 11g, UNIX, Toad, B2B, PL/SQL, XML,SQL Server 2008,Shell Scripting
Confidential, Alpharetta, GA
Software DeveloperResponsibilities:
- Migrated data from various data sources to Hadoop Data lake using Data Meer
- Worked on Hadoop cluster which ranged from 60-80 nodes during pre-production stage and it was sometimes extended up to 120 nodes during production.
- Used Sqoop to import the data from RDBMS to Hadoop Distributed File System (HDFS) and later analysed the imported data using Hadoop Components
- Did various performance optimizations like using distributed cache for small datasets, Partition, Bucketing in hive and Map Side join’s.
- Involved in creating Hive tables, then applied HiveQL on those tables for data validation.
- Involved in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries and Pig scripts
- Created Visualizations on data in HDFS using Tableau and Spot fire
- Developed Slowly Changing Dimension for Type 1 SCD
- Developed common routine mappings. Made use of mapping variables, mapping parameters and variable functions.
- Used Informatica Designer to create complex mappings using different transformations like Filter, Router, Connected & Unconnected lookups, Stored Procedure, Joiner, Update Strategy, Expressions and Aggregator transformations to pipeline data to Data Mart.
- Assisted in designing Logical/Physical Data Models, forward/reverse engineering using Erwin 4.0.
- Conducted Database testing to check Constrains, field size, Indexes, Stored Procedures, etc.
- Extensively used OBIEE Answers and rpd Design design, ad-hoc reporting and advanced report authoring
- Developed Complex Reports in OBIEE
- Created Subject areas in OBIEE for users to be Analyze specific business function
- Developed OLAP solutions using OBIEE and published Reports using BI Publisher
- Creating prompts, conditions, filters and calculations in OBIEE
- Creating multilingual reports in OBIEE
- Created Tasks, Subject areas and Execution Plans in DAC to schedule ant automate the ETL daily, weekly and Monthly Loads
- Defined the KPI(Key Performance indicators) for the business populated them in Warehouse and exposed them in the reporting layer.
Environment: Hadoop,Hive,Pig,Datameer, Tibco Spotfire Informatica v9.0.1. Informatica Power Center, Windows 7, Oracle 11g, OBIEE 11g ( 11.1.1.0.6 ),Oracle Business Intelligence Applications 7.9.6.3, Data Warehouse Administration Console(DAC) 11.1.1.0.6 , UNIX, Toad, B2B, PL/SQL, XML,SQL Server 2008,Shell Scripting
Confidential
Software DeveloperResponsibilities:
- Designed and Developed the Informatica workflows/sessions to extract, transform and load the data into Target.
- Wrote SQL, PL/SQL codes and stored procedures for dropping and recreating indexes, to generate oracle sequences, procedures that update the Process Run table automatically to make the delta process work accurately.
- Extensively worked with Unix Shell scripting to validate and verify the data in the flat files generated by the ETL process.
- Developed the UNIX shell scripts to send out an E-mail on success of the process indicating the destination folder where the files are available.
- Involved in the migration of OBIEE 10g to OBIEE 11.1.1.6.2.
- Implemented Object Level Security for the objects like Dashboards and Reports, and Data Level Security for Region and Product dimensions, using Session Variables.
- Developed Performance Tuning Strategies for optimizing the performance.
- Performed unit testing and system testing of all components ofOBIEE Analytics.
- Performed unit testing and system testing of all components ofOBIEE Analytics.
- Developed different levels of Security set-up for the entities of the application (on both Repository file and Web catalog file) using OBIEE 11g.
- Created an initialization block to retrieve allowable values.
- Configured the BI Delivers scheduler to schedule reports.
Environment: Informatica v9.0.1. Informatica Power Center, Windows 7, Oracle 11g, OBIEE 11g ( 11.1.1.0.6 ),Oracle Business Intelligence Applications 7.9.6.3, Data Warehouse Administration Console(DAC) 11.1.1.0.6 , UNIX, Toad, B2B, PL/SQL, XML,SQL Server 2008,Shell Scripting,Tibco Spotfire, Datameer
Confidential
Software DeveloperResponsibilities:
- Involved in ETL process from development to testing and production environments.
- Responsible for creating interfaces using different types of mappings using various transformations like Expression, filter, router, Aggregator, Look up, Joiner, Stored Procedure, Update Strategy, Etc.
- Extensively worked on Mapping Variables, Mapping Parameters, Workflow Variables and Session Parameters for the delta process to extract only the additional data added during that period.
- Extensively Used debugger in identifying bugs in existing mappings by analyzing data flow, evaluating transformations.
- Worked with mappings to dynamically generate parameter files used by other mappings.
- Involved in performance tuning of the ETL process by addressing various performance issues at the extraction and transformation stages.
- Extensive performance tuning by determining bottlenecks at various points like targets, sources, mappings and sessions.
- Documented the mappings used in ETL processes including the Unit testing and Technical document of the mappings for future .
Environment: Informatica 8.6.1, Informatica Architecture, Informatica Power Center 8.6.1, DB2, Oracle 11g/, Data Transformation, UNIX, Toad, PL/SQL Shell Scripting
Confidential, Alpharetta, GA
Software DeveloperResponsibilities:
- Assisted in designing Logical/Physical Data Models, forward/reverse engineering using Erwin 4.0.
- Developed common routine mappings. Made use of mapping variables, mapping parameters and variable functions.
- Used Informatica Designer to create complex mappings using different transformations like Filter, Router, Connected & Unconnected lookups, Stored Procedure, Joiner, Update Strategy, Expressions and Aggregator transformations to pipeline data to Data Mart.
- Developed Slowly Changing Dimension for Type 1 SCD
- Used mapplets for use in mappings thereby saving valuable design time and effort
- Used Informatica Workflow Manager to create, schedule, and run sessions, worklets and workflows
Environment: Informatica v 7.1.3. Informatica Power Center, Oracle 10g, OBIEE 10g (10.1.3.4.1 ),Data Warehouse Administration Console(DAC), UNIX, Toad, PL/SQL, Shell Scripting