Application Developer Resume
SUMMARY
- 13+ years of professional IT work experience in Analysis, Design, Development, Deployment and Maintenance of critical software and big data applications.
- 2+ years of hands on experience across Hadoop eco systems that includes extensive experience into Big Data technologies like Spark, HDFS, YARN, Hive and Oozie.
- Having knowledge about Hadoop architecture and its different components such as HDFS, Job tracker, Task tracker, Resource Manager, Name Node, Data Node and Map Reduce concepts.
- Extensively worked on Sqoop, Hadoop, Hive, Spark for importing and exporting the data systems having various data sources, data targets and data formats.
- Built SPARK application using Python (PySpark) depending on the business requirement.
- Developed SPARK application using UDF’s, DataFrames, and Hive SQL to write data into Hive tables and retrieve back.
- Implemented Spark SQL queries and the Hive queries with the programmatic data manipulations supported by data frames in python.
- Used SVN version control tool to push and pull functions to get the updated code from repository.
- Worked with different clients and expertise in the Insurance, transportation and banking industry
- Experience in Agile Methodology and completed the scrum master certification.
- Played multiple roles - Scrum master, Tech lead, Developer, and Requirements Analyst.
- Lead daily stand-ups and scrum ceremonies for two scrum teams. Work with product owners to groom the backlog and plan sprints. Track, escalate and remove impediments. Report at daily Scrum of Scrum meetings. Track burn down, issues and progress in Version One. Work with component teams to resolve issues.
- Collaborated with members of the Product, Business and Engineering Teams to develop and maintain Product Backlogs
- Data integration expertise in developing ETL mappings using Informatica PowerCenter 9.x from disparate Data sources like flat files, WSDL, VSAM and XML files.
- Experience in Incremental Aggregation, Dynamic Lookups, Transaction Control and Constraint Based Loading, Target Load Plan, Partitioning, Persistent Caching.
- Experience in mainframe technologies with COBOL, JCL, DB2, VSAM, CICS and Web Services
- Worked in creating DB2 Stored procedures.
- Excellent ability with competitive spirit in finding various approaches for problem solving and highly skillful, creative and innovative in developing efficient logic/code.
- As a Subject matter expert, periodically communicate our operational costs and push future initiatives to produce cost-befit analysis.
- A very Good team player with ability to solve problems, organize and prioritize multiple tasks.
TECHNICAL SKILLS
Operating System: Windows, Linux, MVS, Z/OS, OS/390
Hadoop/Big Data: HDFS, MapReduce, Spark, HIVE, Sqoop, Impala, HBase, Hue
Languages: Python, Spark, SQL, JCL, COBOL, REXX
Databases: Oracle 11g, SQL Server, IBM DB2
Distributed platforms: Cloudera
ETL Tools: Informatica Power center9.x, SSIS
Tools: PyCharm, FILEAID, File manager, SPUFI, QMF, CHANGEMANN, IBM Debugger, CA Workload, ESP, SOAP, TFS&UCDP
Analytical Tool: Power Bi, Qlik
Others: MobaXterm, Putty, WinSCP, SVN, Autosys, GitHub
PROFESSIONAL EXPERIENCE
Confidential
Application Developer
Responsibilities:
- Involved in complete Bigdata flow of the application starting from data ingestion from upstream to HDFS, processing and analyzing the data in HDFS
- The Operational Data Lake (ODL) platform is used to load the single source of data across OBI functional units.
- Experienced in importing and exporting the data using Sqoop from Relational Database systems /SAP application to HDFS and vice-versa.
- Created Partitioned and Bucketed Hive tables in Parquet File Formats with Snappy compression and then loaded data into Parquet hive tables from Avro hive tables.
- Experienced in handling large datasets using Partitions, Spark in Memory capabilities, Broadcasts in Spark, Effective & efficient Joins, Transformations and other during ingestion process itself.
- Developed PySpark application using DataFrame, SPARK SQL, Broadcast, UDF’s functions and Hive SQL.
- Analyzed the SQL scripts and designed the solution to implement using Pyspark
- Developed shell scripts for running Hive scripts in Hive and Impala
- Involved Hive table creation, loading and visualizing the data using Qlik.
- Developed Hive queries to process the data and generate the data cubes for visualizing
- Developed Spark scripts by using Python and shell commands as per the requirement.
- Created Autosys workflow for PySpark application to extract and load the same in Hive partitioned table.
- Performed SQL Joins among Hive tables to get input for Spark batch process
- Provided support for QA team during the test cycle and fixed defects on the PySpark application.
Environment: Hadoop, Python, HDFS, YARN, HIVE, Sqoop, Linux shell scripting, Spark SQL, ServiceNow, Autosys.
Confidential
Application Developer
Responsibilities:
- Analyzed legacy application modules and extract logic to implement in PySpark along with new requirement provided by business.
- Developed PySpark application using DataFrame, SPARK SQL, Broadcast, UDF’s functions and Hive SQL.
- Loaded data from HDFS to hive tables using PySpark application.
- Experienced in handling large datasets using Partitions, Spark in Memory capabilities, Broadcasts in Spark, Effective & efficient Joins, Transformations and other during ingestion process itself.
- Analyzed the SQL scripts and designed the solution to implement using Pyspark
- Writing the HIVE queries to extract the data processed.
- Worked on POC to load various structured files coming from mainframe to hive tables using single PySpark program.
- Involved Hive table creation, loading and visualizing the data using Power Bi.
- Developed Spark scripts by using Python and shell commands as per the requirement.
- Created Oozie workflow for PySpark application to extract and load the same in Hive partitioned table.
- Involved in utilizing Spark API over Hadoop YARN as execution motor for information investigation utilizing Hive and presented the information to BI group for producing reports, after the handling and examining of information in Spark SQL
- Performed SQL Joins among Hive tables to get input for Spark batch process
- Developed oozie workflow for scheduling & orchestrating the ETL process.
- Reviewed the PySpark code developed by offshore team.
- Provided support for QA team during the test cycle and fixed defects on the PySpark application.
- Created mainframe programs and jobs to place data in Linux for HDFS data movement.
- Trained offshore team to develop PySpark applications and Hive SQL’s to retrieve and manipulate the same.
- Manage multiple projects parallel with offshore on both Big Data and ETL applications
- Involved in everyday SCRUM gatherings to examine the advancement/advance and was dynamic in making
- Scrum gatherings progressively gainful.
Environment: Python, Spark SQL, Hadoop YARN, Spark Streaming, Hive, Hue, Oozie, Maestro, HDFS, Zeppelin, SSIS, SQL Server and UNIX
Confidential
Application Developer
Responsibilities:
- Being Advisory Subject Matter Expert, Interacted with Clients directly for requirement gathering and translating the same to High Level Design.
- Translated customer requirements into formal requirements and design documents, establish specific solutions, and lead the efforts including programming and testing that culminate in client acceptance of the results. Delivered new and complex high-quality solutions to clients in response to varying business requirements.
- Worked as Tech Lead to offshore and onsite Developers on the Requirements, Design and Development of CSB application.
- Lead daily stand-ups and scrum calls. Work with product owners to groom the backlog and plan sprints. Track, escalate and remove impediments. Report at daily Scrum of Scrum meetings. Track burn down, issues and progress in Version One. Work with component teams to resolve issues.
- Communicated the progress to senior management thru 'Burndown Charts'. Monitored the Quality thru metrics and mentored team thru the project management processes.
- Provided technical leadership to production support team.
- Worked in creating DB2 Stored procedures.
- Worked on Multiload, Fastload, Fast Export, TPT utilities to Load and unload data to and from the Teradata database.
- Used Informatica client tools such as Informatica Power Center, Workflow Manager, and Workflow Monitor etc. The objective is to extract data from source systems like Oracle, flat files, salesforce application into a staging database and then load into single data warehouse.
- Developed standard and reusable mappings and mapplets using various transformations like Expression, Sorter, Aggregator, Joiner, Router, Lookup (Connected and Unconnected), Filter, Update Strategy, Sequence Generator, Normalizer and Rank.
- Experience in creating Reusable Tasks (Sessions, Command, Email) and Non-Reusable Tasks (Decision, Event Wait, Event Raise, Timer, Assignment, Control).
- Developed shell scripts for Daily and monthly Loads and scheduled jobs using UNIX and ESP.
- Transitioned changes to Run/Support team for each Production release and provide initial problem troubleshooting and resolution for Production Failures.
- Performing the code migration in to the Release packages thru Harvest, Changeman and running the Package level audits.
- Reduce the elapsed processing times and meet required SLAs.
- Monitoring the Jobs and Schedule them using CA-7 and ESP
- Reviewed deliverables of all phases in SDLC i.e. Design document, Code Review, Unit & System test results and Deployment plans.
- Responsible for Estimation & Tracking of the Maintenance requests.
- Responsible for Team building, mentoring and coaching new team members to make them productive in short span of time.
Environment: Informatica 9.5, Teradata, Oracle 11g, UNIX, ESP
Confidential
Application Developer
Responsibilities:
- As a Team Lead and Analyst played a key role in various phases of the project, viz., Estimations, Enhancements and Implementations using Agile project management methodology.
- Coordination between offshore and Client Project Team, Project manager. Providing Technical and Business guidance to Offshore team members.
- Work closely with Business Analysts in Gathering requirements, analyzing feasibility and timelines.
- Assign tasks to offshore resources and help them in their analysis
- As a Senior Developer coded many programs in CICS, Websphere MQ, and DB2 Stored Procs.
- Develop/modify existing COBOL applications to send messages to the presentation layer (JAVA) using Message Queuing and MQI calls.
- Analyzed, designed and developed JAVA-COBOL-CICS-IMS applications for Reengineering and SOA adoption.
- Written application programs to transfer user data from SHIPCSX to Mainframe systems using COBOL/MQ Series for the completion of user request.
- Experience in setting up NDM - Connect Direct connection, FTP and SFTP mailbox for file transmission between CSX and outside vendors.
- Participated in application testing for newly created COBOL/MQ series modules.
- Create, Modify, Enhance Jobs (JCL’s) and Proc’s as part of the Integration.
- Designed complex mappings for date chaining and deduping before loading the data into target database.
- Performing the code migration into the Release packages thru Changeman and running the Package level audits
Environment: Z/OS, COBOL, CICS, DB2, UNIX, SPUFI, XML, JCL, VSAM, TSO/ISPF, File-Manager, FTP, Endevor, CA-7, File-Aid, IBM Debugger, MQ series, Interest and Xpeditor