Senior Big Data Engineer Resume
2.00/5 (Submit Your Rating)
Seattle, WA
SUMMARY
- Building phenomenal Enterprise Data Services by creating new DataLake/Pipeline/Compute capabilities.
- Has been migrating Enterprise Data Warehouse to AWS, and performed Enterprise Data Warehouse transformation to the Data Lake/Snowflake/Looker technology platform.
- Greater emphasis on cloud computing, storage and exporting of big data.
- Performing all duties with dedication and diligence. Ability to use state - of-the-art technologies and techniques for key strengths such as developing, documenting and delivering quality software on independent platforms and languages.
- With a strong educational background and solid work experiences
TECHNICAL SKILLS
- Hadoop 2.7.1, Ambari 2.7.0.0, Spark 2(sparkSql,cli,hive), Apache Tez,Livy Server, Hive(Hue: beewax,beeline; hivecli,beelinecli,UDF), Zookeeper,kafka, zeppelin,Hbase, AWS(DynamoDb,EC2, EMR,RDS, S3,DistCp, Lambda, Elastic Search, S3-Dist-Cp,Cloud Watch, EC2 spot fleet, S3 snopshots), Machine Learning, Data Visualization,Platfora, pentahoo
- Kafka,Sqoop, Flume
- SQlwork bench, DBeaver,MysqWorkbench, CloudBerry Explorer for Amazon S3
- Eclipse,R studio, Notepad++, SPSS, Weka, Genie, NetBeans
- Linux, Centos, Windows
- Java, Scala,Shell, Python, R, SQL,HQL
- Abstract Factory, Factory Method, Composite, Facade, Template Method, MVC, and Singleton mvn, ant,sbt
- MRunit, Mokito, HiveRunner, Beetest, Hive test
- There are no automation tools or frameworks available for Flume, sqoop and oozie unit testing yet. Tested manually.
PROFESSIONAL EXPERIENCE
Confidential, Seattle, WA
Senior Big Data Engineer
Responsibilities:
- Building, maintaining and testing Big Data Ecosystem with its pipelines: HDP, CDH, Spark…
- Data Migration or Ingestion: Ensuring regular data backups, exports, and imports from various sources: SQL, MYSQL, PostgreSQL, salesforce, google analytics, facebook,..
- Tuning for Performance: Assessing as well as implementing policies for maximizing the performance of Big Data availability
- Analyzing Big data technology for improving efficiency integration
- Data Science: carrying out statistic monitoring as well as analysis of trends and reporting
- Monitoring and configuration of Data Lake in cloud, on-premises and different API
- Programming: Java/Scala like marketing API(facebook,DCM,) UDF,UDAF
- Create and manage data flows, metadata, and routines
- Increasing collaboration with a shared repository and continuous delivery
- Manage deployment, execution and monitoring without switching tools:Docker, kubernates, terraform, Artififactory
Confidential
Big Data Developer
Responsibilities:
- Tasked with refining and visualizing Hadoop cluster Log Data for Security (in situation of suspected security breach, how can server log data be used to identify and repair the vulnerability?) and Compliance (for system audits).
- Used Pig and hive, on top of Hadoop fs for processing. Implemented Hadoop Map Reduce Jobs with Java for cleaning and monitoring of metadata.
- Analyze Hadoop log files, support, maintenance and ongoing monitoring
Confidential
Java Developer
Responsibilities:
- Developed efficient user interface for desktop application.
- Ensured reliable and secure data transactions by implementing Windows Communication Foundation (WCF) Framework which provides efficient services oriented application (SOA).
- Optimized system performance. The System helps employees manage and review their work performances (tasks) easily in a day to day basis. I performed secure IM with different endpoints by using services hosted by application.