CakeResume Talent Search

Advanced filters
On
4-6 years
6-10 years
10-15 years
More than 15 years
Avatar of the user.
Avatar of the user.
Data Engineer @Groundhog Technologies Inc.
2021 ~ 2024
Data Analyst、Data Engineer、Data Scientist、Customer Experience Analyst
Within one month
Git
Python
Scala
Employed
Ready to interview
Full-time / Interested in working remotely
4-6 years
University of Illinois at Urbana-Champaign, School of Information Sciences
Information Management
Avatar of the user.
Avatar of the user.
Past
Senior Data Analyst @趨勢科技
2022 ~ Present
Data Scientist, Data Analyst, Machine Learning Engineer
Within one month
python
R
SQL
Unemployed
Ready to interview
Full-time / Interested in working remotely
4-6 years
輔仁大學 Fu Jen Catholic University
統計資訊學系
Avatar of the user.
Avatar of the user.
Data Engineer @Tesla
2023 ~ 2023
Data engineer / Data anyayst
Within two months
python
Linux
R
Employed
Ready to interview
Full-time / Interested in working remotely
4-6 years
University of Texas at Dallas
Information Technology and Management
Avatar of Yiru Chen.
Avatar of Yiru Chen.
資料工程師 @鴻海(富士康)
2021 ~ Present
資料工程師/網路爬蟲、後端工程師、軟體工程師、網管
Within one month
Uvicorn, Gunicorn 開發RESTful API 及設計 MongoDB schema並使用 Docker-compose 進行開發與部署- 負責開發資料庫 Connector 操作頁面(支援 Impala, Kudu, PostgreSQL, MySQL, MS SQL, Oracle)、 Record Controller 功能、平台 Pipeline 中檢視資料的視覺化功能,以及熟悉 Nifi 各種Processor 功能加入系統中 - 協助用戶建置及改善 ETL Pipeline 效能 Line Notify 服務系統: - 設計系統
Python
SQL
Git
Employed
Ready to interview
Full-time / Interested in working remotely
4-6 years
國立成功大學
統計研究所
Avatar of 陳昭儒.
Avatar of 陳昭儒.
Past
Data Engineer @BUBBLEYE | We're hiring!
2021 ~ 2022
Software Enginer
Within two months
size:GB Number of rows: 6,268,519,176 Qudowe Project Lead & Software Engineer Product of Pixnet Travel Hackathon 2019, a trip planner based on Instagram's data Work Experience Vpon, Data Engineer Aug 2018 ~ Oct 2020 Implement Akka-http(Scala) server endpoints for Vpon Data Platform Product Create new ETL pipelines using GCP Spark(Apache Spark) and GCP Dataflow(Apache Beam) to batch input/output hundreds of files Migrate existing ETL pipelines from AWS(Hive SQL) to GCP(BigQuery SQL) using python Migrate datawarehouse from AWS(Hive) to GCP(BigQuery) Setup Prometheus on GKE (Google Kubernetes Engine) to
Python
ETL
Web Scraping
Unemployed
Ready to interview
Full-time / Interested in working remotely
4-6 years
National Taiwan University
電機工程學系
Avatar of 施冠宇.
Avatar of 施冠宇.
Data engineer @H2 Inc.
2021 ~ Present
AI engineer, ML engineer, data scientist
Within three months
助公司 app 增加 6 % 用戶參與度並獲得 70 % 用戶滿意度 - 運用 DVC 建置 model experiment tracking, dataset versioning 以及 model management 【Airflow 定期排程以及維運 data reports】 - 創建 2 個 ETL pipeline - 幫助 3 個不同部門維護 8 個 data reports - 利用 python 去處理並且整合 SQL, Google big query, AWS S3 bucket 不同來源的 data source 【Data analysis】 - 根據各部門所需提
Airflow
Docker
AWS
Employed
Open to opportunities
Full-time / Interested in working remotely
4-6 years
清華大學
動力機械工程學系
Avatar of Emily Ledoux.
Avatar of Emily Ledoux.
Principal @Cascade Data Labs
2016 ~ 2022
Director Data
Within two months
Emily Ledoux Delivery Principal Seasoned Delivery Principal in the Data Practice. Focused on designing robust, scalable data ecosystems in the cloud to feed insights and data visualizations. Well-rounded consultant with experience spanning sales, recruiting, and delivery. Proven Delivery & Client Lead. Portland, OR, USA https://www.linkedin.com/in/emily-ledoux/ Work Experience JanuaryPresent Principal Data Architect Kin + Carta Delivery or Client Lead for over 25 resources, including direct reports, delivery oversight, hours tracking, QBRs, onboarding management, budget ownership and related responsibilities. Cloud Architect, designing Azure and
PowerPoint
Word
Excel
Employed
Open to opportunities
Full-time / Interested in working remotely
6-10 years
University of Pennsylvania
Economics
Avatar of Anmol Ganju.
Avatar of Anmol Ganju.
Senior Data Engineer @Larsen Toubro Infotech Mindtree
2021 ~ Present
Senior Cloud Consultant
Within three months
load for a particular site level. Have been using Data lake, Azure SQL Database, Azure Databricks, Azure Synapse Analytics as intermediate layer for data storage and end products involving Facts and Dimensions that can be further consumed by Azure Analysis Services. Worked on writing SQL Logic that involves pipeline re-run process so that pipeline can be re-run from the same watermark point as it was failed to avoid data loss. Also worked on Logic App to ingest the data from SQL Database into Sharepoint List on the basis of Row Active flag. Cloud Consultant
Time Management Skills
Team Player
Analysis
Employed
Open to opportunities
Full-time / Interested in working remotely
6-10 years
Dehradun Institute Of Technology
B.E Information Technology
Avatar of 蕭仕廷.
Avatar of 蕭仕廷.
Architect Lead @iKala 愛卡拉互動媒體股份有限公司
2023 ~ Present
軟體工程師
Within one month
Chart, Terraform, ArgoCD, Gitlab CI - Workflow: Argo Workflow Key Roles and Achievements: - iKala Interactive Media Inc.: - Served as an Architect, designing cross-team interfaces and leading AI department design documentation efforts. - As a Senior Backend Engineer, refactored crawler services, developed the Datahub system, and led ML Inference and ETL pipeline integration projects. - Lintor International Co., Ltd.: - Led the design and implementation of 33 games, demonstrating leadership in game logic system development. - Developed a Golang server framework and implemented microservices to enhance system functionality. - CHUNG YO INTERNET INFORMATION CO., LTD.: - Maintained platform
c#
JAVA
MSSQL
Employed
Not open to opportunities
Full-time / Interested in working remotely
4-6 years
逢甲大學
資訊工程學系, 資訊安全學位
Avatar of 呂學炘.
Avatar of 呂學炘.
Data Science @Alfred Labs.
2017 ~ Present
Within two months
Yoyo Lu, Data Analyst professional [email protected] Resourceful Data Analyst and Programmer with 5+ years of experience at Alfred Labs. Strong background in Mathematics and Computer Science. Adept at providing crucial analysis to cooperating teams. WORK EXPERIENCE Alfred Labs., Data Science, Jul 2017 ~ Present As an early employee of the company, I played a pivotal role in the development of our data infrastructure. This included the creation of robust ETL pipelines and automated dashboards, key advancements that enabled the entire organization to make data-driven decisions. My expertise in designing
Python
R
SQL
4-6 years
University of Taipei
Computer Science

The Most Lightweight and Effective Recruiting Plan

Search resumes and take the initiative to contact job applicants for higher recruiting efficiency. The Choice of Hundreds of Companies.

  • Browse all search results
  • Unlimited access to start new conversations
  • Resumes accessible for only paid companies
  • View users’ email address & phone numbers
Search Tips
1
Search a precise keyword combination
senior backend php
If the number of the search result is not enough, you can remove the less important keywords
2
Use quotes to search for an exact phrase
"business development"
3
Use the minus sign to eliminate results containing certain words
UI designer -UX
Only public resumes are available with the free plan.
Upgrade to an advanced plan to view all search results including tens of thousands of resumes exclusive on CakeResume.

Definition of Reputation Credits

Technical Skills
Specialized knowledge and expertise within the profession (e.g. familiar with SEO and use of related tools).
Problem-Solving
Ability to identify, analyze, and prepare solutions to problems.
Adaptability
Ability to navigate unexpected situations; and keep up with shifting priorities, projects, clients, and technology.
Communication
Ability to convey information effectively and is willing to give and receive feedback.
Time Management
Ability to prioritize tasks based on importance; and have them completed within the assigned timeline.
Teamwork
Ability to work cooperatively, communicate effectively, and anticipate each other's demands, resulting in coordinated collective action.
Leadership
Ability to coach, guide, and inspire a team to achieve a shared goal or outcome effectively.
Within two months
Sr. Data Engineer / Team Lead
Logo of 新加坡商競舞電競娛樂有限公司 Garena Online Private Ltd.
新加坡商競舞電競娛樂有限公司 Garena Online Private Ltd
2021 ~ Present
Taoyuan, 桃園區桃園市台灣
Professional Background
Current status
Employed
Job Search Progress
Professions
Other
Fields of Employment
Work experience
6-10 years
Management
I've had experience in managing 1-5 people
Skills
Hadoop
Spark
SQL
python
Scala
AWS
GCP
Languages
Chinese
Native or Bilingual
English
Fluent
Job search preferences
Positions
資料工程師
Job types
Full-time
Locations
Taipei, 台灣
Remote
Interested in working remotely
Freelance
Educations
School
Major
Print
Vxftskax4mfwtiqudlao

Yi-Lun Wu (Velen)

Summary

  • 7+ years of experience in big data fields both Cloud(GCP, AWS) and on-premised (Cloudera CDH). 
  • Develop data catalog with hybrid cloud environment(on-perm+AWS) for global commerce at Yahoo.
  • Lead a machine learning team to build up offline/real-time platforms for recommendation systems from scratch at Garena. 
  • Lead and architect large-scale data pipelines/warehouses from Innova Solutions(AWS) and 17LIVE (GCP). And Cooperate with data science/machine learning team/TW HQ data team.
  • Expert of Hadoop ecosystems such as HDFS/Hive/Hbase/Spark in Athemaster and Xuenn.

Sr. Big Data Engineer / Team Lead
Taoyuan,TW
[email protected]

Skills


Languages

  • Python 
  • SQL
  • Linux Shell Script
  • Scala


Big Data Solutions

  • Cloudera CDH
  • Hadoop echosystems
  • AWS EMR
  • GCP Dataporc


Data Warehouse

  • Hive
  • Google BigQuery
  • MySQL
  • HBase
  • Clickhouse
  • AWS RDS
  • AWS Athena


ETL Skills in Big Data

  • Spark / Spark Streaming
  • Hive (for ELT) 
  • Impala 
  • Kafka 
  • Cloud DataFlow  (GCP)


Workflow Skills

  • Airflow 
  • Digdag 
  • NiFi 
  • AWS CloudFormation


Other Skills

  • Great communication 
  • Leaderships 
  • Scrum 
  • JIRA 
  • Linux 
  • Git  

Experience

Senior Engineer at Yahoo.Jan, 2023 - Present

Working at Yahoo. The greatest challenge lies in developing in response to rapid market changes, where the data catalog needs to integrate with various complex systems. Simultaneously, maintaining the highest stability and ensuring high-quality data is essential. 

More responsibilities/details as below.
  • Fetch providers data with various way via Java. Such as fetching data from client's API, GraphQL, FTP, S3 or GCP... etc.
  • standardizing data from above to feed into data warehouse in Hadoop/Hive/HBase using Spark
  • Implement a checking system to guarantee high quality data via Java.
  • Migrate part of services from on-perm(Hadoop) to cloud(AWS) to become a hybrid cloud environment.

 

Senior Data Engineer/ Team Lead, at BOOYAH! Live Garena.Oct, 2021 - Sep, 2022

Working at ML team as a first data engineer. The challenge include build up data warehouse/pipeline from scratch. And design the data flow to support both batch/real-time recommendation systems.

More responsibilities/details as below.
  • Design data model from scratch and Manage Hadoop based data warehouse for the training system.
  • Develop streaming ETL pipeline via Spark from Message queue(Kafka) into in-memory data structure store(Redis) and ClickHouse for real-time recommendation system.
  • Design ELT job for offline report system in Hadoop/Hive using Spark.
  • Build up monitoring dashboard on Grafana
  • Take leadership on TW side.

Senior Data Engineer at 17 Media.Jun, 2020 - Oct, 2021

The big challenge of 17 Media data teams is facing fast-growing data volume (processing 5-10x TB level daily), complex cooperation with stakeholders, the cost optimization of pipeline and refactor big latency systems .etc.

More responsibilities/details as below. 
  • Manage Google BigQuery based data warehouse/lake. 
  • Refactory architecture of data warehouse to enhance 2x performance
  • Develop batch/streaming ETL pipeline to process data from diverse data sources(e.g. MongoDB, MySQL, APIs) into GCP
  • Design workflow using Digdag
  • Implement CI/CD on BIgQuery
  • Build up visualization tool(Superset) via Kubernetes
  • Well leadership and guiding junior members.

Senior Software Engineer at Innova Solutions Ltd.Oct, 2018 - May, 2020

Development Intelligent Healthcare Data Platform(IHDP) for empowering compony solutions using AWS service. 

More responsibilities/details as below. 
  • Build APIs to the processing of patient records and providing access for downstream usages.
  • Build Infrastructure on AWS and compliance for HIPPA and GDPR standard.

IT Consultant at Xuenn Pte Ltd.May, 2018 - Sep, 2018

The biggest challenge in Xuenn is facing performance issues in the original data warehouse. And I lead a project to build up a Hadoop cluster to reduce original EDW loading and improve various data pipelines.

More responsibilities/details as below.  
  • Perform adopting new technologies to implant into, to fuse into or to replace with existing systems for gaining leaps in performance, benefits and capabilities of the users. 
  • Building-up multiple systems and integrating with existing system, implemented Hadoop, data mining or data warehouse systems.
  • Design an architecture which processing real-time data end-to-end using various Hadoop solutions without coding. 

Software Engineer at Athemaster Co., Ltd.Jan, 2016 - Apr, 2018 

Athemaster is a technology company offering solutions and expertise in implementing Enterprise Data Hub and automating Data integration with Open Source technologies such as Apache Hadoop and Spark. More responsibilities/details as below.
  • Focus on Enterprise Big Data solution such as Hadoop and Spark (Cloudera CDH). 
  • Maintain and improved other companies' Hadoop cluster. 
  • Help other companies integrate with Hadoop and resolved the technical issues. 
  • Build data pipelines via python ETL data to Hadoop.

Certification and License


Readings 00 00@2x ce5676dabcce042724a6fc4c3413d6a86ad9c78eecb848896433e32c60b7006b


CCA-175: CCA Spark and Hadoop Developer



Readings 00 01@2x 77cc06c91fae4dd43a069fa4b813524cd022d4a79115524d3f0d6b9220dfd71d

Cloudera Certified Administrator for Hadoop

Education



Undergraduate studies at Tamkang University, with a concentration in Department of Management Sciences.

                                                                                                                                                                                                     2008 - 2012

Resume
Profile
Vxftskax4mfwtiqudlao

Yi-Lun Wu (Velen)

Summary

  • 7+ years of experience in big data fields both Cloud(GCP, AWS) and on-premised (Cloudera CDH). 
  • Develop data catalog with hybrid cloud environment(on-perm+AWS) for global commerce at Yahoo.
  • Lead a machine learning team to build up offline/real-time platforms for recommendation systems from scratch at Garena. 
  • Lead and architect large-scale data pipelines/warehouses from Innova Solutions(AWS) and 17LIVE (GCP). And Cooperate with data science/machine learning team/TW HQ data team.
  • Expert of Hadoop ecosystems such as HDFS/Hive/Hbase/Spark in Athemaster and Xuenn.

Sr. Big Data Engineer / Team Lead
Taoyuan,TW
[email protected]

Skills


Languages

  • Python 
  • SQL
  • Linux Shell Script
  • Scala


Big Data Solutions

  • Cloudera CDH
  • Hadoop echosystems
  • AWS EMR
  • GCP Dataporc


Data Warehouse

  • Hive
  • Google BigQuery
  • MySQL
  • HBase
  • Clickhouse
  • AWS RDS
  • AWS Athena


ETL Skills in Big Data

  • Spark / Spark Streaming
  • Hive (for ELT) 
  • Impala 
  • Kafka 
  • Cloud DataFlow  (GCP)


Workflow Skills

  • Airflow 
  • Digdag 
  • NiFi 
  • AWS CloudFormation


Other Skills

  • Great communication 
  • Leaderships 
  • Scrum 
  • JIRA 
  • Linux 
  • Git  

Experience

Senior Engineer at Yahoo.Jan, 2023 - Present

Working at Yahoo. The greatest challenge lies in developing in response to rapid market changes, where the data catalog needs to integrate with various complex systems. Simultaneously, maintaining the highest stability and ensuring high-quality data is essential. 

More responsibilities/details as below.
  • Fetch providers data with various way via Java. Such as fetching data from client's API, GraphQL, FTP, S3 or GCP... etc.
  • standardizing data from above to feed into data warehouse in Hadoop/Hive/HBase using Spark
  • Implement a checking system to guarantee high quality data via Java.
  • Migrate part of services from on-perm(Hadoop) to cloud(AWS) to become a hybrid cloud environment.

 

Senior Data Engineer/ Team Lead, at BOOYAH! Live Garena.Oct, 2021 - Sep, 2022

Working at ML team as a first data engineer. The challenge include build up data warehouse/pipeline from scratch. And design the data flow to support both batch/real-time recommendation systems.

More responsibilities/details as below.
  • Design data model from scratch and Manage Hadoop based data warehouse for the training system.
  • Develop streaming ETL pipeline via Spark from Message queue(Kafka) into in-memory data structure store(Redis) and ClickHouse for real-time recommendation system.
  • Design ELT job for offline report system in Hadoop/Hive using Spark.
  • Build up monitoring dashboard on Grafana
  • Take leadership on TW side.

Senior Data Engineer at 17 Media.Jun, 2020 - Oct, 2021

The big challenge of 17 Media data teams is facing fast-growing data volume (processing 5-10x TB level daily), complex cooperation with stakeholders, the cost optimization of pipeline and refactor big latency systems .etc.

More responsibilities/details as below. 
  • Manage Google BigQuery based data warehouse/lake. 
  • Refactory architecture of data warehouse to enhance 2x performance
  • Develop batch/streaming ETL pipeline to process data from diverse data sources(e.g. MongoDB, MySQL, APIs) into GCP
  • Design workflow using Digdag
  • Implement CI/CD on BIgQuery
  • Build up visualization tool(Superset) via Kubernetes
  • Well leadership and guiding junior members.

Senior Software Engineer at Innova Solutions Ltd.Oct, 2018 - May, 2020

Development Intelligent Healthcare Data Platform(IHDP) for empowering compony solutions using AWS service. 

More responsibilities/details as below. 
  • Build APIs to the processing of patient records and providing access for downstream usages.
  • Build Infrastructure on AWS and compliance for HIPPA and GDPR standard.

IT Consultant at Xuenn Pte Ltd.May, 2018 - Sep, 2018

The biggest challenge in Xuenn is facing performance issues in the original data warehouse. And I lead a project to build up a Hadoop cluster to reduce original EDW loading and improve various data pipelines.

More responsibilities/details as below.  
  • Perform adopting new technologies to implant into, to fuse into or to replace with existing systems for gaining leaps in performance, benefits and capabilities of the users. 
  • Building-up multiple systems and integrating with existing system, implemented Hadoop, data mining or data warehouse systems.
  • Design an architecture which processing real-time data end-to-end using various Hadoop solutions without coding. 

Software Engineer at Athemaster Co., Ltd.Jan, 2016 - Apr, 2018 

Athemaster is a technology company offering solutions and expertise in implementing Enterprise Data Hub and automating Data integration with Open Source technologies such as Apache Hadoop and Spark. More responsibilities/details as below.
  • Focus on Enterprise Big Data solution such as Hadoop and Spark (Cloudera CDH). 
  • Maintain and improved other companies' Hadoop cluster. 
  • Help other companies integrate with Hadoop and resolved the technical issues. 
  • Build data pipelines via python ETL data to Hadoop.

Certification and License


Readings 00 00@2x ce5676dabcce042724a6fc4c3413d6a86ad9c78eecb848896433e32c60b7006b


CCA-175: CCA Spark and Hadoop Developer



Readings 00 01@2x 77cc06c91fae4dd43a069fa4b813524cd022d4a79115524d3f0d6b9220dfd71d

Cloudera Certified Administrator for Hadoop

Education



Undergraduate studies at Tamkang University, with a concentration in Department of Management Sciences.

                                                                                                                                                                                                     2008 - 2012