CakeResume Talent Search

Advanced filters
On
De 4 a 6 años
6-10 años
10-15 años
Más de 15 años
Avatar of the user.
Avatar of the user.
資深資料工程師 @緯創資通股份有限公司
2020 ~ Presente
Data Analyst、Data Engineer、Data Scientist、Customer Experience Analyst、Solution Architect、Cloud Architect
En un mes
python
PowerBI
Power Platform
Empleado
Listo para la entrevista
A tiempo completo / Interesado en trabajar a distancia
De 4 a 6 años
元智大學 Yuan Ze University
工業工程與管理學所
Avatar of Ching You.
Avatar of Ching You.
Product Design Consultant @Freelancer
2023 ~ 2024
UX Researcher / UIUX Designer / Product Designer
En un mes
Ching You Hi, I'm Ching You, a Product Designer with six years of experience developing digital products. I specialize in creating an indulging atmosphere for digital products and creating logical user flow based on the result of UX research. As I embark on the next chapter of my career, I eagerly seek opportunities to contribute my wealth of experience and pioneering spirit to a prominent digital product company on the global stage. Let's explore new horizons together! Taipei City, Taiwan https://chingyou.webflow.io 工作經歷 Product Designer • 時
User Interfaces
Service Design
User Research
Empleado
Listo para la entrevista
A tiempo completo / Interesado en trabajar a distancia
De 4 a 6 años
Shih Chien University
B.F.A Communication Design
Avatar of 彭靖鈞.
Avatar of 彭靖鈞.
前端工程師 @Lctech_雷麒科技有限公司
2021 ~ Presente
Front-end Engineer
En un mes
彭靖鈞 四年前端經驗,熱愛思考與解決問題所帶來的成就感,希望能找到志同道合的工作夥伴。 Taoyuan City, Taiwan• [email protected] 工作經歷 前端工程師 • Lctech_雷麒科技有限公司 七月Present 開發 JKF 旗下的 JVID 前端系統 與團隊協同將舊有的 PHP 專案重構為前後端分離 編
Vue.js
Nuxt.js
React.js
Empleado
Listo para la entrevista
A tiempo completo / Interesado en trabajar a distancia
De 4 a 6 años
南台科技大學
資訊工程系
Avatar of the user.
Avatar of the user.
Past
UX/UI 設計師 @網際威信股份有限公司
2023 ~ Presente
UX/UI Designer
En un mes
UI/UX Design
Flowchart
UI Flow
Desempleado
Listo para la entrevista
A tiempo completo / Interesado en trabajar a distancia
De 4 a 6 años
iSpan資展國際
前端工程師就業養成班
Avatar of 蔡楓淋.
Avatar of 蔡楓淋.
Past
Account manager @iSPOT Media 艾斯博媒體股份有限公司
2023 ~ 2024
資深數位行銷專員、資深電商行銷專員
En un mes
我是蔡楓淋 總共有四年數位行銷及電子商務 0 ~ 1 經驗,策劃營運與媒體佈局,擬定年度行銷檔期及 D2C 規劃,投放數位媒體廣告( Meta Ads、Google Ads、LINE LAP )並單一媒體百萬投放經驗,搭建完整漏斗佈局,注重數據成效分析與運用 CRM 分眾溝通,過往經營品牌電
廣告投放
廣告企劃案╱文案撰寫
數位行銷
Desempleado
Listo para la entrevista
A tiempo completo / Interesado en trabajar a distancia
De 4 a 6 años
中華醫事科技大學
資訊管理
Avatar of 王心妤.
Avatar of 王心妤.
Past
助理工程師 @第一社會企業股份有限公司
2023 ~ 2023
MIS程式設計師,軟體設計工程師,Web開發工程師
En un mes
— 王心妤 Lorem ipsum dolor sit amet, consectetuer adipiscing elit, sed diam nonummy nibh euismod tincidunt ut laoreet dolore magna aliquam erat volutpat. Ut wisi enim ad minim veniam, quis nostrud exerci tation ullamcorper suscipit lobortis nisl ut aliquip. New Taipei City, Taiwan 工作經歷 三月Present 資訊工程師 車麗屋汽車百貨股份有限公司 1.使用.Net core 6 MVC開發LINE會員專區 (1)前端使用DEVEXPRESS+html+JavaScript (2)後
Word
PowerPoint
Excel
Desempleado
Listo para la entrevista
A tiempo completo / Interesado en trabajar a distancia
De 4 a 6 años
萬能科技大學
資訊管理
Avatar of 陳勤霖.
Avatar of 陳勤霖.
Past
博士後研究員 @洛桑大學神經發育疾病實驗室
2023 ~ 2023
Data Scientist, Data Analyst, Machine Learning Engineer
En un mes
陳勤霖 神經工程博士背景的數據分析師 Ph.D. in Neuroscience from Neuroengineering lab I have 5 years of hands-on experience in image and data analysis with biotechnology innovation projects. Dependable ability in managing collaborative projects to success. Business-driven motivation to apply analytic skills to optimize the product and its development procedure. https://chinlinchen1312.wixsite.com/chin-lin-chen 工作經歷 一月十二月 2023 博士後研究員 洛桑大
Data Science
Data Analysis
Machine Learning
Desempleado
Listo para la entrevista
A tiempo completo / Interesado en trabajar a distancia
De 4 a 6 años
洛桑聯邦理工學院(EPFL)
神經科學
Avatar of 紀孟佐.
Avatar of 紀孟佐.
.NET開發工程師 @趣遊科技有限公司
2023 ~ Presente
全端/後端工程師
En un mes
紀孟佐 曾經是電腦硬體工程師| 網頁後端工程師 | 國立台北科技大學碩士畢業 個性開朗、善於溝通,與各單位合作,目前是全端工程師。 [email protected] 技能 Front-end HTML/CSS JavaScript Back-end C/C++ JAVA C# ASP.MVC/APS.NET.CORE MSQL Elasticsearch Redis Other Git/GitLab/GitHub Docker CI/CD Jenkins
c#.net
ASP.NET Core
html + css + javascript
Empleado
Listo para la entrevista
A tiempo completo / Interesado en trabajar a distancia
De 4 a 6 años
國立台北科技大學
電機工程系(計算機組)
Avatar of the user.
Avatar of the user.
Past
資深前端工程師 Senior Front-End Developer @法樂設計有限公司
2019 ~ Presente
資深前端工程師
En un mes
React.js/Redux
JavaScript / ES6 / jQuery
SASS/SCSS
Desempleado
Listo para la entrevista
A tiempo completo / Interesado en trabajar a distancia
De 4 a 6 años
National Yang Ming University
生醫光電所
Avatar of 李慕全(MuChuan Li).
Avatar of 李慕全(MuChuan Li).
Past
Service Provider @Taron Solutions Limited
2023 ~ 2023
AI工程師、機器學習工程師、電腦視覺工程師、資料科學家、Machine Learning Engineer、Computer Vision Engineer、Data Scientist
En un mes
李慕全(MuChuan Li) 畢業於國立臺北科技大學資工所,研究領域為深度學習、電腦視覺、及影像處理。在學期間致力於應用電腦視覺技術解決交通問題,擁有多項產學合作的專案開發經驗,亦在電腦視覺領域中發表過多篇學術論文,主要研究主題包含物
Machine Learning
Computer Vision
Pytorch/Tensorflow
Desempleado
Listo para la entrevista
A tiempo completo / Interesado en trabajar a distancia
De 4 a 6 años
國立臺北科技大學
資訊工程

El plan de reclutamiento más ligero y eficaz

Busque currículums y tome la iniciativa de ponerse en contacto con los solicitantes de empleo para lograr una mayor eficacia en la contratación. La elección de cientos de empresas.

  • Examinar todos los resultados de la búsqueda
  • Acceso ilimitado para iniciar nuevas conversaciones
  • currículos accesibles sólo para empresas de pago
  • Ver dirección de correo electrónico y números de teléfono de los usuarios
Consejos de búsqueda
1
Search a precise keyword combination
senior backend php
If the number of the search result is not enough, you can remove the less important keywords
2
Use quotes to search for an exact phrase
"business development"
3
Use the minus sign to eliminate results containing certain words
UI designer -UX
Sólo los currículums públicos están disponibles con el plan gratuito.
Actualiza a un plan avanzado para ver todos los resultados de la búsqueda incluyendo decenas de miles de currículums exclusivos en CakeResume.

Definition of Reputation Credits

Technical Skills
Specialized knowledge and expertise within the profession (e.g. familiar with SEO and use of related tools).
Problem-Solving
Ability to identify, analyze, and prepare solutions to problems.
Adaptability
Ability to navigate unexpected situations; and keep up with shifting priorities, projects, clients, and technology.
Communication
Ability to convey information effectively and is willing to give and receive feedback.
Time Management
Ability to prioritize tasks based on importance; and have them completed within the assigned timeline.
Teamwork
Ability to work cooperatively, communicate effectively, and anticipate each other's demands, resulting in coordinated collective action.
Leadership
Ability to coach, guide, and inspire a team to achieve a shared goal or outcome effectively.
En un mes
Senior Data Engineer at Paktor x M17 Entertainment Group | AWS x GCP x Azure Big Data Specialist | Data Architect
Logo of KKCompany.
KKCompany
2023 ~ Presente
Taiwan
Professional Background
Situación actual
Empleado
Progreso en la búsqueda de empleo
Abierto a oportunidades
Professions
Data Engineer, Back-end Engineer
Fields of Employment
Software
Experiencia laboral
10-15 años
Management
I've had experience in managing 1-5 people
Habilidades
Big Data
Data Engineering
ETL
AWS
GCP
Python
BigQuery
Data Warehouse
Data Pipeline
Java
Azure
SQL
Spark
kafka
spark streaming
Scala
Redshift
HBase
SQL Server
AWS S3
AWS Lambda
MongoDB
Hadoop
Hadoop Distributed File System
AWS SQS
Azure Storage
MySQL
PostgreSQL
Postman for API
Snowflake
Idiomas
English
Profesional
Chinese
Nativo o bilingüe
Job search preferences
Posición
Backend Engineer, Data Engineer, MLOps Engineer
Tipo de trabajo
A tiempo completo
Ubicación
Taiwan, 台灣, Singapore, Hong Kong
A distancia
Interesado en trabajar a distancia
Freelance
No.
Educación
Escuela
National Taiwan University
Mayor
EMBA Programs, Business Administration, Accounting, Finance and International Business.
Imprimir
Ahzwaym2ourqm1t0glsc

Chin-Hung (Wilson) Liu

I am a lead architect responsible for designing and implementing a large-scale data pipeline for Lomotif, Paktor x 17LIVE, utilizing GCP/AWS/Python/Scala, in collaboration with data science and machine learning teams in Singapore and TW HQ, as well as with the Hadoop ecosystem (HDFS/HBase/Kafka) at JSpectrum in Hong Kong and Sydney. 


With over 15 years of experience in designing and developing Java/Scala/Python-based applications for daily operations, I bring:

● At least 8 years of experience in data analysis, pipeline design and development, and tool building as a team member. 

● In-depth knowledge of the Spark and Hadoop ecosystems, including Hadoop, HDFS, HBase, and more. 
● Strong skills in designing and developing Big Data services on AWS and GCP. 
 Extensive expertise in developing generic distributed systems, streaming processing, machine learning pipelines, and continuously improving ML models.


Senior Data Engineer at Paktor x 17LIVE| AWS Big Data Specialist | Data Architect 
Singapore / Hong Kong / Taiwan

[email protected]

https://www.linkedin.com/in/chin-hung-wilson-liu-29392957

Nanxing Rd., Xizhi Dist., New Taipei City, Taiwan (R.O.C.)

Experience 

Senior Data Engineer (DataOps / AI) / Lomotif Private Limited / Singapore

Jul. 2021 - Present.

Description and Responsibilities: Lomotif is a leading short video social platform in South America and India that holds PBs of videos in buckets and serves millions of users. DataOps and AI team take part in many challenging projects e.g. Ncanto, XROAD services, Ray Serve, and scalable model serving frameworks for support the recommendation and moderation pipeline, also integrated Universal Music Group music (UMG) and full catalog feed with 7digital. DataOps team handling 10TB+ data for day-to-day operation, moderating model training results, and designing SLIs/SLOs for EKS Clusters. More responsibilities/details as below.

  • Optimize music (UMG) pipeline with queries and memories for Elasticsearch and PostgreSQL, the pipeline saving 90% execution time from 10+ hours to 40 mins.
  • Migrate service from apache spark, AWS Data Lake Formation to AWS MWAA, EKS airflow environment. 
  • Design, and deliver distributed system for Ray Serve with AI team.
  • Design, and implement a modern machine learning pipeline for a recommendation, and moderation pipe.
  • Design SLA and implement alert log reporting system (history logs) for moderation pipeline, histories logs handling application, server levels information for further investigation.
  • Supporting other departments to gather data in the appropriate platforms.
Tech Stacks : 
  • Streaming, Snowpipe/Kinesis/Firehose
  • Monitoring, CloudWatch/Grafana
  • Orchestration, AWS MWAA / Airflow
  • Kubernetes, EKS
  • Message, SQS/SNS
  • MLflow, Ray Serve/EMR/Lambda
  • Storage, Snowflake / RDS (PostgreSQL) / ElastiCache (Redis) / Elastic search
  • Bucket, AWS S3
Reports to : VP of Data Engineering


Senior Data Engineer / Handshakes by DC Frontiers / Singapore

Oct. 2020 - May. 2021.

Description and Responsibilities: The main responsibility of the engineering team is launching ScoutAsia by Nikkei and The Financial Times Nikkei content to SGX TitanOTC's platform. Titan Users will be able to access Nikkei news articles from across 11 categories, including equities, stocks, indices, foreign exchange, and iron ore. DPP (Data team) is processing hundreds of GB articles/market/financial/relationships and organization for day-to-day operation on Azure and on-premise environments. More responsibilities/details as below.

  • Identifying, digging bottlenecks, and problem-solving especially optimizing the performance of SQL Server, NoSQL (Azure Cosmos), resource units, and message queues, reducing/saving almost 50-75% of resources. 
  • Identifying and solving the problems between machine learning/backend/frontend/DDP side and giving the advance logical/physical design of a system. Displayed technical expertise in optimizing the databases and improving the data pipeline to achieve the objective.
  • Bring in industry standards to data management to deliver data at the end objective. 
  • Building, and recruiting the new data engineering staff for the next-generation, enterprise data pipeline.
Tech Stacks : 
  • Storage, Azure Cosmos DB/Gremlin/SQL Server/MYSQL/Redis
  • Storage (Bucket), Azure Blob/AWS S3
  • Streaming/Batch/transform, Spark/Scala (90% codebase coverage)
  • Message, Azure service bus, queue storage
  • Search, Elastic search
  • Algorithm, graph/concordance
Reports to : CTO

Senior Data Engineer / 17LIVE Inc. / Taiwan, Taipei.

Feb. 2020 - Jul. 2020

Description and Responsibilities: The big challenge of 17 Media data teams is facing fast-growing data volume (processing 5-10x TB level daily), complex cooperation with stakeholders, the cost optimization of the pipeline, and refactoring big latency systems .etc. As a senior data member, I’m making a data dictionary and trying to explain/design how the whole pipeline works with each component, especially how to solve those bottlenecks. More responsibilities/details as below. 

  • Leading, and architect large-scale data pipeline for supporting scientists and shareholders. 
  • Optimize, ensure quality and play a tough role in data lake projects/data pipes. infrastructure. 
  • Define, and design stage, dimension, production, and fact tables for data warehouse (BigQuery). 
  • Coordinate with client / QA / backend team for QC lists / MongoDB change stream workers. 
  • Architect workflows with those components, Dataflow, Cloud Functions, and GCS. 
  • Recruiting (Jr./Sr.) data engineering members, setting goals, and sprint management.

Tech Stacks : 

  • Storage, GCS/BigQuery/Firebase/MongoDB/MYSQL 
  • Realtime process and Message system, DataFlow (Apache Beam) / BigQuery Streaming / MongoDB Change Stream / Fluentd / Firebase / Pub/Sub 
  • ETL/ELT workflow, Digdag / Embulk 
  • Data warehouse, Visualization, BigQuery / Superset / Chartio / Data Studio 
  • Continuous deployment, docker, CricleCI 

Reports to : Data Head

Data Engineer / Paktor Pte. Ltd. / Singapore 

Sep. 2015 - Dec. 2019.

Description and Responsibilities : This is another 0 to 1 story. As an early data member, we need to figure out the data driven policy, strategies, engineering requirements from the company. In Paktor, data / backend sides are 100% on AWS, therefore the whole data ingestion, automation and data warehouse etc. are relying on those components. We are processing 50-100x GB realtime / batch jobs and the other data sources (RDBMS, APIs) for ETL/ELT on S3, Redshift, the data platform helps our marketing / HQ scientists team getting data into insights and making good decisions. More responsibilities / details as below. 

  • Supports Big Data and batch, real-time analytical solutions leveraging transformational technologies. 
  • Optimize data pipeline on AWS using Kinesis-Firehose/Lambda/Kinesis Analytics/Data Pipeline, and optimize, resizing Redshift clusters and related scripts. 
  • Translates complex analytics requirements into detailed architecture, design, and high performing software such as machine-learning, CI/CD of recommendation pipeline. 
  • Collaborate with client / backend side developers to formulate innovative solutions to experiment and implement related algorithms. 

Tech Stacks : 

  • Storage, S3/Redshift/Aurora - Realtime process and Message system, Kinesis Firehose / SNS 
  • Data warehouse, Visualization, Redshift / Klipfolio / Metabase 
  • ETL/ELT workflow, Lambda / SNS / Batch / Python 
  • Recommendation, ML, DynamoDB / EMR / Spark / Sagemaker 
  • Metadata management, Athena (presto) / Glue / Redshift Spectrum 
  • Continuous deployment, Elasticbeanstalk / Cloudformation 
  • Operations, PagerDuty / Zapier / Cloud Watch 

Reports to : CTO, Data Head

System Analyst (Data Backend Engineer) / JSpectrum Software Limited / Hong Kong 

 Jan. 2014 - Aug 2015.

Description and Responsibilities : JSPectrum is a leading passive location-based service company in Hong Kong which holds many interesting products such as NetProbe, NetWhere, NetAd etc. In Optus (The main project in Sydney), the main responsibility of system analyst is designing / implementing data ingestion (real-time processing) / load and management data with major components of the Hadoop ecosystem. We meet the challenge to process 15,000 TPS, 60,000 inserts per second and 300 GB daily storages, therefore we are trying to optimize those components with Kafka consumers, HDFS storages and re-designing keys / columns of HBase to fulfill the requirement and deployed NetAd, whole in-house solutions on Optus. More responsibilities / details as below. 

  • Design, implement and optimize Hadoop ecosystems, MLP, real-time processing on Optus in house servers with our main product NetAd, NetWhere. We are focusing on HBase schema, HDFS, balancing Kafka consumers and more issues on data ingestion. 
  • Collaborate with shareholders and LBS team members for further requirements with HeapMap. 

Tech Stacks : 

  • Storage, HDFS / HBase
  • Realtime process and Message system, Kafka streaming, Log systems 
  • Data warehouse, Visualization, HBase / NetWhere (Dashboard) 
  • Hadoop ecosystem, Hadoop / HDFS / Zookeeper / Spark / Hive
  • ETL/ELT workflow, Spark / Hive / Scala / Java

Reports to : CTO


Senior Software Engineer / Toro Development Ltd. / Taiwan, Taipei. 

Oct. 2012 - Dec. 2013.

Description and Responsibilities : TORO is a technology business that provides a mobile platform and its associated systems, services and rules to help Brands (with initial focus on Sports Teams, Smart Cities and Streaming apps) become super-apps to generate additional revenue with minimum effort. Responsibilities as below. 

  • Design, implement and test back-office modules for NFC wallet platform, Trusted Service Managers (TSM) and distributed NFC services to end­ users / stakeholders. 
  • Implement RESTful services and deliver endpoints for wallet managers and collaborating with front­end, backend teams for further business requirements. 

Tech Stacks: MYSQL / Spring / Hibernate / XML / Apache Camel / Java / POJO .etc. 

Reports to : Head of Server Solutions


Software Engineer / Digital River / Taiwan, Taipei. 

Oct. 2011 - Sep. 2012.

Description and Responsibilities : Digital river proactive partners, providing API-based Payments & Risk, Order Management and Commerce services to leading enterprise brands. The big challenge to DR is integrating with the current module and working well with a huge code base (over 2+ millions lines), the strict process including analysis requirements, design, implement, test and code review. More responsibilities as below. 

  • Design, implement custom bundle project, bundle customized by shoppers to pick products of groups and get special discounts, the main stakeholders /users from Logitech, Microsoft. 
  • Analysis, collect business requirements, identify use cases and collaborate with business analysts and deliver related diagrams, documents. 

Tech Stacks: Oracle / Tomcat / Spring / Struts / JDO / XML / JUnit / Java / J2EE .etc. 

Reports to : Technical Development Manager


Technical Supervisor / Stark Technology Inc. / Taiwan, Taipei. 

Oct. 2008 - Sep. 2011.

Description and Responsibilities : Stark Technology (STI) is the largest domestic system integrator in Taiwan. We plan and deliver complete ICT solutions for a wide spectrum of industries through representing and reselling the world's leading products. This is made possible by using the most advanced technology, and providing the best professional services. More responsibilities / projects as below. 

  • Lead, coach JR. programmers for the development process of enterprise modules, and design Fatwire CMS components as Template/Page/Cache .etc. 
  • Design, analyze DMDB systems, and implement functions to meet the requirements of queries / storage. Optimize performance for online servers and GC tuning. 

Tech Stacks : Oracle / Sybase / Tomcat / Weblogic / Spring / Struts / Hibernate / Fatwire / Java / J2EE .etc. 

Reports to : Technical Manager


Relevant Skills and Qualifications


Big Data Tech Stacks

  • AWS Services, EC2/S3/Lambda/EMR/CloudWatch/SNS/SQS/Elastic Beanstalk 
  • AWS Big Data Solutions, Kinesis/Firehose/Athena/Redshift/Dynamodb 
  • GCP Big Data Solutions, BigQuery/PubSub/Dataflow/Cloud Functions 
  • Hadoop ecosystem, Hadoop/HDFS/Zookeeper/Hbase/Hive 
  • Spark Streaming/Apache Kafka 
  • CI/CD: Jenkins/Cloud Formation/GitLab/Grafana

Specific Skills

  • Solid, well-designed real-time streaming/batch processing, ETL systems.
  • Monitors and conducts data-pipeline / machine learning pipeline development requests through lifecycle management and ensures that the technical solution meets.
  • Diagnosing and troubleshooting Redshift and specific clusters management.
  • Development of micro-services and endpoints based on enterprise integration patterns. Knowledge over garbage collection (JVM) tuning technologies for various servers.
  • Developed multi-threading processing consuming work and managed transactions.

Certifications and Training

  • Sun Certified Web Component Developer Java 2 Platform, Enterprise Edition. 
  • Sun Certified Programmer for the Java 2 Platform. 
  • Red Hat Enterprise Directory Services and Authentication Attended. 
  • Project Management Professional (PMP)® Attended. 
  • AWS Certified Solutions Architect Attended. 
  • Big Data on AWS Attended. 
  • Azure Data Engineer AssociateAttended. 

Education


National Taiwan University, 2010 – 2011

EMBA Programs, Business Administration, Accounting, Finance and International Business.


Chinese Culture University Master of Information Management, 2002 – 2005

Computer Science, Data Mining, Expert Systems and Knowledge Base as major concentration.


Chinese Culture University, Bachelor Degree of Science in Journalism, 1998 - 2002

Resume
Perfil
Ahzwaym2ourqm1t0glsc

Chin-Hung (Wilson) Liu

I am a lead architect responsible for designing and implementing a large-scale data pipeline for Lomotif, Paktor x 17LIVE, utilizing GCP/AWS/Python/Scala, in collaboration with data science and machine learning teams in Singapore and TW HQ, as well as with the Hadoop ecosystem (HDFS/HBase/Kafka) at JSpectrum in Hong Kong and Sydney. 


With over 15 years of experience in designing and developing Java/Scala/Python-based applications for daily operations, I bring:

● At least 8 years of experience in data analysis, pipeline design and development, and tool building as a team member. 

● In-depth knowledge of the Spark and Hadoop ecosystems, including Hadoop, HDFS, HBase, and more. 
● Strong skills in designing and developing Big Data services on AWS and GCP. 
 Extensive expertise in developing generic distributed systems, streaming processing, machine learning pipelines, and continuously improving ML models.


Senior Data Engineer at Paktor x 17LIVE| AWS Big Data Specialist | Data Architect 
Singapore / Hong Kong / Taiwan

[email protected]

https://www.linkedin.com/in/chin-hung-wilson-liu-29392957

Nanxing Rd., Xizhi Dist., New Taipei City, Taiwan (R.O.C.)

Experience 

Senior Data Engineer (DataOps / AI) / Lomotif Private Limited / Singapore

Jul. 2021 - Present.

Description and Responsibilities: Lomotif is a leading short video social platform in South America and India that holds PBs of videos in buckets and serves millions of users. DataOps and AI team take part in many challenging projects e.g. Ncanto, XROAD services, Ray Serve, and scalable model serving frameworks for support the recommendation and moderation pipeline, also integrated Universal Music Group music (UMG) and full catalog feed with 7digital. DataOps team handling 10TB+ data for day-to-day operation, moderating model training results, and designing SLIs/SLOs for EKS Clusters. More responsibilities/details as below.

  • Optimize music (UMG) pipeline with queries and memories for Elasticsearch and PostgreSQL, the pipeline saving 90% execution time from 10+ hours to 40 mins.
  • Migrate service from apache spark, AWS Data Lake Formation to AWS MWAA, EKS airflow environment. 
  • Design, and deliver distributed system for Ray Serve with AI team.
  • Design, and implement a modern machine learning pipeline for a recommendation, and moderation pipe.
  • Design SLA and implement alert log reporting system (history logs) for moderation pipeline, histories logs handling application, server levels information for further investigation.
  • Supporting other departments to gather data in the appropriate platforms.
Tech Stacks : 
  • Streaming, Snowpipe/Kinesis/Firehose
  • Monitoring, CloudWatch/Grafana
  • Orchestration, AWS MWAA / Airflow
  • Kubernetes, EKS
  • Message, SQS/SNS
  • MLflow, Ray Serve/EMR/Lambda
  • Storage, Snowflake / RDS (PostgreSQL) / ElastiCache (Redis) / Elastic search
  • Bucket, AWS S3
Reports to : VP of Data Engineering


Senior Data Engineer / Handshakes by DC Frontiers / Singapore

Oct. 2020 - May. 2021.

Description and Responsibilities: The main responsibility of the engineering team is launching ScoutAsia by Nikkei and The Financial Times Nikkei content to SGX TitanOTC's platform. Titan Users will be able to access Nikkei news articles from across 11 categories, including equities, stocks, indices, foreign exchange, and iron ore. DPP (Data team) is processing hundreds of GB articles/market/financial/relationships and organization for day-to-day operation on Azure and on-premise environments. More responsibilities/details as below.

  • Identifying, digging bottlenecks, and problem-solving especially optimizing the performance of SQL Server, NoSQL (Azure Cosmos), resource units, and message queues, reducing/saving almost 50-75% of resources. 
  • Identifying and solving the problems between machine learning/backend/frontend/DDP side and giving the advance logical/physical design of a system. Displayed technical expertise in optimizing the databases and improving the data pipeline to achieve the objective.
  • Bring in industry standards to data management to deliver data at the end objective. 
  • Building, and recruiting the new data engineering staff for the next-generation, enterprise data pipeline.
Tech Stacks : 
  • Storage, Azure Cosmos DB/Gremlin/SQL Server/MYSQL/Redis
  • Storage (Bucket), Azure Blob/AWS S3
  • Streaming/Batch/transform, Spark/Scala (90% codebase coverage)
  • Message, Azure service bus, queue storage
  • Search, Elastic search
  • Algorithm, graph/concordance
Reports to : CTO

Senior Data Engineer / 17LIVE Inc. / Taiwan, Taipei.

Feb. 2020 - Jul. 2020

Description and Responsibilities: The big challenge of 17 Media data teams is facing fast-growing data volume (processing 5-10x TB level daily), complex cooperation with stakeholders, the cost optimization of the pipeline, and refactoring big latency systems .etc. As a senior data member, I’m making a data dictionary and trying to explain/design how the whole pipeline works with each component, especially how to solve those bottlenecks. More responsibilities/details as below. 

  • Leading, and architect large-scale data pipeline for supporting scientists and shareholders. 
  • Optimize, ensure quality and play a tough role in data lake projects/data pipes. infrastructure. 
  • Define, and design stage, dimension, production, and fact tables for data warehouse (BigQuery). 
  • Coordinate with client / QA / backend team for QC lists / MongoDB change stream workers. 
  • Architect workflows with those components, Dataflow, Cloud Functions, and GCS. 
  • Recruiting (Jr./Sr.) data engineering members, setting goals, and sprint management.

Tech Stacks : 

  • Storage, GCS/BigQuery/Firebase/MongoDB/MYSQL 
  • Realtime process and Message system, DataFlow (Apache Beam) / BigQuery Streaming / MongoDB Change Stream / Fluentd / Firebase / Pub/Sub 
  • ETL/ELT workflow, Digdag / Embulk 
  • Data warehouse, Visualization, BigQuery / Superset / Chartio / Data Studio 
  • Continuous deployment, docker, CricleCI 

Reports to : Data Head

Data Engineer / Paktor Pte. Ltd. / Singapore 

Sep. 2015 - Dec. 2019.

Description and Responsibilities : This is another 0 to 1 story. As an early data member, we need to figure out the data driven policy, strategies, engineering requirements from the company. In Paktor, data / backend sides are 100% on AWS, therefore the whole data ingestion, automation and data warehouse etc. are relying on those components. We are processing 50-100x GB realtime / batch jobs and the other data sources (RDBMS, APIs) for ETL/ELT on S3, Redshift, the data platform helps our marketing / HQ scientists team getting data into insights and making good decisions. More responsibilities / details as below. 

  • Supports Big Data and batch, real-time analytical solutions leveraging transformational technologies. 
  • Optimize data pipeline on AWS using Kinesis-Firehose/Lambda/Kinesis Analytics/Data Pipeline, and optimize, resizing Redshift clusters and related scripts. 
  • Translates complex analytics requirements into detailed architecture, design, and high performing software such as machine-learning, CI/CD of recommendation pipeline. 
  • Collaborate with client / backend side developers to formulate innovative solutions to experiment and implement related algorithms. 

Tech Stacks : 

  • Storage, S3/Redshift/Aurora - Realtime process and Message system, Kinesis Firehose / SNS 
  • Data warehouse, Visualization, Redshift / Klipfolio / Metabase 
  • ETL/ELT workflow, Lambda / SNS / Batch / Python 
  • Recommendation, ML, DynamoDB / EMR / Spark / Sagemaker 
  • Metadata management, Athena (presto) / Glue / Redshift Spectrum 
  • Continuous deployment, Elasticbeanstalk / Cloudformation 
  • Operations, PagerDuty / Zapier / Cloud Watch 

Reports to : CTO, Data Head

System Analyst (Data Backend Engineer) / JSpectrum Software Limited / Hong Kong 

 Jan. 2014 - Aug 2015.

Description and Responsibilities : JSPectrum is a leading passive location-based service company in Hong Kong which holds many interesting products such as NetProbe, NetWhere, NetAd etc. In Optus (The main project in Sydney), the main responsibility of system analyst is designing / implementing data ingestion (real-time processing) / load and management data with major components of the Hadoop ecosystem. We meet the challenge to process 15,000 TPS, 60,000 inserts per second and 300 GB daily storages, therefore we are trying to optimize those components with Kafka consumers, HDFS storages and re-designing keys / columns of HBase to fulfill the requirement and deployed NetAd, whole in-house solutions on Optus. More responsibilities / details as below. 

  • Design, implement and optimize Hadoop ecosystems, MLP, real-time processing on Optus in house servers with our main product NetAd, NetWhere. We are focusing on HBase schema, HDFS, balancing Kafka consumers and more issues on data ingestion. 
  • Collaborate with shareholders and LBS team members for further requirements with HeapMap. 

Tech Stacks : 

  • Storage, HDFS / HBase
  • Realtime process and Message system, Kafka streaming, Log systems 
  • Data warehouse, Visualization, HBase / NetWhere (Dashboard) 
  • Hadoop ecosystem, Hadoop / HDFS / Zookeeper / Spark / Hive
  • ETL/ELT workflow, Spark / Hive / Scala / Java

Reports to : CTO


Senior Software Engineer / Toro Development Ltd. / Taiwan, Taipei. 

Oct. 2012 - Dec. 2013.

Description and Responsibilities : TORO is a technology business that provides a mobile platform and its associated systems, services and rules to help Brands (with initial focus on Sports Teams, Smart Cities and Streaming apps) become super-apps to generate additional revenue with minimum effort. Responsibilities as below. 

  • Design, implement and test back-office modules for NFC wallet platform, Trusted Service Managers (TSM) and distributed NFC services to end­ users / stakeholders. 
  • Implement RESTful services and deliver endpoints for wallet managers and collaborating with front­end, backend teams for further business requirements. 

Tech Stacks: MYSQL / Spring / Hibernate / XML / Apache Camel / Java / POJO .etc. 

Reports to : Head of Server Solutions


Software Engineer / Digital River / Taiwan, Taipei. 

Oct. 2011 - Sep. 2012.

Description and Responsibilities : Digital river proactive partners, providing API-based Payments & Risk, Order Management and Commerce services to leading enterprise brands. The big challenge to DR is integrating with the current module and working well with a huge code base (over 2+ millions lines), the strict process including analysis requirements, design, implement, test and code review. More responsibilities as below. 

  • Design, implement custom bundle project, bundle customized by shoppers to pick products of groups and get special discounts, the main stakeholders /users from Logitech, Microsoft. 
  • Analysis, collect business requirements, identify use cases and collaborate with business analysts and deliver related diagrams, documents. 

Tech Stacks: Oracle / Tomcat / Spring / Struts / JDO / XML / JUnit / Java / J2EE .etc. 

Reports to : Technical Development Manager


Technical Supervisor / Stark Technology Inc. / Taiwan, Taipei. 

Oct. 2008 - Sep. 2011.

Description and Responsibilities : Stark Technology (STI) is the largest domestic system integrator in Taiwan. We plan and deliver complete ICT solutions for a wide spectrum of industries through representing and reselling the world's leading products. This is made possible by using the most advanced technology, and providing the best professional services. More responsibilities / projects as below. 

  • Lead, coach JR. programmers for the development process of enterprise modules, and design Fatwire CMS components as Template/Page/Cache .etc. 
  • Design, analyze DMDB systems, and implement functions to meet the requirements of queries / storage. Optimize performance for online servers and GC tuning. 

Tech Stacks : Oracle / Sybase / Tomcat / Weblogic / Spring / Struts / Hibernate / Fatwire / Java / J2EE .etc. 

Reports to : Technical Manager


Relevant Skills and Qualifications


Big Data Tech Stacks

  • AWS Services, EC2/S3/Lambda/EMR/CloudWatch/SNS/SQS/Elastic Beanstalk 
  • AWS Big Data Solutions, Kinesis/Firehose/Athena/Redshift/Dynamodb 
  • GCP Big Data Solutions, BigQuery/PubSub/Dataflow/Cloud Functions 
  • Hadoop ecosystem, Hadoop/HDFS/Zookeeper/Hbase/Hive 
  • Spark Streaming/Apache Kafka 
  • CI/CD: Jenkins/Cloud Formation/GitLab/Grafana

Specific Skills

  • Solid, well-designed real-time streaming/batch processing, ETL systems.
  • Monitors and conducts data-pipeline / machine learning pipeline development requests through lifecycle management and ensures that the technical solution meets.
  • Diagnosing and troubleshooting Redshift and specific clusters management.
  • Development of micro-services and endpoints based on enterprise integration patterns. Knowledge over garbage collection (JVM) tuning technologies for various servers.
  • Developed multi-threading processing consuming work and managed transactions.

Certifications and Training

  • Sun Certified Web Component Developer Java 2 Platform, Enterprise Edition. 
  • Sun Certified Programmer for the Java 2 Platform. 
  • Red Hat Enterprise Directory Services and Authentication Attended. 
  • Project Management Professional (PMP)® Attended. 
  • AWS Certified Solutions Architect Attended. 
  • Big Data on AWS Attended. 
  • Azure Data Engineer AssociateAttended. 

Education


National Taiwan University, 2010 – 2011

EMBA Programs, Business Administration, Accounting, Finance and International Business.


Chinese Culture University Master of Information Management, 2002 – 2005

Computer Science, Data Mining, Expert Systems and Knowledge Base as major concentration.


Chinese Culture University, Bachelor Degree of Science in Journalism, 1998 - 2002