CakeResume 找人才

進階搜尋
On
4 到 6 年
6 到 10 年
10 到 15 年
15 年以上
Avatar of 陳奕妤.
Avatar of 陳奕妤.
曾任
Senior Data Analyst @趨勢科技
2022 ~ 現在
Data Scientist, Data Analyst, Machine Learning Engineer
一個月內
Cathy Chen Sr. Data Analyst Senior data analyst with over 6 years experience in ETL, data visualization, exploratory data analysis, machine learning, deep learning, customized online dashboard using SQL , R , Python and data analytics tools. Data Scientist, Data Analyst Taipei, Taiwan [email protected] Experience Sr. Data Analyst • TrendMicro NovNow Work with cross-functional teams(UI/UX designer, Front-end, Back-end, Marketing, PM, Sales) to provide related data, design metrics, report and dashboard. Cross app data tracking and user journey analysis. VisionOne customers engagement score - the metrics can help fields to
python
R
SQL
待業中
正在積極求職中
全職 / 對遠端工作有興趣
4 到 6 年
輔仁大學 Fu Jen Catholic University
統計資訊學系
Avatar of Vel Tien-Yun Wu.
Avatar of Vel Tien-Yun Wu.
Data Engineer @Groundhog Technologies Inc.
2021 ~ 2024
Data Analyst、Data Engineer、Data Scientist、Customer Experience Analyst
一個月內
Vel Tien-Yun Wu I bring 5 years of hands-on experience in data engineering and software development, with a focus on building scalable data processing systems utilizing Hadoop, Spark, Kafka and Docker. My expertise in developing efficient ETL pipelines has been fundamental in optimizing data workflows for various data warehouses, enhancing data integrity and availability. My track record includes managing high-volume data pipelines, automating scheduling processes to improve operational efficiency, and deploying monitoring solutions that have reduced Mean-Time-To-Repair (MTTR) by 40%. I have a strong foundation in SQL, especially PostgreSQL, which enables
Git
Python
Scala
就職中
正在積極求職中
全職 / 對遠端工作有興趣
4 到 6 年
University of Illinois at Urbana-Champaign, School of Information Sciences
Information Management
Avatar of 朱建銘.
Avatar of 朱建銘.
Soft Engineer @銓鍇國際股份有限公司
2023 ~ 現在
java程式開發
一個月內
務,並測試Dremio作為數據處理和分析工具。 建置Sam Project 進行Aws Lambda 部署。 建置Promethus, Grafana 監控Backend, Mongo 資源監控。 嘗試理解帳務Domain前處理技術, spark 相關技術的資料進行ETL處理。 嘗試使用Terrform 部署 參與重構的規劃討論,參與後端開發,並與前端進行整合,理解商務邏輯運
Java EE
JavaScript / ES6 / jQuery
JBoss Application Server
就職中
正在積極求職中
全職 / 對遠端工作有興趣
4 到 6 年
東南科技大學
資訊科技與通訊學系
Avatar of the user.
Avatar of the user.
曾任
資深前端工程師 Senior Front-End Developer @法樂設計有限公司
2019 ~ 現在
資深前端工程師
一個月內
React.js/Redux
JavaScript / ES6 / jQuery
SASS/SCSS
待業中
正在積極求職中
全職 / 對遠端工作有興趣
4 到 6 年
National Yang Ming University
生醫光電所
Avatar of 李雅涵.
Avatar of 李雅涵.
網頁設計 @震豪網路媒體股份有限公司
2023 ~ 現在
Visual Designer
一個月內
旅途探索那些枝微末節的細節,然後將生活中那份細膩練習連結在設計思考中 ;) & My Portfolio Design Tools Figma|Illustrator|Photoshop 專案 Lemon Scan Branding , UI/UX , Web Digit Spark Website - China UI , Web AquaFeb Branding , Graphic , Social Media Judy Barton Dept. Branding , Graphic 自我介紹 Hi ! 我是李雅涵 Hannah,擁有天馬行空創意的水瓶座,擁有 3-4 的設計經驗並對
Photoshop
Illustrator
Figma
職場能力評價7
就職中
正在積極求職中
全職 / 對遠端工作有興趣
4 到 6 年
華夏科技大學
室內設計
Avatar of the user.
Avatar of the user.
曾任
Software Engineer @17LIVE
2022 ~ 2024
Game Developer, Technical Game Designer, Game Programmer, Interactive Developer
一個月內
C#
UNITY
待業中
正在積極求職中
全職 / 對遠端工作有興趣
6 到 10 年
National ChengChi University (NCCU)
Computer Science
Avatar of Vu Nguyen Ngoc Quang.
Avatar of Vu Nguyen Ngoc Quang.
曾任
Mobile App Developer @Apple Inc.
2014 ~ 現在
Lead Infrastructure Engineer
兩個月內
and Scikit-Learn libraries for predictive analysis of customer behavior. Designed and implemented a scalable data warehouse architecture using Apache Cassandra, PostgresDB, and Redis. Optimized database performance by tuning queries in SQL Server, Oracle and PostgreSQL databases. Implemented efficient data processing algorithms on large datasets with Apache Spark, MapReduce, and Pandas Python. Created dashboards in Tableau Desktop Professional Edition to visualize complex datasets in an interactive manner. Created custom scripts to automate the extraction, transformation, and loading of Big Data into distributed systems. Utilized Amazon Web Services components such as EMR and S3 buckets
Machine learning
Virtualization Technologies
Pandas Python
待業中
正在積極求職中
全職 / 對遠端工作有興趣
6 到 10 年
Avatar of 林承緯.
Avatar of 林承緯.
Senior Software Engineer @集客數據行銷(震豪網路媒體)
2023 ~ 現在
Software Engineer / Backend Engineer
兩個月內
Justin Lin Taiwan, Taipei · [email protected] ·· github.com/jubeatwww Experience Lead Software Engineer • Digit Spark JanPresent | Taipei, Taiwan Manage and execute strategic refactoring projects for cost-efficient services, achieving a 35% reduction in database expenses. Guide my team through comprehensive refactoring of a 3TB database (2 billion records) and oversee architectural enhancements, achieving a drastic reduction in feature response times from over 10 minutes to under 5 seconds. Lead the infrastructure design and collaborate with the execution team to develop an advertising delivery system from scratch, achieving 100K daily ad views and accommodating
TypeScript
Django
Docker
就職中
正在積極求職中
全職 / 對遠端工作有興趣
4 到 6 年
National Chiao Tung University
Computer Science
Avatar of the user.
Avatar of the user.
Product Design Consultant @Freelancer
2023 ~ 2024
UX Researcher / UIUX Designer / Product Designer
一個月內
User Interfaces
Service Design
User Research
就職中
正在積極求職中
全職 / 對遠端工作有興趣
4 到 6 年
Shih Chien University
B.F.A Communication Design
Avatar of Leo Chang.
Avatar of Leo Chang.
資深工程師 @太引資訊系統股份有限公司
2014 ~ 現在
前後端工程師
一個月內
信現在所需技能都可以很快的銜接不同的工作 技能 程式語言 JAVA Python Angular HTML JavaScript SQL Oracle stored procedure Bash/Shell 資料處理/儲存/系統 Oracle PostgreSQL Spark Windows Linux (CentOS, RedHat, Ubuntu) Language Chinese English 專案管理 時程控管, 資源協調 跨部門溝通 客戶關係 語言 英文 - 可閱讀 , 基本溝通 學歷逢甲大學 資訊工
python programming
JAVA
SQL
就職中
正在積極求職中
全職 / 對遠端工作有興趣
10 到 15 年
逢甲大學
資訊工程

最輕量、快速的招募方案,數百家企業的選擇

搜尋履歷,主動聯繫求職者,提升招募效率。

  • 瀏覽所有搜尋結果
  • 每日可無限次數開啟陌生對話
  • 搜尋僅開放付費企業檢視的履歷
  • 檢視使用者信箱 & 電話
搜尋技巧
1
嘗試搜尋最精準的關鍵字組合
資深 後端 php laravel
如果結果不夠多,再逐一刪除較不重要的關鍵字
2
將須完全符合的字詞放在雙引號中
"社群行銷"
3
在不想搜尋到的字詞前面加上減號,如果想濾掉中文字,需搭配雙引號使用 (-"人資")
UI designer -UX
免費方案僅能搜尋公開履歷。
升級至進階方案,即可瀏覽所有搜尋結果(包含數萬筆覽僅在 CakeResume 平台上公開的履歷)。

職場能力評價定義

專業技能
該領域中具備哪些專業能力(例如熟悉 SEO 操作,且會使用相關工具)。
問題解決能力
能洞察、分析問題,並擬定方案有效解決問題。
變通能力
遇到突發事件能冷靜應對,並隨時調整專案、客戶、技術的相對優先序。
溝通能力
有效傳達個人想法,且願意傾聽他人意見並給予反饋。
時間管理能力
了解工作項目的優先順序,有效運用時間,準時完成工作內容。
團隊合作能力
具有向心力與團隊責任感,願意傾聽他人意見並主動溝通協調。
領導力
專注於團隊發展,有效引領團隊採取行動,達成共同目標。
一個月內
Senior Data Engineer at Paktor x M17 Entertainment Group | AWS x GCP x Azure Big Data Specialist | Data Architect
Logo of KKCompany.
KKCompany
2023 ~ 現在
Taiwan
專業背景
目前狀態
就職中
求職階段
目前會考慮了解新的機會
專業
數據工程師, 後端開發人員
產業
軟體
工作年資
10 到 15 年
管理經歷
我有管理 1~5 人的經驗
技能
Big Data
Data Engineering
ETL
AWS
GCP
Python
BigQuery
Data Warehouse
Data Pipeline
Java
Azure
SQL
Spark
kafka
spark streaming
Scala
Redshift
HBase
SQL Server
AWS S3
AWS Lambda
MongoDB
Hadoop
Hadoop Distributed File System
AWS SQS
Azure Storage
MySQL
PostgreSQL
Postman for API
Snowflake
語言能力
English
專業
Chinese
母語或雙語
求職偏好
希望獲得的職位
Backend Engineer, Data Engineer, MLOps Engineer
預期工作模式
全職
期望的工作地點
Taiwan, 台灣, Singapore, Hong Kong
遠端工作意願
對遠端工作有興趣
接案服務
學歷
學校
National Taiwan University
主修科系
EMBA Programs, Business Administration, Accounting, Finance and International Business.
列印
Ahzwaym2ourqm1t0glsc

Chin-Hung (Wilson) Liu

I am a lead architect responsible for designing and implementing a large-scale data pipeline for Lomotif, Paktor x 17LIVE, utilizing GCP/AWS/Python/Scala, in collaboration with data science and machine learning teams in Singapore and TW HQ, as well as with the Hadoop ecosystem (HDFS/HBase/Kafka) at JSpectrum in Hong Kong and Sydney. 


With over 15 years of experience in designing and developing Java/Scala/Python-based applications for daily operations, I bring:

● At least 8 years of experience in data analysis, pipeline design and development, and tool building as a team member. 

● In-depth knowledge of the Spark and Hadoop ecosystems, including Hadoop, HDFS, HBase, and more. 
● Strong skills in designing and developing Big Data services on AWS and GCP. 
 Extensive expertise in developing generic distributed systems, streaming processing, machine learning pipelines, and continuously improving ML models.


Senior Data Engineer at Paktor x 17LIVE| AWS Big Data Specialist | Data Architect 
Singapore / Hong Kong / Taiwan

[email protected]

https://www.linkedin.com/in/chin-hung-wilson-liu-29392957

Nanxing Rd., Xizhi Dist., New Taipei City, Taiwan (R.O.C.)

Experience 

Senior Data Engineer (DataOps / AI) / Lomotif Private Limited / Singapore

Jul. 2021 - Present.

Description and Responsibilities: Lomotif is a leading short video social platform in South America and India that holds PBs of videos in buckets and serves millions of users. DataOps and AI team take part in many challenging projects e.g. Ncanto, XROAD services, Ray Serve, and scalable model serving frameworks for support the recommendation and moderation pipeline, also integrated Universal Music Group music (UMG) and full catalog feed with 7digital. DataOps team handling 10TB+ data for day-to-day operation, moderating model training results, and designing SLIs/SLOs for EKS Clusters. More responsibilities/details as below.

  • Optimize music (UMG) pipeline with queries and memories for Elasticsearch and PostgreSQL, the pipeline saving 90% execution time from 10+ hours to 40 mins.
  • Migrate service from apache spark, AWS Data Lake Formation to AWS MWAA, EKS airflow environment. 
  • Design, and deliver distributed system for Ray Serve with AI team.
  • Design, and implement a modern machine learning pipeline for a recommendation, and moderation pipe.
  • Design SLA and implement alert log reporting system (history logs) for moderation pipeline, histories logs handling application, server levels information for further investigation.
  • Supporting other departments to gather data in the appropriate platforms.
Tech Stacks : 
  • Streaming, Snowpipe/Kinesis/Firehose
  • Monitoring, CloudWatch/Grafana
  • Orchestration, AWS MWAA / Airflow
  • Kubernetes, EKS
  • Message, SQS/SNS
  • MLflow, Ray Serve/EMR/Lambda
  • Storage, Snowflake / RDS (PostgreSQL) / ElastiCache (Redis) / Elastic search
  • Bucket, AWS S3
Reports to : VP of Data Engineering


Senior Data Engineer / Handshakes by DC Frontiers / Singapore

Oct. 2020 - May. 2021.

Description and Responsibilities: The main responsibility of the engineering team is launching ScoutAsia by Nikkei and The Financial Times Nikkei content to SGX TitanOTC's platform. Titan Users will be able to access Nikkei news articles from across 11 categories, including equities, stocks, indices, foreign exchange, and iron ore. DPP (Data team) is processing hundreds of GB articles/market/financial/relationships and organization for day-to-day operation on Azure and on-premise environments. More responsibilities/details as below.

  • Identifying, digging bottlenecks, and problem-solving especially optimizing the performance of SQL Server, NoSQL (Azure Cosmos), resource units, and message queues, reducing/saving almost 50-75% of resources. 
  • Identifying and solving the problems between machine learning/backend/frontend/DDP side and giving the advance logical/physical design of a system. Displayed technical expertise in optimizing the databases and improving the data pipeline to achieve the objective.
  • Bring in industry standards to data management to deliver data at the end objective. 
  • Building, and recruiting the new data engineering staff for the next-generation, enterprise data pipeline.
Tech Stacks : 
  • Storage, Azure Cosmos DB/Gremlin/SQL Server/MYSQL/Redis
  • Storage (Bucket), Azure Blob/AWS S3
  • Streaming/Batch/transform, Spark/Scala (90% codebase coverage)
  • Message, Azure service bus, queue storage
  • Search, Elastic search
  • Algorithm, graph/concordance
Reports to : CTO

Senior Data Engineer / 17LIVE Inc. / Taiwan, Taipei.

Feb. 2020 - Jul. 2020

Description and Responsibilities: The big challenge of 17 Media data teams is facing fast-growing data volume (processing 5-10x TB level daily), complex cooperation with stakeholders, the cost optimization of the pipeline, and refactoring big latency systems .etc. As a senior data member, I’m making a data dictionary and trying to explain/design how the whole pipeline works with each component, especially how to solve those bottlenecks. More responsibilities/details as below. 

  • Leading, and architect large-scale data pipeline for supporting scientists and shareholders. 
  • Optimize, ensure quality and play a tough role in data lake projects/data pipes. infrastructure. 
  • Define, and design stage, dimension, production, and fact tables for data warehouse (BigQuery). 
  • Coordinate with client / QA / backend team for QC lists / MongoDB change stream workers. 
  • Architect workflows with those components, Dataflow, Cloud Functions, and GCS. 
  • Recruiting (Jr./Sr.) data engineering members, setting goals, and sprint management.

Tech Stacks : 

  • Storage, GCS/BigQuery/Firebase/MongoDB/MYSQL 
  • Realtime process and Message system, DataFlow (Apache Beam) / BigQuery Streaming / MongoDB Change Stream / Fluentd / Firebase / Pub/Sub 
  • ETL/ELT workflow, Digdag / Embulk 
  • Data warehouse, Visualization, BigQuery / Superset / Chartio / Data Studio 
  • Continuous deployment, docker, CricleCI 

Reports to : Data Head

Data Engineer / Paktor Pte. Ltd. / Singapore 

Sep. 2015 - Dec. 2019.

Description and Responsibilities : This is another 0 to 1 story. As an early data member, we need to figure out the data driven policy, strategies, engineering requirements from the company. In Paktor, data / backend sides are 100% on AWS, therefore the whole data ingestion, automation and data warehouse etc. are relying on those components. We are processing 50-100x GB realtime / batch jobs and the other data sources (RDBMS, APIs) for ETL/ELT on S3, Redshift, the data platform helps our marketing / HQ scientists team getting data into insights and making good decisions. More responsibilities / details as below. 

  • Supports Big Data and batch, real-time analytical solutions leveraging transformational technologies. 
  • Optimize data pipeline on AWS using Kinesis-Firehose/Lambda/Kinesis Analytics/Data Pipeline, and optimize, resizing Redshift clusters and related scripts. 
  • Translates complex analytics requirements into detailed architecture, design, and high performing software such as machine-learning, CI/CD of recommendation pipeline. 
  • Collaborate with client / backend side developers to formulate innovative solutions to experiment and implement related algorithms. 

Tech Stacks : 

  • Storage, S3/Redshift/Aurora - Realtime process and Message system, Kinesis Firehose / SNS 
  • Data warehouse, Visualization, Redshift / Klipfolio / Metabase 
  • ETL/ELT workflow, Lambda / SNS / Batch / Python 
  • Recommendation, ML, DynamoDB / EMR / Spark / Sagemaker 
  • Metadata management, Athena (presto) / Glue / Redshift Spectrum 
  • Continuous deployment, Elasticbeanstalk / Cloudformation 
  • Operations, PagerDuty / Zapier / Cloud Watch 

Reports to : CTO, Data Head

System Analyst (Data Backend Engineer) / JSpectrum Software Limited / Hong Kong 

 Jan. 2014 - Aug 2015.

Description and Responsibilities : JSPectrum is a leading passive location-based service company in Hong Kong which holds many interesting products such as NetProbe, NetWhere, NetAd etc. In Optus (The main project in Sydney), the main responsibility of system analyst is designing / implementing data ingestion (real-time processing) / load and management data with major components of the Hadoop ecosystem. We meet the challenge to process 15,000 TPS, 60,000 inserts per second and 300 GB daily storages, therefore we are trying to optimize those components with Kafka consumers, HDFS storages and re-designing keys / columns of HBase to fulfill the requirement and deployed NetAd, whole in-house solutions on Optus. More responsibilities / details as below. 

  • Design, implement and optimize Hadoop ecosystems, MLP, real-time processing on Optus in house servers with our main product NetAd, NetWhere. We are focusing on HBase schema, HDFS, balancing Kafka consumers and more issues on data ingestion. 
  • Collaborate with shareholders and LBS team members for further requirements with HeapMap. 

Tech Stacks : 

  • Storage, HDFS / HBase
  • Realtime process and Message system, Kafka streaming, Log systems 
  • Data warehouse, Visualization, HBase / NetWhere (Dashboard) 
  • Hadoop ecosystem, Hadoop / HDFS / Zookeeper / Spark / Hive
  • ETL/ELT workflow, Spark / Hive / Scala / Java

Reports to : CTO


Senior Software Engineer / Toro Development Ltd. / Taiwan, Taipei. 

Oct. 2012 - Dec. 2013.

Description and Responsibilities : TORO is a technology business that provides a mobile platform and its associated systems, services and rules to help Brands (with initial focus on Sports Teams, Smart Cities and Streaming apps) become super-apps to generate additional revenue with minimum effort. Responsibilities as below. 

  • Design, implement and test back-office modules for NFC wallet platform, Trusted Service Managers (TSM) and distributed NFC services to end­ users / stakeholders. 
  • Implement RESTful services and deliver endpoints for wallet managers and collaborating with front­end, backend teams for further business requirements. 

Tech Stacks: MYSQL / Spring / Hibernate / XML / Apache Camel / Java / POJO .etc. 

Reports to : Head of Server Solutions


Software Engineer / Digital River / Taiwan, Taipei. 

Oct. 2011 - Sep. 2012.

Description and Responsibilities : Digital river proactive partners, providing API-based Payments & Risk, Order Management and Commerce services to leading enterprise brands. The big challenge to DR is integrating with the current module and working well with a huge code base (over 2+ millions lines), the strict process including analysis requirements, design, implement, test and code review. More responsibilities as below. 

  • Design, implement custom bundle project, bundle customized by shoppers to pick products of groups and get special discounts, the main stakeholders /users from Logitech, Microsoft. 
  • Analysis, collect business requirements, identify use cases and collaborate with business analysts and deliver related diagrams, documents. 

Tech Stacks: Oracle / Tomcat / Spring / Struts / JDO / XML / JUnit / Java / J2EE .etc. 

Reports to : Technical Development Manager


Technical Supervisor / Stark Technology Inc. / Taiwan, Taipei. 

Oct. 2008 - Sep. 2011.

Description and Responsibilities : Stark Technology (STI) is the largest domestic system integrator in Taiwan. We plan and deliver complete ICT solutions for a wide spectrum of industries through representing and reselling the world's leading products. This is made possible by using the most advanced technology, and providing the best professional services. More responsibilities / projects as below. 

  • Lead, coach JR. programmers for the development process of enterprise modules, and design Fatwire CMS components as Template/Page/Cache .etc. 
  • Design, analyze DMDB systems, and implement functions to meet the requirements of queries / storage. Optimize performance for online servers and GC tuning. 

Tech Stacks : Oracle / Sybase / Tomcat / Weblogic / Spring / Struts / Hibernate / Fatwire / Java / J2EE .etc. 

Reports to : Technical Manager


Relevant Skills and Qualifications


Big Data Tech Stacks

  • AWS Services, EC2/S3/Lambda/EMR/CloudWatch/SNS/SQS/Elastic Beanstalk 
  • AWS Big Data Solutions, Kinesis/Firehose/Athena/Redshift/Dynamodb 
  • GCP Big Data Solutions, BigQuery/PubSub/Dataflow/Cloud Functions 
  • Hadoop ecosystem, Hadoop/HDFS/Zookeeper/Hbase/Hive 
  • Spark Streaming/Apache Kafka 
  • CI/CD: Jenkins/Cloud Formation/GitLab/Grafana

Specific Skills

  • Solid, well-designed real-time streaming/batch processing, ETL systems.
  • Monitors and conducts data-pipeline / machine learning pipeline development requests through lifecycle management and ensures that the technical solution meets.
  • Diagnosing and troubleshooting Redshift and specific clusters management.
  • Development of micro-services and endpoints based on enterprise integration patterns. Knowledge over garbage collection (JVM) tuning technologies for various servers.
  • Developed multi-threading processing consuming work and managed transactions.

Certifications and Training

  • Sun Certified Web Component Developer Java 2 Platform, Enterprise Edition. 
  • Sun Certified Programmer for the Java 2 Platform. 
  • Red Hat Enterprise Directory Services and Authentication Attended. 
  • Project Management Professional (PMP)® Attended. 
  • AWS Certified Solutions Architect Attended. 
  • Big Data on AWS Attended. 
  • Azure Data Engineer AssociateAttended. 

Education


National Taiwan University, 2010 – 2011

EMBA Programs, Business Administration, Accounting, Finance and International Business.


Chinese Culture University Master of Information Management, 2002 – 2005

Computer Science, Data Mining, Expert Systems and Knowledge Base as major concentration.


Chinese Culture University, Bachelor Degree of Science in Journalism, 1998 - 2002

履歷
個人檔案
Ahzwaym2ourqm1t0glsc

Chin-Hung (Wilson) Liu

I am a lead architect responsible for designing and implementing a large-scale data pipeline for Lomotif, Paktor x 17LIVE, utilizing GCP/AWS/Python/Scala, in collaboration with data science and machine learning teams in Singapore and TW HQ, as well as with the Hadoop ecosystem (HDFS/HBase/Kafka) at JSpectrum in Hong Kong and Sydney. 


With over 15 years of experience in designing and developing Java/Scala/Python-based applications for daily operations, I bring:

● At least 8 years of experience in data analysis, pipeline design and development, and tool building as a team member. 

● In-depth knowledge of the Spark and Hadoop ecosystems, including Hadoop, HDFS, HBase, and more. 
● Strong skills in designing and developing Big Data services on AWS and GCP. 
 Extensive expertise in developing generic distributed systems, streaming processing, machine learning pipelines, and continuously improving ML models.


Senior Data Engineer at Paktor x 17LIVE| AWS Big Data Specialist | Data Architect 
Singapore / Hong Kong / Taiwan

[email protected]

https://www.linkedin.com/in/chin-hung-wilson-liu-29392957

Nanxing Rd., Xizhi Dist., New Taipei City, Taiwan (R.O.C.)

Experience 

Senior Data Engineer (DataOps / AI) / Lomotif Private Limited / Singapore

Jul. 2021 - Present.

Description and Responsibilities: Lomotif is a leading short video social platform in South America and India that holds PBs of videos in buckets and serves millions of users. DataOps and AI team take part in many challenging projects e.g. Ncanto, XROAD services, Ray Serve, and scalable model serving frameworks for support the recommendation and moderation pipeline, also integrated Universal Music Group music (UMG) and full catalog feed with 7digital. DataOps team handling 10TB+ data for day-to-day operation, moderating model training results, and designing SLIs/SLOs for EKS Clusters. More responsibilities/details as below.

  • Optimize music (UMG) pipeline with queries and memories for Elasticsearch and PostgreSQL, the pipeline saving 90% execution time from 10+ hours to 40 mins.
  • Migrate service from apache spark, AWS Data Lake Formation to AWS MWAA, EKS airflow environment. 
  • Design, and deliver distributed system for Ray Serve with AI team.
  • Design, and implement a modern machine learning pipeline for a recommendation, and moderation pipe.
  • Design SLA and implement alert log reporting system (history logs) for moderation pipeline, histories logs handling application, server levels information for further investigation.
  • Supporting other departments to gather data in the appropriate platforms.
Tech Stacks : 
  • Streaming, Snowpipe/Kinesis/Firehose
  • Monitoring, CloudWatch/Grafana
  • Orchestration, AWS MWAA / Airflow
  • Kubernetes, EKS
  • Message, SQS/SNS
  • MLflow, Ray Serve/EMR/Lambda
  • Storage, Snowflake / RDS (PostgreSQL) / ElastiCache (Redis) / Elastic search
  • Bucket, AWS S3
Reports to : VP of Data Engineering


Senior Data Engineer / Handshakes by DC Frontiers / Singapore

Oct. 2020 - May. 2021.

Description and Responsibilities: The main responsibility of the engineering team is launching ScoutAsia by Nikkei and The Financial Times Nikkei content to SGX TitanOTC's platform. Titan Users will be able to access Nikkei news articles from across 11 categories, including equities, stocks, indices, foreign exchange, and iron ore. DPP (Data team) is processing hundreds of GB articles/market/financial/relationships and organization for day-to-day operation on Azure and on-premise environments. More responsibilities/details as below.

  • Identifying, digging bottlenecks, and problem-solving especially optimizing the performance of SQL Server, NoSQL (Azure Cosmos), resource units, and message queues, reducing/saving almost 50-75% of resources. 
  • Identifying and solving the problems between machine learning/backend/frontend/DDP side and giving the advance logical/physical design of a system. Displayed technical expertise in optimizing the databases and improving the data pipeline to achieve the objective.
  • Bring in industry standards to data management to deliver data at the end objective. 
  • Building, and recruiting the new data engineering staff for the next-generation, enterprise data pipeline.
Tech Stacks : 
  • Storage, Azure Cosmos DB/Gremlin/SQL Server/MYSQL/Redis
  • Storage (Bucket), Azure Blob/AWS S3
  • Streaming/Batch/transform, Spark/Scala (90% codebase coverage)
  • Message, Azure service bus, queue storage
  • Search, Elastic search
  • Algorithm, graph/concordance
Reports to : CTO

Senior Data Engineer / 17LIVE Inc. / Taiwan, Taipei.

Feb. 2020 - Jul. 2020

Description and Responsibilities: The big challenge of 17 Media data teams is facing fast-growing data volume (processing 5-10x TB level daily), complex cooperation with stakeholders, the cost optimization of the pipeline, and refactoring big latency systems .etc. As a senior data member, I’m making a data dictionary and trying to explain/design how the whole pipeline works with each component, especially how to solve those bottlenecks. More responsibilities/details as below. 

  • Leading, and architect large-scale data pipeline for supporting scientists and shareholders. 
  • Optimize, ensure quality and play a tough role in data lake projects/data pipes. infrastructure. 
  • Define, and design stage, dimension, production, and fact tables for data warehouse (BigQuery). 
  • Coordinate with client / QA / backend team for QC lists / MongoDB change stream workers. 
  • Architect workflows with those components, Dataflow, Cloud Functions, and GCS. 
  • Recruiting (Jr./Sr.) data engineering members, setting goals, and sprint management.

Tech Stacks : 

  • Storage, GCS/BigQuery/Firebase/MongoDB/MYSQL 
  • Realtime process and Message system, DataFlow (Apache Beam) / BigQuery Streaming / MongoDB Change Stream / Fluentd / Firebase / Pub/Sub 
  • ETL/ELT workflow, Digdag / Embulk 
  • Data warehouse, Visualization, BigQuery / Superset / Chartio / Data Studio 
  • Continuous deployment, docker, CricleCI 

Reports to : Data Head

Data Engineer / Paktor Pte. Ltd. / Singapore 

Sep. 2015 - Dec. 2019.

Description and Responsibilities : This is another 0 to 1 story. As an early data member, we need to figure out the data driven policy, strategies, engineering requirements from the company. In Paktor, data / backend sides are 100% on AWS, therefore the whole data ingestion, automation and data warehouse etc. are relying on those components. We are processing 50-100x GB realtime / batch jobs and the other data sources (RDBMS, APIs) for ETL/ELT on S3, Redshift, the data platform helps our marketing / HQ scientists team getting data into insights and making good decisions. More responsibilities / details as below. 

  • Supports Big Data and batch, real-time analytical solutions leveraging transformational technologies. 
  • Optimize data pipeline on AWS using Kinesis-Firehose/Lambda/Kinesis Analytics/Data Pipeline, and optimize, resizing Redshift clusters and related scripts. 
  • Translates complex analytics requirements into detailed architecture, design, and high performing software such as machine-learning, CI/CD of recommendation pipeline. 
  • Collaborate with client / backend side developers to formulate innovative solutions to experiment and implement related algorithms. 

Tech Stacks : 

  • Storage, S3/Redshift/Aurora - Realtime process and Message system, Kinesis Firehose / SNS 
  • Data warehouse, Visualization, Redshift / Klipfolio / Metabase 
  • ETL/ELT workflow, Lambda / SNS / Batch / Python 
  • Recommendation, ML, DynamoDB / EMR / Spark / Sagemaker 
  • Metadata management, Athena (presto) / Glue / Redshift Spectrum 
  • Continuous deployment, Elasticbeanstalk / Cloudformation 
  • Operations, PagerDuty / Zapier / Cloud Watch 

Reports to : CTO, Data Head

System Analyst (Data Backend Engineer) / JSpectrum Software Limited / Hong Kong 

 Jan. 2014 - Aug 2015.

Description and Responsibilities : JSPectrum is a leading passive location-based service company in Hong Kong which holds many interesting products such as NetProbe, NetWhere, NetAd etc. In Optus (The main project in Sydney), the main responsibility of system analyst is designing / implementing data ingestion (real-time processing) / load and management data with major components of the Hadoop ecosystem. We meet the challenge to process 15,000 TPS, 60,000 inserts per second and 300 GB daily storages, therefore we are trying to optimize those components with Kafka consumers, HDFS storages and re-designing keys / columns of HBase to fulfill the requirement and deployed NetAd, whole in-house solutions on Optus. More responsibilities / details as below. 

  • Design, implement and optimize Hadoop ecosystems, MLP, real-time processing on Optus in house servers with our main product NetAd, NetWhere. We are focusing on HBase schema, HDFS, balancing Kafka consumers and more issues on data ingestion. 
  • Collaborate with shareholders and LBS team members for further requirements with HeapMap. 

Tech Stacks : 

  • Storage, HDFS / HBase
  • Realtime process and Message system, Kafka streaming, Log systems 
  • Data warehouse, Visualization, HBase / NetWhere (Dashboard) 
  • Hadoop ecosystem, Hadoop / HDFS / Zookeeper / Spark / Hive
  • ETL/ELT workflow, Spark / Hive / Scala / Java

Reports to : CTO


Senior Software Engineer / Toro Development Ltd. / Taiwan, Taipei. 

Oct. 2012 - Dec. 2013.

Description and Responsibilities : TORO is a technology business that provides a mobile platform and its associated systems, services and rules to help Brands (with initial focus on Sports Teams, Smart Cities and Streaming apps) become super-apps to generate additional revenue with minimum effort. Responsibilities as below. 

  • Design, implement and test back-office modules for NFC wallet platform, Trusted Service Managers (TSM) and distributed NFC services to end­ users / stakeholders. 
  • Implement RESTful services and deliver endpoints for wallet managers and collaborating with front­end, backend teams for further business requirements. 

Tech Stacks: MYSQL / Spring / Hibernate / XML / Apache Camel / Java / POJO .etc. 

Reports to : Head of Server Solutions


Software Engineer / Digital River / Taiwan, Taipei. 

Oct. 2011 - Sep. 2012.

Description and Responsibilities : Digital river proactive partners, providing API-based Payments & Risk, Order Management and Commerce services to leading enterprise brands. The big challenge to DR is integrating with the current module and working well with a huge code base (over 2+ millions lines), the strict process including analysis requirements, design, implement, test and code review. More responsibilities as below. 

  • Design, implement custom bundle project, bundle customized by shoppers to pick products of groups and get special discounts, the main stakeholders /users from Logitech, Microsoft. 
  • Analysis, collect business requirements, identify use cases and collaborate with business analysts and deliver related diagrams, documents. 

Tech Stacks: Oracle / Tomcat / Spring / Struts / JDO / XML / JUnit / Java / J2EE .etc. 

Reports to : Technical Development Manager


Technical Supervisor / Stark Technology Inc. / Taiwan, Taipei. 

Oct. 2008 - Sep. 2011.

Description and Responsibilities : Stark Technology (STI) is the largest domestic system integrator in Taiwan. We plan and deliver complete ICT solutions for a wide spectrum of industries through representing and reselling the world's leading products. This is made possible by using the most advanced technology, and providing the best professional services. More responsibilities / projects as below. 

  • Lead, coach JR. programmers for the development process of enterprise modules, and design Fatwire CMS components as Template/Page/Cache .etc. 
  • Design, analyze DMDB systems, and implement functions to meet the requirements of queries / storage. Optimize performance for online servers and GC tuning. 

Tech Stacks : Oracle / Sybase / Tomcat / Weblogic / Spring / Struts / Hibernate / Fatwire / Java / J2EE .etc. 

Reports to : Technical Manager


Relevant Skills and Qualifications


Big Data Tech Stacks

  • AWS Services, EC2/S3/Lambda/EMR/CloudWatch/SNS/SQS/Elastic Beanstalk 
  • AWS Big Data Solutions, Kinesis/Firehose/Athena/Redshift/Dynamodb 
  • GCP Big Data Solutions, BigQuery/PubSub/Dataflow/Cloud Functions 
  • Hadoop ecosystem, Hadoop/HDFS/Zookeeper/Hbase/Hive 
  • Spark Streaming/Apache Kafka 
  • CI/CD: Jenkins/Cloud Formation/GitLab/Grafana

Specific Skills

  • Solid, well-designed real-time streaming/batch processing, ETL systems.
  • Monitors and conducts data-pipeline / machine learning pipeline development requests through lifecycle management and ensures that the technical solution meets.
  • Diagnosing and troubleshooting Redshift and specific clusters management.
  • Development of micro-services and endpoints based on enterprise integration patterns. Knowledge over garbage collection (JVM) tuning technologies for various servers.
  • Developed multi-threading processing consuming work and managed transactions.

Certifications and Training

  • Sun Certified Web Component Developer Java 2 Platform, Enterprise Edition. 
  • Sun Certified Programmer for the Java 2 Platform. 
  • Red Hat Enterprise Directory Services and Authentication Attended. 
  • Project Management Professional (PMP)® Attended. 
  • AWS Certified Solutions Architect Attended. 
  • Big Data on AWS Attended. 
  • Azure Data Engineer AssociateAttended. 

Education


National Taiwan University, 2010 – 2011

EMBA Programs, Business Administration, Accounting, Finance and International Business.


Chinese Culture University Master of Information Management, 2002 – 2005

Computer Science, Data Mining, Expert Systems and Knowledge Base as major concentration.


Chinese Culture University, Bachelor Degree of Science in Journalism, 1998 - 2002