CakeResume 找人才

进阶搜寻
On
4 到 6 年
6 到 10 年
10 到 15 年
15 年以上
Avatar of P.Koteswar.
Avatar of P.Koteswar.
Cloud Services Manager @AIA
2021 ~ 2022
DevOps Engineer, Site Reliability Engineer
一個月內
DevOps & SRE Jenkins, ArgoCD, Docker, Packers Git , GitLabs , BitBucket, GitOps Maven, Ansible, Jforg, Chef, Vagrant, Msbuild, Build and Release, Deployment Strategies, Management , Vantage, Infra Cost Optimisation. AWS : EC2, S3, IAM, Route53, VPC, Autoscaling, EBS, ELB, ElasticSearch, CloudWatch, RDS, Lambda, EKS, MSK, AWS Secrets , Lambda, AWS DevOps, Kinesis , GaurdDuty, Boto3, AWS CLI Monitoring & Logging Prometheus , AlertManager Grafana Dashboards, AWS CloudWatch, Azure LogAnalytics Azure Monitor, Kibana, Thanos, Graylog , Opsgenie , Slack Network & Security Istio Service Mesh, SSL and TLS Certs Network Firewalls Ingress & Egress Controllers Routing, TGW and VPC Peering Infra As Code Terraform Terragrunt Go Lang ARM Templates, AWS CloudFormation Database
DevOps / CI / CD
Site Reliability Engineering
Terraform/Ansible/Jenkins
就职中
正在积极求职中
全职 / 对远端工作有兴趣
10 到 15 年
Sikkim Manipal University
Information Technology
Avatar of 黃季承.
Avatar of 黃季承.
曾任
後端工程師 & DevOps @創業家兄弟Kuobrothers Corp.
2022 ~ 2024
Senior Backend Engineer | DevOps | SRE
一個月內
協助紅隊演練漏洞修復與資安工具導入 POC [工作經驗] 1. 使用 Drone CI 搭配 GitOps 建立自動化 CI/CD 流程 2. 使用 IaC (AWS CloudFormation) 管理 AWS 雲端資源 3. 使用 Cloudflare CDN 功能來保護後方雲端服務 4. 協助修復資安漏洞並使用 CrowdStrike 進行異常行為監測 [使用技能] AWS: VPC
AWS
CI/CD Drone
Cloudflare
待业中
正在积极求职中
全职 / 对远端工作有兴趣
4 到 6 年
National Taipei University of Technology
資工系
Avatar of 鄭一翔.
Avatar of 鄭一翔.
Director @朕來雲端工作室
2023 ~ 现在
管理職
一個月內
Jenkins 3.Docker 4.Kubernetes 測試工具 1.POSTMAN 2.cucumber 3.Mocha 4. SonarQube 5.OWASP ZAP SQL & NoSQL 1. MySQL 2. PostgreSQL 3. Sqlite 4. MongoDB 5. Redis 雲端服務 1.AWS 2.Azure 3.GCP 4.Alexa 5.FireBase 系統 1. Arch Linux 2. Ubuntu Linux 3. Redhat Linux 4. Windows Server 工作經歷 朕來雲端工作室 , Director, Aug 2023 ~ 現在 工作室
AWS
Node.js
PHP
就职中
正在积极求职中
全职 / 对远端工作有兴趣
15 年以上
輔仁大學
資訊管理
Avatar of the user.
Avatar of the user.
Consultant @Global Advanced IT Connections
2024 ~ 2024
Product Owner, Business Analyst, System Analyst
一個月內
YAML
AWS
Azure
就职中
正在积极求职中
全职 / 对远端工作有兴趣
6 到 10 年
Adamson University
Information Technology
Avatar of the user.
Avatar of the user.
Frontend Architect / Lead @Elemental
2021 ~ 2023
Frontend Developer
一個月內
PHP
NoSQL
AWS
就职中
正在积极求职中
全职 / 我只想远端工作
6 到 10 年
ETS "Nikola Tesla"
IT
Avatar of Yao.
Avatar of Yao.
Chief Architect and Director @Innotech ME
2022 ~ 现在
Development Director or Head of Product
一個月內
cache by last updated date to get latest data, not to retrieve entire redundant old data. 2. Build client-server socket between application level and actual service to respond data via AsyncResponse. ˙Framework Spring Boot, Spring Cloud, Eureka, Nacos, Redis, Kafka , ELK ˙Container Docker ˙Production running environment AWS, GCP ˙Database Oracle, MySql, PostgrelSQL ˙Tool Maven, Git, Svn, Ant ˙IDE Intellij , Eclipse Certification SCJP (Sun Certified Programmer for Java 4) Taipei City, Taiwan 專案 Project Sports Exchange/Book Sports : Cricket/Soccer/Tennis Game category : Exchange, SportsBook, FancyBet, BookMaker, Casino, BPoker, Binary Concurrent users : 250K
JAVA
Linux
AWS
就职中
正在积极求职中
全职 / 对远端工作有兴趣
15 年以上
Chung Hua University
Avatar of Adam Prasetya.
Avatar of Adam Prasetya.
Full Stack JS @KARLO
2021 ~ 现在
Software Engineer
兩個月內
MERN Stack with fully integration on Mapbox, Here, Tollguru & IOT GPS Tracker. Dev. Backend API Service using Nodejs and mongoose . Dev. Integration system Map using Mapbox, Tollguru and Here. Dev. Platform Menu Agreement, Order, Invoice, Report using React. [ Reactjs, Nodejs/Express, Mongodb, AWS, Mapbox] Full Stack Engineer RebelWorks SeptemberAugust 2021 Jakarta, Indonesia products Campaign and Voucher & transaction from Shopping carts into Payment Gateway. Create Notifications (Callback) from midtrans (integrations payment notifications) Integrations with Bellesomics APIs for update and get SKIN result patient/customer using Guzzle Http. Handle Send to
Laravel
Node.js / Express.js
ReactJS
就职中
正在积极求职中
全职 / 对远端工作有兴趣
6 到 10 年
University of Brawijaya
Information Technology
Avatar of 陶俊良.
Avatar of 陶俊良.
資料分析師 Data Analyst @Portto 門戶科技| Blocto
2022 ~ 2024
Data Analyst、Data Engineer、Data Scientist、Customer Experience Analyst
一個月內
陶俊良 (Tao,Chun-Liang) Taipei, Taiwan Email: [email protected] Phone:I am very sensitive to data and enjoy finding inspiration and ideas from them. I am proficient in machine learning, text analysis, and recommendation systems, EVM blockchain analytics, and currently use Python as my primary programming languages. I am always open to learning new things, such as learning new data structure from blockchain. I am currently very interested in blockchain data and on-chain user segamentation. I was working in digital media, advertising (DSP, SSP, DMP platforms), gaming user analyst, blockchain
python
R
MySQL
就职中
正在积极求职中
全职 / 对远端工作有兴趣
4 到 6 年
臺灣大學
流行病學與預防醫學所 生物統計組
Avatar of the user.
Avatar of the user.
曾任
Senior Back End Software Developer @PT Profeed Social Media Management
2023 ~ 2024
Front-End / Back-End / Full Stack Web Developer / Software Engineer
一個月內
MySQL Database
PHP
Git
待业中
正在积极求职中
全职 / 对远端工作有兴趣
6 到 10 年
Universitas Pembangunan Nasional Veteran Jawa Timur
Informatics Engineering
Avatar of 陳韋燁.
Avatar of 陳韋燁.
曾任
工程師 @博彥科技有限公司
2018 ~ 2023
後端工程師
一個月內
程,讓外籍員工也能輕易上手 巡檢系統 簡化巡檢相關流程,降低人工疏失造成風險 學歷 龍華科技大學 資訊網路工程 •Lorem ipsum dolor sit amet, consectetuer adipiscing elit, sed diam nonummy nibh euismod tincidunt ut laoreet dolore magna aliquam erat volutpat. 技能 C、C++、golang lua、linux shell script mysql、postgresql、redis linux aws、docker、jenkis rabbitmq http、mqtt、grpc i2c、uart、spi 語言 English — 中階
C
C++
Golang
待业中
正在积极求职中
全职 / 对远端工作有兴趣
4 到 6 年
龍華科技大學
資訊網路工程

最轻量、快速的招募方案,数百家企业的选择

搜寻简历,主动联系求职者,提升招募效率。

  • 浏览所有搜寻结果
  • 每日可无限次数开启陌生对话
  • 搜尋僅開放付費企業檢視的简历
  • 检视使用者信箱 & 电话
搜寻技巧
1
Search a precise keyword combination
senior backend php
If the number of the search result is not enough, you can remove the less important keywords
2
Use quotes to search for an exact phrase
"business development"
3
Use the minus sign to eliminate results containing certain words
UI designer -UX
免费方案仅能搜寻公开简历。
升级至进阶方案,即可浏览所有搜寻结果(包含数万笔览仅在 CakeResume 平台上公开的简历)。

职场能力评价定义

专业技能
该领域中具备哪些专业能力(例如熟悉 SEO 操作,且会使用相关工具)。
问题解决能力
能洞察、分析问题,并拟定方案有效解决问题。
变通能力
遇到突发事件能冷静应对,并随时调整专案、客户、技术的相对优先序。
沟通能力
有效传达个人想法,且愿意倾听他人意见并给予反馈。
时间管理能力
了解工作项目的优先顺序,有效运用时间,准时完成工作内容。
团队合作能力
具有向心力与团队责任感,愿意倾听他人意见并主动沟通协调。
领导力
专注于团队发展,有效引领团队采取行动,达成共同目标。
一個月內
Senior Data Engineer at Paktor x M17 Entertainment Group | AWS x GCP x Azure Big Data Specialist | Data Architect
Logo of KKCompany.
KKCompany
2023 ~ 现在
Taiwan
专业背景
目前状态
就职中
求职阶段
目前会考虑了解新的机会
专业
数据工程师, 后端开发人员
产业
软件
工作年资
10 到 15 年
管理经历
我有管理 1~5 人的经验
技能
Big Data
Data Engineering
ETL
AWS
GCP
Python
BigQuery
Data Warehouse
Data Pipeline
Java
Azure
SQL
Spark
kafka
spark streaming
Scala
Redshift
HBase
SQL Server
AWS S3
AWS Lambda
MongoDB
Hadoop
Hadoop Distributed File System
AWS SQS
Azure Storage
MySQL
PostgreSQL
Postman for API
Snowflake
语言能力
English
专业
Chinese
母语或双语
求职偏好
希望获得的职位
Backend Engineer, Data Engineer, MLOps Engineer
预期工作模式
全职
期望的工作地点
Taiwan, 台灣, Singapore, Hong Kong
远端工作意愿
对远端工作有兴趣
接案服务
学历
学校
National Taiwan University
主修科系
EMBA Programs, Business Administration, Accounting, Finance and International Business.
列印
Ahzwaym2ourqm1t0glsc

Chin-Hung (Wilson) Liu

I am a lead architect responsible for designing and implementing a large-scale data pipeline for Lomotif, Paktor x 17LIVE, utilizing GCP/AWS/Python/Scala, in collaboration with data science and machine learning teams in Singapore and TW HQ, as well as with the Hadoop ecosystem (HDFS/HBase/Kafka) at JSpectrum in Hong Kong and Sydney. 


With over 15 years of experience in designing and developing Java/Scala/Python-based applications for daily operations, I bring:

● At least 8 years of experience in data analysis, pipeline design and development, and tool building as a team member. 

● In-depth knowledge of the Spark and Hadoop ecosystems, including Hadoop, HDFS, HBase, and more. 
● Strong skills in designing and developing Big Data services on AWS and GCP. 
 Extensive expertise in developing generic distributed systems, streaming processing, machine learning pipelines, and continuously improving ML models.


Senior Data Engineer at Paktor x 17LIVE| AWS Big Data Specialist | Data Architect 
Singapore / Hong Kong / Taiwan

[email protected]

https://www.linkedin.com/in/chin-hung-wilson-liu-29392957

Nanxing Rd., Xizhi Dist., New Taipei City, Taiwan (R.O.C.)

Experience 

Senior Data Engineer (DataOps / AI) / Lomotif Private Limited / Singapore

Jul. 2021 - Present.

Description and Responsibilities: Lomotif is a leading short video social platform in South America and India that holds PBs of videos in buckets and serves millions of users. DataOps and AI team take part in many challenging projects e.g. Ncanto, XROAD services, Ray Serve, and scalable model serving frameworks for support the recommendation and moderation pipeline, also integrated Universal Music Group music (UMG) and full catalog feed with 7digital. DataOps team handling 10TB+ data for day-to-day operation, moderating model training results, and designing SLIs/SLOs for EKS Clusters. More responsibilities/details as below.

  • Optimize music (UMG) pipeline with queries and memories for Elasticsearch and PostgreSQL, the pipeline saving 90% execution time from 10+ hours to 40 mins.
  • Migrate service from apache spark, AWS Data Lake Formation to AWS MWAA, EKS airflow environment. 
  • Design, and deliver distributed system for Ray Serve with AI team.
  • Design, and implement a modern machine learning pipeline for a recommendation, and moderation pipe.
  • Design SLA and implement alert log reporting system (history logs) for moderation pipeline, histories logs handling application, server levels information for further investigation.
  • Supporting other departments to gather data in the appropriate platforms.
Tech Stacks : 
  • Streaming, Snowpipe/Kinesis/Firehose
  • Monitoring, CloudWatch/Grafana
  • Orchestration, AWS MWAA / Airflow
  • Kubernetes, EKS
  • Message, SQS/SNS
  • MLflow, Ray Serve/EMR/Lambda
  • Storage, Snowflake / RDS (PostgreSQL) / ElastiCache (Redis) / Elastic search
  • Bucket, AWS S3
Reports to : VP of Data Engineering


Senior Data Engineer / Handshakes by DC Frontiers / Singapore

Oct. 2020 - May. 2021.

Description and Responsibilities: The main responsibility of the engineering team is launching ScoutAsia by Nikkei and The Financial Times Nikkei content to SGX TitanOTC's platform. Titan Users will be able to access Nikkei news articles from across 11 categories, including equities, stocks, indices, foreign exchange, and iron ore. DPP (Data team) is processing hundreds of GB articles/market/financial/relationships and organization for day-to-day operation on Azure and on-premise environments. More responsibilities/details as below.

  • Identifying, digging bottlenecks, and problem-solving especially optimizing the performance of SQL Server, NoSQL (Azure Cosmos), resource units, and message queues, reducing/saving almost 50-75% of resources. 
  • Identifying and solving the problems between machine learning/backend/frontend/DDP side and giving the advance logical/physical design of a system. Displayed technical expertise in optimizing the databases and improving the data pipeline to achieve the objective.
  • Bring in industry standards to data management to deliver data at the end objective. 
  • Building, and recruiting the new data engineering staff for the next-generation, enterprise data pipeline.
Tech Stacks : 
  • Storage, Azure Cosmos DB/Gremlin/SQL Server/MYSQL/Redis
  • Storage (Bucket), Azure Blob/AWS S3
  • Streaming/Batch/transform, Spark/Scala (90% codebase coverage)
  • Message, Azure service bus, queue storage
  • Search, Elastic search
  • Algorithm, graph/concordance
Reports to : CTO

Senior Data Engineer / 17LIVE Inc. / Taiwan, Taipei.

Feb. 2020 - Jul. 2020

Description and Responsibilities: The big challenge of 17 Media data teams is facing fast-growing data volume (processing 5-10x TB level daily), complex cooperation with stakeholders, the cost optimization of the pipeline, and refactoring big latency systems .etc. As a senior data member, I’m making a data dictionary and trying to explain/design how the whole pipeline works with each component, especially how to solve those bottlenecks. More responsibilities/details as below. 

  • Leading, and architect large-scale data pipeline for supporting scientists and shareholders. 
  • Optimize, ensure quality and play a tough role in data lake projects/data pipes. infrastructure. 
  • Define, and design stage, dimension, production, and fact tables for data warehouse (BigQuery). 
  • Coordinate with client / QA / backend team for QC lists / MongoDB change stream workers. 
  • Architect workflows with those components, Dataflow, Cloud Functions, and GCS. 
  • Recruiting (Jr./Sr.) data engineering members, setting goals, and sprint management.

Tech Stacks : 

  • Storage, GCS/BigQuery/Firebase/MongoDB/MYSQL 
  • Realtime process and Message system, DataFlow (Apache Beam) / BigQuery Streaming / MongoDB Change Stream / Fluentd / Firebase / Pub/Sub 
  • ETL/ELT workflow, Digdag / Embulk 
  • Data warehouse, Visualization, BigQuery / Superset / Chartio / Data Studio 
  • Continuous deployment, docker, CricleCI 

Reports to : Data Head

Data Engineer / Paktor Pte. Ltd. / Singapore 

Sep. 2015 - Dec. 2019.

Description and Responsibilities : This is another 0 to 1 story. As an early data member, we need to figure out the data driven policy, strategies, engineering requirements from the company. In Paktor, data / backend sides are 100% on AWS, therefore the whole data ingestion, automation and data warehouse etc. are relying on those components. We are processing 50-100x GB realtime / batch jobs and the other data sources (RDBMS, APIs) for ETL/ELT on S3, Redshift, the data platform helps our marketing / HQ scientists team getting data into insights and making good decisions. More responsibilities / details as below. 

  • Supports Big Data and batch, real-time analytical solutions leveraging transformational technologies. 
  • Optimize data pipeline on AWS using Kinesis-Firehose/Lambda/Kinesis Analytics/Data Pipeline, and optimize, resizing Redshift clusters and related scripts. 
  • Translates complex analytics requirements into detailed architecture, design, and high performing software such as machine-learning, CI/CD of recommendation pipeline. 
  • Collaborate with client / backend side developers to formulate innovative solutions to experiment and implement related algorithms. 

Tech Stacks : 

  • Storage, S3/Redshift/Aurora - Realtime process and Message system, Kinesis Firehose / SNS 
  • Data warehouse, Visualization, Redshift / Klipfolio / Metabase 
  • ETL/ELT workflow, Lambda / SNS / Batch / Python 
  • Recommendation, ML, DynamoDB / EMR / Spark / Sagemaker 
  • Metadata management, Athena (presto) / Glue / Redshift Spectrum 
  • Continuous deployment, Elasticbeanstalk / Cloudformation 
  • Operations, PagerDuty / Zapier / Cloud Watch 

Reports to : CTO, Data Head

System Analyst (Data Backend Engineer) / JSpectrum Software Limited / Hong Kong 

 Jan. 2014 - Aug 2015.

Description and Responsibilities : JSPectrum is a leading passive location-based service company in Hong Kong which holds many interesting products such as NetProbe, NetWhere, NetAd etc. In Optus (The main project in Sydney), the main responsibility of system analyst is designing / implementing data ingestion (real-time processing) / load and management data with major components of the Hadoop ecosystem. We meet the challenge to process 15,000 TPS, 60,000 inserts per second and 300 GB daily storages, therefore we are trying to optimize those components with Kafka consumers, HDFS storages and re-designing keys / columns of HBase to fulfill the requirement and deployed NetAd, whole in-house solutions on Optus. More responsibilities / details as below. 

  • Design, implement and optimize Hadoop ecosystems, MLP, real-time processing on Optus in house servers with our main product NetAd, NetWhere. We are focusing on HBase schema, HDFS, balancing Kafka consumers and more issues on data ingestion. 
  • Collaborate with shareholders and LBS team members for further requirements with HeapMap. 

Tech Stacks : 

  • Storage, HDFS / HBase
  • Realtime process and Message system, Kafka streaming, Log systems 
  • Data warehouse, Visualization, HBase / NetWhere (Dashboard) 
  • Hadoop ecosystem, Hadoop / HDFS / Zookeeper / Spark / Hive
  • ETL/ELT workflow, Spark / Hive / Scala / Java

Reports to : CTO


Senior Software Engineer / Toro Development Ltd. / Taiwan, Taipei. 

Oct. 2012 - Dec. 2013.

Description and Responsibilities : TORO is a technology business that provides a mobile platform and its associated systems, services and rules to help Brands (with initial focus on Sports Teams, Smart Cities and Streaming apps) become super-apps to generate additional revenue with minimum effort. Responsibilities as below. 

  • Design, implement and test back-office modules for NFC wallet platform, Trusted Service Managers (TSM) and distributed NFC services to end­ users / stakeholders. 
  • Implement RESTful services and deliver endpoints for wallet managers and collaborating with front­end, backend teams for further business requirements. 

Tech Stacks: MYSQL / Spring / Hibernate / XML / Apache Camel / Java / POJO .etc. 

Reports to : Head of Server Solutions


Software Engineer / Digital River / Taiwan, Taipei. 

Oct. 2011 - Sep. 2012.

Description and Responsibilities : Digital river proactive partners, providing API-based Payments & Risk, Order Management and Commerce services to leading enterprise brands. The big challenge to DR is integrating with the current module and working well with a huge code base (over 2+ millions lines), the strict process including analysis requirements, design, implement, test and code review. More responsibilities as below. 

  • Design, implement custom bundle project, bundle customized by shoppers to pick products of groups and get special discounts, the main stakeholders /users from Logitech, Microsoft. 
  • Analysis, collect business requirements, identify use cases and collaborate with business analysts and deliver related diagrams, documents. 

Tech Stacks: Oracle / Tomcat / Spring / Struts / JDO / XML / JUnit / Java / J2EE .etc. 

Reports to : Technical Development Manager


Technical Supervisor / Stark Technology Inc. / Taiwan, Taipei. 

Oct. 2008 - Sep. 2011.

Description and Responsibilities : Stark Technology (STI) is the largest domestic system integrator in Taiwan. We plan and deliver complete ICT solutions for a wide spectrum of industries through representing and reselling the world's leading products. This is made possible by using the most advanced technology, and providing the best professional services. More responsibilities / projects as below. 

  • Lead, coach JR. programmers for the development process of enterprise modules, and design Fatwire CMS components as Template/Page/Cache .etc. 
  • Design, analyze DMDB systems, and implement functions to meet the requirements of queries / storage. Optimize performance for online servers and GC tuning. 

Tech Stacks : Oracle / Sybase / Tomcat / Weblogic / Spring / Struts / Hibernate / Fatwire / Java / J2EE .etc. 

Reports to : Technical Manager


Relevant Skills and Qualifications


Big Data Tech Stacks

  • AWS Services, EC2/S3/Lambda/EMR/CloudWatch/SNS/SQS/Elastic Beanstalk 
  • AWS Big Data Solutions, Kinesis/Firehose/Athena/Redshift/Dynamodb 
  • GCP Big Data Solutions, BigQuery/PubSub/Dataflow/Cloud Functions 
  • Hadoop ecosystem, Hadoop/HDFS/Zookeeper/Hbase/Hive 
  • Spark Streaming/Apache Kafka 
  • CI/CD: Jenkins/Cloud Formation/GitLab/Grafana

Specific Skills

  • Solid, well-designed real-time streaming/batch processing, ETL systems.
  • Monitors and conducts data-pipeline / machine learning pipeline development requests through lifecycle management and ensures that the technical solution meets.
  • Diagnosing and troubleshooting Redshift and specific clusters management.
  • Development of micro-services and endpoints based on enterprise integration patterns. Knowledge over garbage collection (JVM) tuning technologies for various servers.
  • Developed multi-threading processing consuming work and managed transactions.

Certifications and Training

  • Sun Certified Web Component Developer Java 2 Platform, Enterprise Edition. 
  • Sun Certified Programmer for the Java 2 Platform. 
  • Red Hat Enterprise Directory Services and Authentication Attended. 
  • Project Management Professional (PMP)® Attended. 
  • AWS Certified Solutions Architect Attended. 
  • Big Data on AWS Attended. 
  • Azure Data Engineer AssociateAttended. 

Education


National Taiwan University, 2010 – 2011

EMBA Programs, Business Administration, Accounting, Finance and International Business.


Chinese Culture University Master of Information Management, 2002 – 2005

Computer Science, Data Mining, Expert Systems and Knowledge Base as major concentration.


Chinese Culture University, Bachelor Degree of Science in Journalism, 1998 - 2002

简历
个人档案
Ahzwaym2ourqm1t0glsc

Chin-Hung (Wilson) Liu

I am a lead architect responsible for designing and implementing a large-scale data pipeline for Lomotif, Paktor x 17LIVE, utilizing GCP/AWS/Python/Scala, in collaboration with data science and machine learning teams in Singapore and TW HQ, as well as with the Hadoop ecosystem (HDFS/HBase/Kafka) at JSpectrum in Hong Kong and Sydney. 


With over 15 years of experience in designing and developing Java/Scala/Python-based applications for daily operations, I bring:

● At least 8 years of experience in data analysis, pipeline design and development, and tool building as a team member. 

● In-depth knowledge of the Spark and Hadoop ecosystems, including Hadoop, HDFS, HBase, and more. 
● Strong skills in designing and developing Big Data services on AWS and GCP. 
 Extensive expertise in developing generic distributed systems, streaming processing, machine learning pipelines, and continuously improving ML models.


Senior Data Engineer at Paktor x 17LIVE| AWS Big Data Specialist | Data Architect 
Singapore / Hong Kong / Taiwan

[email protected]

https://www.linkedin.com/in/chin-hung-wilson-liu-29392957

Nanxing Rd., Xizhi Dist., New Taipei City, Taiwan (R.O.C.)

Experience 

Senior Data Engineer (DataOps / AI) / Lomotif Private Limited / Singapore

Jul. 2021 - Present.

Description and Responsibilities: Lomotif is a leading short video social platform in South America and India that holds PBs of videos in buckets and serves millions of users. DataOps and AI team take part in many challenging projects e.g. Ncanto, XROAD services, Ray Serve, and scalable model serving frameworks for support the recommendation and moderation pipeline, also integrated Universal Music Group music (UMG) and full catalog feed with 7digital. DataOps team handling 10TB+ data for day-to-day operation, moderating model training results, and designing SLIs/SLOs for EKS Clusters. More responsibilities/details as below.

  • Optimize music (UMG) pipeline with queries and memories for Elasticsearch and PostgreSQL, the pipeline saving 90% execution time from 10+ hours to 40 mins.
  • Migrate service from apache spark, AWS Data Lake Formation to AWS MWAA, EKS airflow environment. 
  • Design, and deliver distributed system for Ray Serve with AI team.
  • Design, and implement a modern machine learning pipeline for a recommendation, and moderation pipe.
  • Design SLA and implement alert log reporting system (history logs) for moderation pipeline, histories logs handling application, server levels information for further investigation.
  • Supporting other departments to gather data in the appropriate platforms.
Tech Stacks : 
  • Streaming, Snowpipe/Kinesis/Firehose
  • Monitoring, CloudWatch/Grafana
  • Orchestration, AWS MWAA / Airflow
  • Kubernetes, EKS
  • Message, SQS/SNS
  • MLflow, Ray Serve/EMR/Lambda
  • Storage, Snowflake / RDS (PostgreSQL) / ElastiCache (Redis) / Elastic search
  • Bucket, AWS S3
Reports to : VP of Data Engineering


Senior Data Engineer / Handshakes by DC Frontiers / Singapore

Oct. 2020 - May. 2021.

Description and Responsibilities: The main responsibility of the engineering team is launching ScoutAsia by Nikkei and The Financial Times Nikkei content to SGX TitanOTC's platform. Titan Users will be able to access Nikkei news articles from across 11 categories, including equities, stocks, indices, foreign exchange, and iron ore. DPP (Data team) is processing hundreds of GB articles/market/financial/relationships and organization for day-to-day operation on Azure and on-premise environments. More responsibilities/details as below.

  • Identifying, digging bottlenecks, and problem-solving especially optimizing the performance of SQL Server, NoSQL (Azure Cosmos), resource units, and message queues, reducing/saving almost 50-75% of resources. 
  • Identifying and solving the problems between machine learning/backend/frontend/DDP side and giving the advance logical/physical design of a system. Displayed technical expertise in optimizing the databases and improving the data pipeline to achieve the objective.
  • Bring in industry standards to data management to deliver data at the end objective. 
  • Building, and recruiting the new data engineering staff for the next-generation, enterprise data pipeline.
Tech Stacks : 
  • Storage, Azure Cosmos DB/Gremlin/SQL Server/MYSQL/Redis
  • Storage (Bucket), Azure Blob/AWS S3
  • Streaming/Batch/transform, Spark/Scala (90% codebase coverage)
  • Message, Azure service bus, queue storage
  • Search, Elastic search
  • Algorithm, graph/concordance
Reports to : CTO

Senior Data Engineer / 17LIVE Inc. / Taiwan, Taipei.

Feb. 2020 - Jul. 2020

Description and Responsibilities: The big challenge of 17 Media data teams is facing fast-growing data volume (processing 5-10x TB level daily), complex cooperation with stakeholders, the cost optimization of the pipeline, and refactoring big latency systems .etc. As a senior data member, I’m making a data dictionary and trying to explain/design how the whole pipeline works with each component, especially how to solve those bottlenecks. More responsibilities/details as below. 

  • Leading, and architect large-scale data pipeline for supporting scientists and shareholders. 
  • Optimize, ensure quality and play a tough role in data lake projects/data pipes. infrastructure. 
  • Define, and design stage, dimension, production, and fact tables for data warehouse (BigQuery). 
  • Coordinate with client / QA / backend team for QC lists / MongoDB change stream workers. 
  • Architect workflows with those components, Dataflow, Cloud Functions, and GCS. 
  • Recruiting (Jr./Sr.) data engineering members, setting goals, and sprint management.

Tech Stacks : 

  • Storage, GCS/BigQuery/Firebase/MongoDB/MYSQL 
  • Realtime process and Message system, DataFlow (Apache Beam) / BigQuery Streaming / MongoDB Change Stream / Fluentd / Firebase / Pub/Sub 
  • ETL/ELT workflow, Digdag / Embulk 
  • Data warehouse, Visualization, BigQuery / Superset / Chartio / Data Studio 
  • Continuous deployment, docker, CricleCI 

Reports to : Data Head

Data Engineer / Paktor Pte. Ltd. / Singapore 

Sep. 2015 - Dec. 2019.

Description and Responsibilities : This is another 0 to 1 story. As an early data member, we need to figure out the data driven policy, strategies, engineering requirements from the company. In Paktor, data / backend sides are 100% on AWS, therefore the whole data ingestion, automation and data warehouse etc. are relying on those components. We are processing 50-100x GB realtime / batch jobs and the other data sources (RDBMS, APIs) for ETL/ELT on S3, Redshift, the data platform helps our marketing / HQ scientists team getting data into insights and making good decisions. More responsibilities / details as below. 

  • Supports Big Data and batch, real-time analytical solutions leveraging transformational technologies. 
  • Optimize data pipeline on AWS using Kinesis-Firehose/Lambda/Kinesis Analytics/Data Pipeline, and optimize, resizing Redshift clusters and related scripts. 
  • Translates complex analytics requirements into detailed architecture, design, and high performing software such as machine-learning, CI/CD of recommendation pipeline. 
  • Collaborate with client / backend side developers to formulate innovative solutions to experiment and implement related algorithms. 

Tech Stacks : 

  • Storage, S3/Redshift/Aurora - Realtime process and Message system, Kinesis Firehose / SNS 
  • Data warehouse, Visualization, Redshift / Klipfolio / Metabase 
  • ETL/ELT workflow, Lambda / SNS / Batch / Python 
  • Recommendation, ML, DynamoDB / EMR / Spark / Sagemaker 
  • Metadata management, Athena (presto) / Glue / Redshift Spectrum 
  • Continuous deployment, Elasticbeanstalk / Cloudformation 
  • Operations, PagerDuty / Zapier / Cloud Watch 

Reports to : CTO, Data Head

System Analyst (Data Backend Engineer) / JSpectrum Software Limited / Hong Kong 

 Jan. 2014 - Aug 2015.

Description and Responsibilities : JSPectrum is a leading passive location-based service company in Hong Kong which holds many interesting products such as NetProbe, NetWhere, NetAd etc. In Optus (The main project in Sydney), the main responsibility of system analyst is designing / implementing data ingestion (real-time processing) / load and management data with major components of the Hadoop ecosystem. We meet the challenge to process 15,000 TPS, 60,000 inserts per second and 300 GB daily storages, therefore we are trying to optimize those components with Kafka consumers, HDFS storages and re-designing keys / columns of HBase to fulfill the requirement and deployed NetAd, whole in-house solutions on Optus. More responsibilities / details as below. 

  • Design, implement and optimize Hadoop ecosystems, MLP, real-time processing on Optus in house servers with our main product NetAd, NetWhere. We are focusing on HBase schema, HDFS, balancing Kafka consumers and more issues on data ingestion. 
  • Collaborate with shareholders and LBS team members for further requirements with HeapMap. 

Tech Stacks : 

  • Storage, HDFS / HBase
  • Realtime process and Message system, Kafka streaming, Log systems 
  • Data warehouse, Visualization, HBase / NetWhere (Dashboard) 
  • Hadoop ecosystem, Hadoop / HDFS / Zookeeper / Spark / Hive
  • ETL/ELT workflow, Spark / Hive / Scala / Java

Reports to : CTO


Senior Software Engineer / Toro Development Ltd. / Taiwan, Taipei. 

Oct. 2012 - Dec. 2013.

Description and Responsibilities : TORO is a technology business that provides a mobile platform and its associated systems, services and rules to help Brands (with initial focus on Sports Teams, Smart Cities and Streaming apps) become super-apps to generate additional revenue with minimum effort. Responsibilities as below. 

  • Design, implement and test back-office modules for NFC wallet platform, Trusted Service Managers (TSM) and distributed NFC services to end­ users / stakeholders. 
  • Implement RESTful services and deliver endpoints for wallet managers and collaborating with front­end, backend teams for further business requirements. 

Tech Stacks: MYSQL / Spring / Hibernate / XML / Apache Camel / Java / POJO .etc. 

Reports to : Head of Server Solutions


Software Engineer / Digital River / Taiwan, Taipei. 

Oct. 2011 - Sep. 2012.

Description and Responsibilities : Digital river proactive partners, providing API-based Payments & Risk, Order Management and Commerce services to leading enterprise brands. The big challenge to DR is integrating with the current module and working well with a huge code base (over 2+ millions lines), the strict process including analysis requirements, design, implement, test and code review. More responsibilities as below. 

  • Design, implement custom bundle project, bundle customized by shoppers to pick products of groups and get special discounts, the main stakeholders /users from Logitech, Microsoft. 
  • Analysis, collect business requirements, identify use cases and collaborate with business analysts and deliver related diagrams, documents. 

Tech Stacks: Oracle / Tomcat / Spring / Struts / JDO / XML / JUnit / Java / J2EE .etc. 

Reports to : Technical Development Manager


Technical Supervisor / Stark Technology Inc. / Taiwan, Taipei. 

Oct. 2008 - Sep. 2011.

Description and Responsibilities : Stark Technology (STI) is the largest domestic system integrator in Taiwan. We plan and deliver complete ICT solutions for a wide spectrum of industries through representing and reselling the world's leading products. This is made possible by using the most advanced technology, and providing the best professional services. More responsibilities / projects as below. 

  • Lead, coach JR. programmers for the development process of enterprise modules, and design Fatwire CMS components as Template/Page/Cache .etc. 
  • Design, analyze DMDB systems, and implement functions to meet the requirements of queries / storage. Optimize performance for online servers and GC tuning. 

Tech Stacks : Oracle / Sybase / Tomcat / Weblogic / Spring / Struts / Hibernate / Fatwire / Java / J2EE .etc. 

Reports to : Technical Manager


Relevant Skills and Qualifications


Big Data Tech Stacks

  • AWS Services, EC2/S3/Lambda/EMR/CloudWatch/SNS/SQS/Elastic Beanstalk 
  • AWS Big Data Solutions, Kinesis/Firehose/Athena/Redshift/Dynamodb 
  • GCP Big Data Solutions, BigQuery/PubSub/Dataflow/Cloud Functions 
  • Hadoop ecosystem, Hadoop/HDFS/Zookeeper/Hbase/Hive 
  • Spark Streaming/Apache Kafka 
  • CI/CD: Jenkins/Cloud Formation/GitLab/Grafana

Specific Skills

  • Solid, well-designed real-time streaming/batch processing, ETL systems.
  • Monitors and conducts data-pipeline / machine learning pipeline development requests through lifecycle management and ensures that the technical solution meets.
  • Diagnosing and troubleshooting Redshift and specific clusters management.
  • Development of micro-services and endpoints based on enterprise integration patterns. Knowledge over garbage collection (JVM) tuning technologies for various servers.
  • Developed multi-threading processing consuming work and managed transactions.

Certifications and Training

  • Sun Certified Web Component Developer Java 2 Platform, Enterprise Edition. 
  • Sun Certified Programmer for the Java 2 Platform. 
  • Red Hat Enterprise Directory Services and Authentication Attended. 
  • Project Management Professional (PMP)® Attended. 
  • AWS Certified Solutions Architect Attended. 
  • Big Data on AWS Attended. 
  • Azure Data Engineer AssociateAttended. 

Education


National Taiwan University, 2010 – 2011

EMBA Programs, Business Administration, Accounting, Finance and International Business.


Chinese Culture University Master of Information Management, 2002 – 2005

Computer Science, Data Mining, Expert Systems and Knowledge Base as major concentration.


Chinese Culture University, Bachelor Degree of Science in Journalism, 1998 - 2002