CakeResume Talent Search

Advanced filters
On
4-6 years
6-10 years
10-15 years
More than 15 years
Avatar of the user.
Avatar of the user.
Past
Senior Data Analyst @趨勢科技
2022 ~ Present
Data Scientist, Data Analyst, Machine Learning Engineer
Within one month
python
R
SQL
Unemployed
Ready to interview
Full-time / Interested in working remotely
4-6 years
輔仁大學 Fu Jen Catholic University
統計資訊學系
Avatar of Yen-Ting Liu.
Avatar of Yen-Ting Liu.
Data Engineer @Tesla
2023 ~ 2023
Data engineer / Data anyayst
Within two months
Yen-Ting Liu 我具有5年python資料分析,熟悉以Docker搭配nginx, redis部屬api及系統於GCP上。熟悉Airflow程式及報表自動化分析流程,並有Hadoop,Elasticsearch群集管理實務、pyspark數據ETL經驗。我喜歡學習新技術,並追求以更高效率進行資料處理流程。 Santa Clara, CA, USA [email protected] 工作經歷 Data Engineer
python
Linux
R
Employed
Ready to interview
Full-time / Interested in working remotely
4-6 years
University of Texas at Dallas
Information Technology and Management
Avatar of 陳昭儒.
Avatar of 陳昭儒.
Past
Data Engineer @BUBBLEYE | We're hiring!
2021 ~ 2022
Software Enginer
Within two months
陳昭儒(Chao-Ju Chen) Github [email protected] Education National Taiwan University Bachelor’s Degree, Electrical Engineering 2012 ~ 2017 Project Highlights Aggregating Files in one ETL, output 60B row to Data Warehouse Input :gzipped files(200GB in total) Task : Loading columns with values parsed from each gzipped file name. Wrote to BigQuery existing table(specific schema) in parallel. Tool: GCP Dataflow(Hosted Serverless Apache Beam) Result : The job took 40min to finish. Machine Type: n1-standard-1(1 vcpu, 3.75GB memory) Autoscaled up to 122 workers at peak. The data
Python
ETL
Web Scraping
Unemployed
Ready to interview
Full-time / Interested in working remotely
4-6 years
National Taiwan University
電機工程學系
Avatar of Clark Wang.
Avatar of Clark Wang.
Senior Data Scientist @PTI 力成科技
2016 ~ Present
大數據分析,資料科學家,資料工程師,AI工程師
Within one month
known their requirement and current difficulty, and guide end-user to establish their own analysis flow, thus reducing and replacing many daily manual analysis processes. In the meantime, i have experience on In-house user training too. iv. ETL for Tableau. I write python script on pyspark to summary daily output, machine error code, quality checking data, and pass it to Tableau for visualization. v. Unscheduled AI and statistical education and training for production line person and engineers. 學歷 SepJun 2012 逢甲大學 Applied Mathematics - Master degree 技能 Data
Data Augmentation for Rare Defect Images
Signal Processing & Recognition
Administrator for Engineering Data Analysis System
Employed
Open to opportunities
Full-time / Interested in working remotely
6-10 years
逢甲大學
Applied Mathematics
Avatar of 謝欣宏.
Avatar of 謝欣宏.
高級工程師 @電商
2023 ~ Present
軟體工程師、影像處理工程師、AI處理工程師、演算法工程師
Within one month
市集 六月至今 成功實作整合了 Kubernetes、Airflow、Gitlab、Gitlab Runner 和 Docker Registry 的微服務系統架構 。 重構 TensorFlow Model Server AI 專案,實作 gRPC 協定減少通訊延遲 。 使用 PySpark 和 Apache Beam 處理深度學習的億級資料前處理 。 利用 Spring Boot 改進了 NLP 專案,增加模組化設計、新增單元測試,加入微服務同步機制
C++
Java
JavaScript
Employed
Open to opportunities
Full-time / Interested in working remotely
More than 15 years
國立台灣科技大學
資訊與通訊
Avatar of 許立農.
Offline
Avatar of 許立農.
Offline
數據科學家 @中國信託商業銀行股份有限公司
2021 ~ Present
AI工程師、機器學習工程師、深度學習工程師、資料科學家、Machine Learning Engineer、Deep Learning Engineer、Data Scientist
Within three months
許立農 | Hsu, Li-Nung Data Scientist、Data Engineer Taipei [email protected] Education National Chenchi University, MS, Statistics, 2015 – 2017 GPA : 3.84 / 4.0 Master Thesis: Entropy Based Feature Selection, Professor Pei-Ting, Chou Objective: Build a similarity matrix based on Mutual Entropy under Hierarchical Clustering. Afterwards, select clustered features as the final selection. Compare the model with other feature selection methods like RF, Lasso, F-score. National Chen-Kung University, BS, Mathematics, 2011 – 2015 Skills Programing Python Scala R MSSQL Data-related Tools Tensorflow (Keras) PyTorch Spark Docker
Python
R
MSSQL
Employed
Open to opportunities
Full-time / Interested in working remotely
4-6 years
政治大學
統計
Avatar of the user.
Avatar of the user.
Data Engineer @美好金融
2022 ~ 2023
軟體工程師
Within two months
Java
Python
MongoDB
Employed
Open to opportunities
Full-time / Interested in working remotely
6-10 years
國立中央大學
物理
Avatar of 孫煜凱.
Avatar of 孫煜凱.
Past
機器學習工程師 @順豐科技公司
2021 ~ 2022
AI工程師、機器學習工程師、深度學習工程師、資料科學家、Machine Learning Engineer、Deep Learning Engineer、Data Scientist
Within one month
5. 技術能力 熟練Hadoop生態和集群運行原理 ,與 Hive、 Presto組件 ,熟練掌握SQL語言編寫與模型落地,具有其調優經驗 ; 熟練使用Python與Pyspark, 熟練 使用RDD算子與Spark SQL、Sklearn等,進行機器學習與數據分析 ; 熟練使用Git與Docker,能進行YAML、Dockerfile編寫與CICD部署流程、搭建RESTful API等 ; 熟
Word
PowerPoint
Excel
Unemployed
Full-time / Interested in working remotely
4-6 years
國立政治大學(National Chengchi University)
統計系
Avatar of the user.
Software Engineer / Backend Engineer
Within three months
FastAPI(Python)
System Design
GCP Compute Engine
Unemployed
Full-time / Interested in working remotely
6-10 years
中原大學 Chung Yuan Christian University
資訊工程學系
Avatar of 翁崇恒.
大數據資料工程師、演算法工程師
Within two months
翁崇恒 我的名字是翁崇恒,自2018年從事資料科學與AI領域約5年的時間,目前正在尋求資料科學與人工智慧的職務。我曾在工作上研發新的深度學習演算法,能夠在非常少的資源下運行模型來做自動控制;在學術研究上,我能夠發明新的方法
Python
machine learning
Data Analysis
Full-time / Interested in working remotely
4-6 years
國立中央大學 National Central University
人工智慧

The Most Lightweight and Effective Recruiting Plan

Search resumes and take the initiative to contact job applicants for higher recruiting efficiency. The Choice of Hundreds of Companies.

  • Browse all search results
  • Unlimited access to start new conversations
  • Resumes accessible for only paid companies
  • View users’ email address & phone numbers
Search Tips
1
Search a precise keyword combination
senior backend php
If the number of the search result is not enough, you can remove the less important keywords
2
Use quotes to search for an exact phrase
"business development"
3
Use the minus sign to eliminate results containing certain words
UI designer -UX
Only public resumes are available with the free plan.
Upgrade to an advanced plan to view all search results including tens of thousands of resumes exclusive on CakeResume.

Definition of Reputation Credits

Technical Skills
Specialized knowledge and expertise within the profession (e.g. familiar with SEO and use of related tools).
Problem-Solving
Ability to identify, analyze, and prepare solutions to problems.
Adaptability
Ability to navigate unexpected situations; and keep up with shifting priorities, projects, clients, and technology.
Communication
Ability to convey information effectively and is willing to give and receive feedback.
Time Management
Ability to prioritize tasks based on importance; and have them completed within the assigned timeline.
Teamwork
Ability to work cooperatively, communicate effectively, and anticipate each other's demands, resulting in coordinated collective action.
Leadership
Ability to coach, guide, and inspire a team to achieve a shared goal or outcome effectively.
Within six months
Data Scientist, Data Engineer
Logo of 中國信託商業銀行股份有限公司.
中國信託商業銀行股份有限公司
2021 ~ Present
台灣台北市
Professional Background
Current status
Employed
Job Search Progress
Open to opportunities
Professions
Data Scientist, Machine Learning Engineer
Fields of Employment
Banking, Artificial Intelligence / Machine Learning, AdTech / MarTech
Work experience
4-6 years
Management
None
Skills
Python
R
MSSQL
Scala
Linux
PyTorch
Tensorflow (Keras)
AWS
GCP
Spark
Tensorflow
pyspark
Languages
English
Fluent
Job search preferences
Positions
AI工程師、機器學習工程師、深度學習工程師、資料科學家、Machine Learning Engineer、Deep Learning Engineer、Data Scientist
Job types
Full-time
Locations
台灣台北, 台灣新北市
Remote
Interested in working remotely
Freelance
Yes, I freelance in my spare time
Educations
School
政治大學
Major
統計
Print
E3uoaqcxyy6dppaet0kg

許立農 | Hsu, Li-Nung


Data Scientist、Data Engineer
Taipei
[email protected]

Education

National Chenchi University, MS, Statistics, 2015 – 2017

  • GPA : 3.84 / 4.0
  • Master Thesis: Entropy Based Feature Selection, Professor Pei-Ting, Chou
    • Objective: Build a similarity matrix based on Mutual Entropy under Hierarchical Clustering. Afterwards, select clustered features as the final selection.
    • Compare the model with other feature selection methods like RF, Lasso, F-score.

Igtt7bfqhad2uml5y0ki

National Chen-Kung University, BS, Mathematics, 2011 – 2015


Kxc0f0caus5l9rwo4qji

Skills


Programing

  • Python
  • Scala
  • R
  • MSSQL


Data-related Tools

  • Tensorflow (Keras)
  • PyTorch
  • Spark
  • Docker
  • Scikit-Learn
  • Pandas


Cloud Platform

  • AWS
  • GCP


Language

  • English: TOEFL 98 / 120

Work Experience

CTBC Bank, Model Development Department, Data Scientist

2021.12 – present

  • About the department:
    • Responsible for developing models related to bank recommendations and risks, including projects such as coupon recommendations, account opening marketing lists, and fraud detection.
  • Job responsibilities:
    • Throughout the entire project lifecycle, my primary responsibilities included model design, model training, end-to-end process development, feature design, performance tracking, and method research.
Lqnpwfiwbu3f99i6zod4

Fraud Alert Project

  • Objective:
    • Predicting potential fraudulent accounts based on transaction data, restricting transactions in advance to prevent harm.
  • Responsibilities/Achievements:
    • Development and deployment of credit card and financial features.
    • Managing the data flow process from receiving variables to model predictions, identifying risk factors, and updating alert lists.
    • Implemented Autoencoder + contrastive learning to achieve a 1.81% improvement in model effectiveness.

Coupon Recommendation

  • Objective:
    • Personalized coupon recommendations for mobile banking users to increase click-through rates and redemption rates.
  • Responsibilities/Achievements:
    • Utilized multi-task learning to simultaneously predict click-through behavior and coupon redemptions, resulting in a 14% increase in click-through rate and a 74% increase in redemption rate.
    • Created performance tracking reports to monitor online model performance and provide insights to Business Units.

Financial Product Recommendations

  • Objective:
    • Tailored financial product recommendations for mobile banking users to enhance click-through rates without compromising conversion rates.
  • Responsibilities/Achievements:
    • Applied multi-task learning to jointly learn click-through and conversion behaviors, fine-tuned model architecture, achieving a 90% outperformance against competitor models in online testing.

Marketing List for Digital Savings Accounts

  • Objective:
    • Optimized conversion rates for marketing lists related to digital savings accounts
  • Responsibilities/Achievements:
    • successfully raising conversion rates from 0.23% to 1.16%

Work Experience

CLICKFORCE, Data Engineer Supervisor, 2020.1 – 2021.11

  • About the company:
    • As a top domestic digital advertisement company, CLICKFORCE cooperates with over 900 web media and over 400 mobile media to build a huge advertising environment. CLICKFORCE considers data-driven solution as the core concept of the company, and dedicates to help advertisers to achieve their commercial goals.
    • At 2020, CLICKFORCE won 2 awards at Agency & Advertiser of the Year.
    • Successfully acquire the exclusive advertising agency qualification for Tokyo 2020 Olympics in Taiwan.
  • Job responsibilities:
    • Optimize ad performance from all aspects, including the system, target audience tags, etc.
    • Do researches for new ML model (recommender model, NLP model) or architecture which is suitable for our system.
    • Develop data-related products or projects.
    • Analyze data to help improve our system or inspect whether the demands from business side is doable.
Lqnpwfiwbu3f99i6zod4

Real-time AD Recommender System

  • Objective:
    • Building a real-time ad recommender system to upgrade our ad server and get better performance.
  • Responsibilities:
    • Figure out what kind of recommender system components that is suitable for our ad system.
    • Build a tower-like and feature-cross model refer to other famous recommender system model.
    • Responsible for system engineering, which includes data preprocessing, embedding generates, memory cache, cold start, model API, etc.

Interest Tags

  • Objective:
    • Build interest tags for ads to help ad optimizers choose their target audience.
  • Responsibilities:
    • Create the features from what articles they saw, what website they viewed, and what ads they interacted.
    • Deal with 20 million rows data and 120 million inference samples.
    • Build ML model to predict each user's behavior on certain ads.
    • Using Spark through AWS EMR to accelerate the speed of producing tags.
  • Achievements:
    • Raise CTR performance up to 200-300% of the original tags depends on different tags, and gain more impression while maintain better performance.
    • After accomplishing this project, we terminated the cost on purchasing interest tags from other company, and successfully turned the original cost into revenue by providing profitable data.

First Party Cookie Mapping

  • Objective:
    • Deal with the Google 3rd party Cookie issue, figure out a method to map numerous 1st party Cookies to a user.
  • Responsibility:
    • Transform this problem into a ML mission. Design the label of the data, figure out what feature we can get or produce and whether the feature is useful for the goal.
    • Apply XGboost on this mission.
    • Build a small test to prove this method works.
  • Achievement:
    • 70% of precision.
    • One of the solution of our company while the cancelation of 3rd party Cookie happen.

Invoice Data Application

  • Objective:
    • Develop invoice data application.
  • Responsibility:
    • Responsible for fine-tuning BERT to predict category for each product.
    • Produce invoice data report to brands or business unit. It demonstrates the sales volume across different channel, what kind of products are frequently bought together, and also shows comparison of target brand to the other brands.
  • Achievements:
    • Produce an invoice data report product.
    • Produce invoice tags for ad system.

Other Experience

E.Sun AI 2020 Summer Competition, 2020.7 – 2020.8

  • Objective:
    • Extract names of money laundering suspects from an article.
  • Responsibilities:
    • Crawl the articles from different media, and parse them by using Selenium, Requests, and Beautiful Soup.
    • Construct 2-step model: First, identify whether the article is related to money laundering. Second, extract the suspects' names.
    • Build model serving API by Tensorflow Serving.
    • Build REST API for preprocessing request data and return the prediction.
  • Achievement:
    • 23rd place among 409 teams.

Youtube Data-Driven Marketing System, Institute for Information Industry, 2019.8 – 2019.11

  • Objectives:
    • Use the title and the description of videos to automatically classify videos.
    • Use the title and the description of videos to identify whether a video is sponsored.
    • Give suggestions for Youtubers or companies who desire to sponsor in a video based on data analysis.
  •  Responsibilities:
    • Apply Google API and write Python functions to get structured raw data.
    • Train word vectors using Gensim based on Wiki's open data. 
    • Use the frequency of each sentence as a criteria to eliminate useless words.
    • Tune LSTM, Conv1D, BERT on the NLP mission.
    • Use EDA methods to see the insights of the data under different classes and different sponsored status.
  • Achievement:
    • 71% accuracy in classifying video’s type.
    • 89% accuracy in detecting sponsored content.

E.Sun Real Estate Price Prediction Competition, 2019.7 – 2019.8

  • Objective:
    • Use the real estate training data to build a model and predict the real estate price within 10% residual.
  • Responsibilities:
    • Apply XGBoost, LGBM and other ML models to train the model.
    • Collect the outputs as new features from each ML model and add them into the original data set to enhance the performance of the final model.
  • Achievement:
    • 150th place out of 1200 teams.


KKTV Data Game,2017.5 – 2017.6

  • Objective:
    • Predict the next video a user watch in the next time interval.
  • Responsibilities:
    • Extract different features from raw data, such as the latest video, the video which got the longest viewing time, the video which got the largest number of viewing.
    • Use the user viewing data to construct a similarity matrix of each video as additional features.
  • Achievement:
    • 10th place out of 50 teams.


MRT Open Data Competition, 2017.4 – 2017.5

  • Objective:
    • Study the changes of passenger volume of MRT by surrounding geometric data.
  • Responsibilities:
    • Apply bisection method to build the edges between MRT stations.
    • Combine other geometric data based on these borders.
    • Use Lasso feature selection method to explore the importance of each feature.
    • Add noises into features to check the features are not randomly selected.
  • Achievement:
    • Certificate of Honorable Mention.


Resume
Profile
E3uoaqcxyy6dppaet0kg

許立農 | Hsu, Li-Nung


Data Scientist、Data Engineer
Taipei
[email protected]

Education

National Chenchi University, MS, Statistics, 2015 – 2017

  • GPA : 3.84 / 4.0
  • Master Thesis: Entropy Based Feature Selection, Professor Pei-Ting, Chou
    • Objective: Build a similarity matrix based on Mutual Entropy under Hierarchical Clustering. Afterwards, select clustered features as the final selection.
    • Compare the model with other feature selection methods like RF, Lasso, F-score.

Igtt7bfqhad2uml5y0ki

National Chen-Kung University, BS, Mathematics, 2011 – 2015


Kxc0f0caus5l9rwo4qji

Skills


Programing

  • Python
  • Scala
  • R
  • MSSQL


Data-related Tools

  • Tensorflow (Keras)
  • PyTorch
  • Spark
  • Docker
  • Scikit-Learn
  • Pandas


Cloud Platform

  • AWS
  • GCP


Language

  • English: TOEFL 98 / 120

Work Experience

CTBC Bank, Model Development Department, Data Scientist

2021.12 – present

  • About the department:
    • Responsible for developing models related to bank recommendations and risks, including projects such as coupon recommendations, account opening marketing lists, and fraud detection.
  • Job responsibilities:
    • Throughout the entire project lifecycle, my primary responsibilities included model design, model training, end-to-end process development, feature design, performance tracking, and method research.
Lqnpwfiwbu3f99i6zod4

Fraud Alert Project

  • Objective:
    • Predicting potential fraudulent accounts based on transaction data, restricting transactions in advance to prevent harm.
  • Responsibilities/Achievements:
    • Development and deployment of credit card and financial features.
    • Managing the data flow process from receiving variables to model predictions, identifying risk factors, and updating alert lists.
    • Implemented Autoencoder + contrastive learning to achieve a 1.81% improvement in model effectiveness.

Coupon Recommendation

  • Objective:
    • Personalized coupon recommendations for mobile banking users to increase click-through rates and redemption rates.
  • Responsibilities/Achievements:
    • Utilized multi-task learning to simultaneously predict click-through behavior and coupon redemptions, resulting in a 14% increase in click-through rate and a 74% increase in redemption rate.
    • Created performance tracking reports to monitor online model performance and provide insights to Business Units.

Financial Product Recommendations

  • Objective:
    • Tailored financial product recommendations for mobile banking users to enhance click-through rates without compromising conversion rates.
  • Responsibilities/Achievements:
    • Applied multi-task learning to jointly learn click-through and conversion behaviors, fine-tuned model architecture, achieving a 90% outperformance against competitor models in online testing.

Marketing List for Digital Savings Accounts

  • Objective:
    • Optimized conversion rates for marketing lists related to digital savings accounts
  • Responsibilities/Achievements:
    • successfully raising conversion rates from 0.23% to 1.16%

Work Experience

CLICKFORCE, Data Engineer Supervisor, 2020.1 – 2021.11

  • About the company:
    • As a top domestic digital advertisement company, CLICKFORCE cooperates with over 900 web media and over 400 mobile media to build a huge advertising environment. CLICKFORCE considers data-driven solution as the core concept of the company, and dedicates to help advertisers to achieve their commercial goals.
    • At 2020, CLICKFORCE won 2 awards at Agency & Advertiser of the Year.
    • Successfully acquire the exclusive advertising agency qualification for Tokyo 2020 Olympics in Taiwan.
  • Job responsibilities:
    • Optimize ad performance from all aspects, including the system, target audience tags, etc.
    • Do researches for new ML model (recommender model, NLP model) or architecture which is suitable for our system.
    • Develop data-related products or projects.
    • Analyze data to help improve our system or inspect whether the demands from business side is doable.
Lqnpwfiwbu3f99i6zod4

Real-time AD Recommender System

  • Objective:
    • Building a real-time ad recommender system to upgrade our ad server and get better performance.
  • Responsibilities:
    • Figure out what kind of recommender system components that is suitable for our ad system.
    • Build a tower-like and feature-cross model refer to other famous recommender system model.
    • Responsible for system engineering, which includes data preprocessing, embedding generates, memory cache, cold start, model API, etc.

Interest Tags

  • Objective:
    • Build interest tags for ads to help ad optimizers choose their target audience.
  • Responsibilities:
    • Create the features from what articles they saw, what website they viewed, and what ads they interacted.
    • Deal with 20 million rows data and 120 million inference samples.
    • Build ML model to predict each user's behavior on certain ads.
    • Using Spark through AWS EMR to accelerate the speed of producing tags.
  • Achievements:
    • Raise CTR performance up to 200-300% of the original tags depends on different tags, and gain more impression while maintain better performance.
    • After accomplishing this project, we terminated the cost on purchasing interest tags from other company, and successfully turned the original cost into revenue by providing profitable data.

First Party Cookie Mapping

  • Objective:
    • Deal with the Google 3rd party Cookie issue, figure out a method to map numerous 1st party Cookies to a user.
  • Responsibility:
    • Transform this problem into a ML mission. Design the label of the data, figure out what feature we can get or produce and whether the feature is useful for the goal.
    • Apply XGboost on this mission.
    • Build a small test to prove this method works.
  • Achievement:
    • 70% of precision.
    • One of the solution of our company while the cancelation of 3rd party Cookie happen.

Invoice Data Application

  • Objective:
    • Develop invoice data application.
  • Responsibility:
    • Responsible for fine-tuning BERT to predict category for each product.
    • Produce invoice data report to brands or business unit. It demonstrates the sales volume across different channel, what kind of products are frequently bought together, and also shows comparison of target brand to the other brands.
  • Achievements:
    • Produce an invoice data report product.
    • Produce invoice tags for ad system.

Other Experience

E.Sun AI 2020 Summer Competition, 2020.7 – 2020.8

  • Objective:
    • Extract names of money laundering suspects from an article.
  • Responsibilities:
    • Crawl the articles from different media, and parse them by using Selenium, Requests, and Beautiful Soup.
    • Construct 2-step model: First, identify whether the article is related to money laundering. Second, extract the suspects' names.
    • Build model serving API by Tensorflow Serving.
    • Build REST API for preprocessing request data and return the prediction.
  • Achievement:
    • 23rd place among 409 teams.

Youtube Data-Driven Marketing System, Institute for Information Industry, 2019.8 – 2019.11

  • Objectives:
    • Use the title and the description of videos to automatically classify videos.
    • Use the title and the description of videos to identify whether a video is sponsored.
    • Give suggestions for Youtubers or companies who desire to sponsor in a video based on data analysis.
  •  Responsibilities:
    • Apply Google API and write Python functions to get structured raw data.
    • Train word vectors using Gensim based on Wiki's open data. 
    • Use the frequency of each sentence as a criteria to eliminate useless words.
    • Tune LSTM, Conv1D, BERT on the NLP mission.
    • Use EDA methods to see the insights of the data under different classes and different sponsored status.
  • Achievement:
    • 71% accuracy in classifying video’s type.
    • 89% accuracy in detecting sponsored content.

E.Sun Real Estate Price Prediction Competition, 2019.7 – 2019.8

  • Objective:
    • Use the real estate training data to build a model and predict the real estate price within 10% residual.
  • Responsibilities:
    • Apply XGBoost, LGBM and other ML models to train the model.
    • Collect the outputs as new features from each ML model and add them into the original data set to enhance the performance of the final model.
  • Achievement:
    • 150th place out of 1200 teams.


KKTV Data Game,2017.5 – 2017.6

  • Objective:
    • Predict the next video a user watch in the next time interval.
  • Responsibilities:
    • Extract different features from raw data, such as the latest video, the video which got the longest viewing time, the video which got the largest number of viewing.
    • Use the user viewing data to construct a similarity matrix of each video as additional features.
  • Achievement:
    • 10th place out of 50 teams.


MRT Open Data Competition, 2017.4 – 2017.5

  • Objective:
    • Study the changes of passenger volume of MRT by surrounding geometric data.
  • Responsibilities:
    • Apply bisection method to build the edges between MRT stations.
    • Combine other geometric data based on these borders.
    • Use Lasso feature selection method to explore the importance of each feature.
    • Add noises into features to check the features are not randomly selected.
  • Achievement:
    • Certificate of Honorable Mention.