Email: [email protected]
Tel: 0934039069
Python
C++
Tensorflow
Pandas
Numpy
Matplotlib
Data Processing
Web Crawler
Machine Learning
Practices of Deep Learning
Cloud Computing and Services
Practice of Social Media Analytics
Neural Networks
Information Retrieval
Algorithm Design and Application
- 熟悉RESTful API,公司內部服務互相串接 (Gerrit、Jenkins、Jira)
- CI/CD intergration,自動化code-review流程以及自動化驗證程式碼的安全性
- 熟悉WAF build,完善軟體構建系統
- 熟悉Git版本控制
- Hyperedge Aware Greedy (HAG) 演算法實作
- HAG演算法效能改善
Hang-Yang Wu, Yi-Ling Chen, "Graph Sparsification with Generative Adversarial Network", IEEE International Conference on Data Mining (ICDM 2020)
After completing the courses. I applied for the internship program of the Academia Sinica. During the period of Academia Sinica, my job is to improve the performance of the algorithm.
Our work is to make a model that uses deep learning to predict weather profiles. The main information is tides, waves, and wind direction. Mainly for people who love surfing.
Graph sparsification aims to reduce the number of edges of a network while maintaining its accuracy for given tasks. In this study, we propose a novel method that is able to sparsify networks for community detection tasks. Our method is able to capture those relationships that are not shown in the original graph but are relatively important, and creating artificial edges to reflect these relationships and further increase the effectiveness of the community detection task.
This is the seed diffusion algorithm I implemented during my internship at the SINICA. It can be applied to graphs with hyperedge. In addition, I also make three pruning techniques to speed up the execution time, and use multithread to parallelize the algorithm.
The order of the text affects the overall semantics, Using Deep Learning, let the first and last words as input, and the middle as output. If each word is made as a one-hot vector, data-sparse problems may occur. In order to avoid this, the word embedding technique is used to reach the compression of the text vector representation.
It's a 2015 ECML PKDD competition, I use Deep Learning approaches to predict the taxi driving time. Although it's expired when I joined this competition on Kaggle. I still submit my result and the score is 0.60 (mean-squared-error). The first place score was 0.53.
From the perspective of probability, use the Expectation-Maximization algorithm to calculate the probability of generating the current Topic for each document to match the degree of the query. After that, I have more knowledge of information retrieval.
The classic data retrieval approaches converts the documents and query into vectors, then calculates the cosine similarity between matching query and the document. This helps me know how the search engine work although it's just a basic approach and not advanced.
MNIST handwriting recognition is a very basic machine learning example. As a beginner for machine learning, I also try to start with this.
In order to know well with python and machine learning, I got some practices on the internet. This is a water level prediction with the Scikit tool.