I'm leading, architect a large-scale data pipeline from Paktor x 17LIVE (GCP/AWS/python/scala), cooperating with Singapore data science/machine learning team/TW HQ data team, and Hadoop ecosystem related experiences such as HDFS/Hbase/Kafka in JSpectrum (Hong Kong, Sydney).
● Having 12+ years of experience in designing, developing various Java/Scala/Python based applications that support day to day operations.
● 6+ years of working on multiple issues as a data team member or driving analysis and elaboration, design, development of data-pipeline applications and builds automation tools.
● Extensive knowledge of Spark and Hadoop ecosystem, such as Hadoop, HDFS, HBase, etc.
● Strong knowledge in developing/designing AWS/GCP/Azure Big Data services.
● Extensive skill towards developing generic distributed systems, streaming processing, deployed machine learning pipeline and continuous development of ML models.
Senior Data Engineer at Paktor x 17LIVE| AWS x GCP x Azure Big Data Specialist
Singapore / Hong Kong / Taiwan
https://www.linkedin.com/in/chin-hung-wilson-liu-29392957
Nanxing Rd., Xizhi Dist., New Taipei City, Taiwan (R.O.C.)
Description and Responsibilities : The big challenge of 17 Media data teams is facing fast-growing data volume (processing 5-10x TB level daily), complex cooperation with stakeholders, the cost optimization of pipeline and refactor big latency systems .etc. As a senior data member, I’m making a data dictionary and trying to explain/design how the whole pipeline working with each components, especially how to solve those bottlenecks. More responsibilities / details as below.
Tech Stacks :
Reports to : Data Head
Description and Responsibilities : This is another 0 to 1 story. As an early data member, we need to figure out the data driven policy, strategies, engineering requirements from the company. In Paktor, data / backend sides are 100% on AWS, therefore the whole data ingestion, automation and data warehouse etc. are relying on those components. We are processing 50-100x GB realtime / batch jobs and the other data sources (RDBMS, APIs) for ETL/ELT on S3, Redshift, the data platform helps our marketing / HQ scientists team getting data into insights and making good decisions. More responsibilities / details as below.
Tech Stacks :
Reports to : CTO, Data Head
Description and Responsibilities : JSPectrum is a leading passive location-based service company in Hong Kong which holds many interesting products such as NetProbe, NetWhere, NetAd etc. In Optus (The main project in Sydney), the main responsibility of system analyst is designing / implementing data ingestion (real-time processing) / load and management data with major components of the Hadoop ecosystem. We meet the challenge to process 15,000 TPS, 60,000 inserts per second and 300 GB daily storages, therefore we are trying to optimize those components with Kafka consumers, HDFS storages and re-designing keys / columns of HBase to fulfill the requirement and deployed NetAd, whole in-house solutions on Optus. More responsibilities / details as below.
Tech Stacks :
Reports to : CTO
Description and Responsibilities : TORO is a technology business that provides a mobile platform and its associated systems, services and rules to help Brands (with initial focus on Sports Teams, Smart Cities and Streaming apps) become super-apps to generate additional revenue with minimum effort. Responsibilities as below.
Tech Stacks: MYSQL / Spring / Hibernate / XML / Apache Camel / Java / POJO .etc.
Reports to : Head of Server Solutions
Description and Responsibilities : Digital river proactive partners, providing API-based Payments & Risk, Order Management and Commerce services to leading enterprise brands. The big challenge to DR is integrating with the current module and working well with a huge code base (over 2+ millions lines), the strict process including analysis requirements, design, implement, test and code review. More responsibilities as below.
Tech Stacks: Oracle / Tomcat / Spring / Struts / JDO / XML / JUnit / Java / J2EE .etc.
Reports to : Technical Development Manager
Description and Responsibilities : Stark Technology (STI) is the largest domestic system integrator in Taiwan. We plan and deliver complete ICT solutions for a wide spectrum of industries through representing and reselling the world's leading products. This is made possible by using the most advanced technology, and providing the best professional services. More responsibilities / projects as below.
Tech Stacks : Oracle / Sybase / Tomcat / Weblogic / Spring / Struts / Hibernate / Fatwire / Java / J2EE .etc.
Reports to : Technical Manager
National Taiwan University, 2010 – 2011
EMBA Programs, Business Administration, Accounting, Finance and International Business.
Chinese Culture University Master of Information Management, 2002 – 2005
Computer Science, Data Mining, Expert Systems and Knowledge Base as major concentration.
Chinese Culture University, Bachelor Degree of Science in Journalism, 1998 - 2002