1. Large-scale data ingestion and integration – design, implement scalable ETL processes to collect and store large amounts of data from multiple data centers and diverse external partners.
2. Real-time query engine – design, implement our state-of-the-art in-memory query engine to bring quick insights to customers.
3. Data Reporting – design, implement our internal BI reporting platform and provide insights to clients.
1. BSc/MSc in Computer Science or other related field.
2. 2+ years working experience in data related systems, e.g. data warehouse, ETL, etc.
3. Strong in Java/Scala/Python.
4. Solid understanding and working knowledge of relational and non-relational databases is a must.
5. Experienced in Spark, Hadoop Yarn or similar distributed computation work is a plus.
6. Experienced in search engine development and familiar with ElasticSearch is a plus.