Data Warehouse Engineer

Singapore / Taiwan, Taipei
Engineering – Big Data /
On-site
Binance is the leading global blockchain ecosystem and cryptocurrency infrastructure provider whose suite of financial products includes the world’s largest digital-asset exchange.
Our mission is to accelerate cryptocurrency adoption and increase the freedom of money.

If you’re looking for a fast-paced, mission-driven organization where opportunities to learn and excel are endless, then Binance is the place for you.

Requirements

    • According to the company's data warehouse specifications and business understanding, build a universal and flexible data warehouse system that can quickly support the needs and reduce repetitive development work efforts.
    • Data model design, development, testing, deployment, online data job monitoring, and the ability to quickly solve complex problems, especially the optimization of complex calculation logic and performance tuning, etc.
    • Participate in Data governance, including the construction of the company’s metadata management system and data quality monitoring system. 
    • Design and implement a data platform integrated with data lake warehouse to support real-time data processing and analysis requirements.
    • Build knowledge graph, and provide in-depth business insight.
    • Participate in technical team building and learning growth, and contribute to the team’s overall knowledge accumulation and skill improvement.

Responsibilities

    • 5+ years experiences of data lake and data warehouse design and development experience.
    • Deeply understanding of data warehouse modeling and data governance. Solid knowledge of data warehouse development methodology, including dimensional modeling, information factory etc.
    • Proficient in Java / Scala / Python (at least one language) and Hive & Spark SQL programming languages.
    • Familiar with OLAP technology (such as: kylin, impala, presto, druid, etc.).
    • Proficient in Big Data batch pipeline development.
    • Familiar with Big Data components including but not limited to Hadoop, Hive, Spark, Delta lake, Hudi, Presto, Hbase, Kafka, Zookeeper, Airflow, Elastic search, Redis, etc.
    • Experiences with AWS Big Data services are a plus.
    • Have a strong team collaboration attitude and develop partnerships with other teams and businesses.
    • Rich experience in real-time data processing, familiar with stream processing frameworks such as Apache Kafka, Apache Flink, in-depth knowledge of Lakehouse technology, practical project experience, proficiency in StarRocks, including its data model design, query optimization and performance tuning.
    • Experience in knowledge graph construction and application, and knowledge of graph databases such as Nebula, etc.