Data Warehouse Engineer

  • Applications may have closed

Binance is the leading global blockchain ecosystem and cryptocurrency infrastructure provider whose suite of financial products includes the world’s largest digital-asset exchange
Our mission is to accelerate cryptocurrency adoption and increase the freedom of money
If you’re looking for a fast-paced, mission-driven organization where opportunities to learn and excel are endless, then Binance is the place for you

Requirements

    • According to the company's data warehouse specifications and business understanding, build a universal and flexible data warehouse system that can quickly support the needs and reduce repetitive development work efforts
    • Data model design, development, testing, deployment, online data job monitoring, and the ability to quickly solve complex problems, especially the optimization of complex calculation logic and performance tuning, etc
    • Participate in Data governance, including the construction of the company’s metadata management system and data quality monitoring system
       
    • Design and implement a data platform integrated with data lake warehouse to support real-time data processing and analysis requirements
    • Build knowledge graph, and provide in-depth business insight
    • Participate in technical team building and learning growth, and contribute to the team’s overall knowledge accumulation and skill improvement

Responsibilities

    • 5+ years experiences of data lake and data warehouse design and development experience
    • Deeply understanding of data warehouse modeling and data governance
      Solid knowledge of data warehouse development methodology, including dimensional modeling, information factory etc
    • Proficient in Java / Scala / Python (at least one language) and Hive & Spark SQL programming languages
    • Familiar with OLAP technology (such as: kylin, impala, presto, druid, etc
      )
    • Proficient in Big Data batch pipeline development
    • Familiar with Big Data components including but not limited to Hadoop, Hive, Spark, Delta lake, Hudi, Presto, Hbase, Kafka, Zookeeper, Airflow, Elastic search, Redis, etc
    • Experiences with AWS Big Data services are a plus
    • Have a strong team collaboration attitude and develop partnerships with other teams and businesses
    • Rich experience in real-time data processing, familiar with stream processing frameworks such as Apache Kafka, Apache Flink, in-depth knowledge of Lakehouse technology, practical project experience, proficiency in StarRocks, including its data model design, query optimization and performance tuning
    • Experience in knowledge graph construction and application, and knowledge of graph databases such as Nebula, etc

Listed in: , , , , , , , ,