• Design, build, launch and maintain extremely efficient & reliable large-scale batch and real-time data pipelines with data processing frameworks like Dataflow, Flink, Spark and Kafka, which would enable teams to consume and understand data faster
  • Run extensive and intensive POCs on data processing frameworks, technologies, and platforms
  • Maintain and improve data quality through testing, tooling and continuously evaluation of performance
  • Strong understanding of data streaming concepts
  • Experience with building production-grade data processing tools such as BigQuery, Kafka, Pub/Sub, Dataflow, Flink
  • Experience in writing codes in Java, Python, SQL
  • Passionate about crafting clean code and story-like documentation
  • Knowledgeable about system design, data structures, and algorithms
  • Strong familiarity about data modeling, data access, and data storage techniques, like Data Lake and Data Warehousing concepts
  • Understand the value of collaboration within teams and be very customer oriented
  • Interested in being the glue between engineering and analysis

Other jobs at Gojek