• Lead the team technically in improving scalability, stability and efficiency of our Data platform
• Design, build and deploy new libraries and frameworks for our core systems while keeping highest standards of code quality
• Cooperate with other engineers and Product owners to identify and build incremental products through Scrum sprint
• Understand the following technologies and navigate them properly: Python, Scala, Spark, SQL etc.
• Collaborate with the Data Architect team on regular basic to design/review data models, and application architectures
• Maintain software clean architecture, clean code, and high quality.
• Participate in code reviews, pair programming, mob programming, and coaching other members
1. Must have:
• Bachelor’s or Master’s degree in Computer Science, Software Engineering, Information Technology, or a related technical field
• English is required
• Have 3+ years of experience as a Data Engineer or Software Engineer
• Have experience in Cloud (AWS/Azure/GCP)
• Extremely proficient in at least 1 programming language (Python/Scala/Java)
• Strong experience in systems architecture – particularly in complex, scalable, and fault-tolerant distributed systems
• Good at multi-threading, atomic operations, computation framework: Spark (DataFrame, SQL, ...), distributed storage, distributed computing
• Understand designs of resilience, fault-tolerance, high availability, and high scalability, ...
• Tools: CI/CD, Gitlab, ...
• Good at communication & team working
• Being open-minded, willing to learn new things
2. Nice to have:
• Databricks is a plus
• Experience in performance tuning/optimizing Big Data programs