Job Description
Join our team in building a modern, high-impact Analytical Platform for one of the largest integrated resort and entertainment companies in Southeast Asia. This platform will serve as a unified environment for data collection, transformation, analytics, and AI-driven insights—powering decisions across marketing, operations, gaming, and more.
You’ll work closely with Data Architects, ML Engineers, Business Analysts, and DevOps to design and implement scalable data solutions.
Requirements
- 5+ years of experience in data engineering or backend data development.
- Strong knowledge of data pipeline design, integration frameworks, and ETL tools.
- Experience working with cloud or hybrid data architectures.
- Proficiency in SQL and at least one programming language (e.g., Python, Scala).
- Hands-on experience with distributed data processing (e.g., Spark, Flink) is a plus.
- Familiarity with data lake, data warehouse, or lakehouse architectures.
- Experience with real-time data streaming and ingestion frameworks is a strong advantage.
- Understanding of data security, privacy, and compliance best practices.
- Experience working in Agile/Scrum environments.
Nice to have skills
- Experience with modern open-source tools (e.g., Airflow, dbt, Delta Lake, Apache Kafka).
- Exposure to machine learning pipelines or working alongside ML teams.
- Familiarity with BI tools and data visualization concepts.
- Experience working in regulated industries (e.g., gaming, finance, hospitality).
Responsibilities:
- Design, build, and maintain scalable and reliable data pipelines for ingesting data from various sources (internal systems, APIs, external platforms).
- Work with structured, semi-structured, and unstructured data, ensuring data quality, consistency, and integrity.
- Develop and maintain ETL/ELT processes to support real-time and batch analytics.
- Collaborate with Data Architects to design optimal data models and storage structures for analytics workloads.
- Implement data validation, deduplication, and transformation logic.
- Contribute to the definition of data governance, security, and access policies.
- Participate in platform scaling and performance optimization initiatives.
- Work closely with business and analytics teams to understand data needs and translate them into technical solutions.
Benefits
- 35 absence days per year for work-life balance
- Udemy courses of your choice
- English courses with native-speaker
- Regular soft-skills trainings
- Excellence Сenters meetups
- Online/offline team-buildings
- Business trips