Kpler

Data Engineer

Job Description

The role involves end-to-end ownership of development tasks, beginning with a thorough understanding of assigned tickets and requirements. The individual designs and builds functionality—including APIs and data processing components—ensuring code is deployed to development environments and reviewed through peer and product testing. 

They are responsible for writing and executing unit, integration, and functional tests aligned with defined test scenarios, while ensuring full compliance through detailed validation. After release, the role includes monitoring system performance, alerts, and SLOs to ensure optimal functionality and reliability.


Responsibilities
  • Design, operate, and document versioned RESTful APIs using FastAPI and JVM-based frameworks, ensuring scalability, reliability, and backward compatibility.
  • Implement and enforce data schema evolution and versioning strategies to support reliable data exchange across systems.
  • Develop and maintain batch and streaming data pipelines using technologies such as Kafka and Spark, handling backpressure, orchestration, retries, and data quality controls.
  • Instrument services with metrics, logs, and traces; contribute to CI/CD pipelines, automated testing, and participate in incident response to ensure system resilience and SLO adherence.
  • Partner closely with Product and cross-functional teams to translate requirements into high-quality technical solutions that deliver business outcomes.
  • Adhere to clean code and architectural standards through code reviews, testing, and Agile development practices, ensuring maintainable and compliant solutions. 

  • Skills and Experience
  • 3–5 years of experience in data-focused software engineering roles.
  • Strong programming skills in Python (including FastAPI) and Scala or Java on the JVM.
  • Proven experience designing and operating RESTful APIs, including secure and versioned interfaces.
  • Solid understanding of data modeling, schema evolution, and serialization technologies such as Avro or Protobuf.
  • Hands-on experience with SQL and NoSQL databases, including query optimization and performance tuning.
  • Experience building and maintaining batch or streaming data systems, with knowledge of streaming patterns and reliability concerns.
  • Familiarity with caching strategies, CI/CD pipelines, and modern monitoring and alerting practices.
  • Proficiency with Git-based workflows, code reviews, and Agile development methodologies.
  • Strong sense of ownership, with pragmatic problem-solving skills and the ability to deliver end-to-end solutions.
  • Excellent communication skills and fluency in English, with the ability to collaborate effectively across product and engineering teams.

  • Nice to have
  • Experience with Apache Airflow for workflow orchestration.
  • Exposure to cloud platforms (preferably AWS) and infrastructure as code using Terraform.
  • Experience with Docker and Kubernetes in production environments.
  • Hands-on knowledge of Kafka and event-driven or microservices architectures.
  • Familiarity with JVM build and tooling ecosystems such as Gradle or Maven.