Kpler

Senior Data Engineer

Job Description

Data Engineers play a crucial role in creating the architecture and systems needed to collect, store, transform, and access data. They design, build, and maintain large-scale processing systems, data pipelines, and infrastructure that power mission-critical products.

As part of the Ship-Tracking crew, you will work on the core infrastructure behind MarineTraffic.com, a platform that serves 10 billion requests per month and transfers over 90 terabytes of data monthly. You’ll contribute to the largest maritime tracking system in the world and help scale and evolve the backend systems that make real-time vessel monitoring possible.

In this role, you’ll be deeply involved in architecture decisions, product development, and scalable system design. You will be instrumental in launching solutions that bring transparency and innovation to the maritime world, offering real-time and historical insights to millions of global users.


Responsibilities
  • Lead the design, execution, and deployment of engineering solutions to address ship-tracking business cases.
  • Collaborate with stakeholders to define the technical roadmap and deliver high-quality, scalable software.
  • Take ownership of system performance, feature maintenance, issue troubleshooting, and ensuring high availability.
  • Uphold and enforce the highest engineering standards, contributing to continuous improvement of practices and development paradigms.
  • Provide technical mentoring and leadership to team members, helping to elevate engineering capabilities across the team and business.
  • Deliver well-documented, maintainable code and ensure ongoing support and monitoring of developed pipelines and systems.
  • Demonstrate a strong sense of ownership, a growth mindset, and a proactive approach to learning the domain and solving problems independently

  • Skills and Experience
  • Have 8+ years of experience in a data engineering or backend-focused role.
  • Have strong expertise in RDBMS databases such as MSSQL and PostgreSQL.
  • Have strong Python skills, including libraries like pandas and numpy, and follow PEP8 standards.
  • Have experience with streaming data technologies such as Kafka.
  • Have experience with batch data processing and orchestration frameworks.
  • Have worked with CI/CD pipelines and follow Git best practicesConsider testing, logging, monitoring, and alerting as integral part of your deliveries.
  • Have experience working with cloud platforms, especially AWS and have a deep understanding of containerisation tools.
  • Have strong collaboration and communication skills.
  • Demonstrate analytical thinking and a problem-solving mindsetThrive in fast-paced and evolving environments.

  • Nice to have
  • Have experience with Airflow or Apache SparkHave worked with Scala
  • Have used Infrastructure as code tools like Terraform
  • Have experience with GitOps workflows and Continuous Deployment systems
  • Familiar with DBT