About The Position

Who are we?

Skai (formerly Kenshoo) is a leading omnichannel marketing platform that leverages advanced AI and machine learning to deliver intelligent solutions based on data with performance media, enabling smarter decision-making, increased efficiency, and maximized returns - Revenue enabler for businesses around the world. Its partners include Google, Meta, Amazon, and Microsoft and more. ~$7 billion in ad spending is managed on the SkaiTM platform every year.

Established in 2006, We’re 700 employees strong. We work hybrid with a great home/ office work mix.


What will you do?

Join a team that builds and operates the core data services behind Skai’s products.

As a Senior Data Engineer, you will design, build and maintain:

  • Data pipelines and data services
  • Microservices and internal APIs
  • The infrastructure and tooling that keep our data systems reliable, observable, and scalable

This is a hands-on role that combines Data Engineering, Backend Development and DevOps.

It’s ideal for someone who enjoys working across layers: from schema design and data flows, through microservice code, to Kubernetes, CI/CD and observability.


Responsibilities:

  • Design and maintain robust infrastructure for large-scale data processing and streaming systems.
  • Develop automation and deployment processes using CI/CD pipelines.
  • Build and operate Kubernetes-based environments and containerized workloads.
  • Collaborate with data engineers to optimize performance, cost, and reliability of data platforms.
  • Design and develop REST-API microservices.
  • Troubleshoot and resolve complex issues in production and staging environments.
  • Drive initiatives that enhance observability, scalability, and developer productivity.
  • Lead by example - share knowledge, mentor teammates, and promote technical best practices.

Requirements

  • 5 years of experience as a Data Engineer, Backend Developer, or DevOps.
  • 5+ years of experience with Python/Java microservices (Flask, Spring, Dropwizard) and component testing.
  • Deep understanding of Kubernetes, Docker, and container orchestration.
  • Hands-on experience with CI/CD pipelines (e.g., Jenkins, GitHub Actions).
  • Proven experience with Snowflake, MySQL, RDS, or similar databases.
  • Familiarity with streaming systems(e.g., Kafka, FireHose), databases, or data pipelines.
  • Self-learner, proactive, and passionate about improving systems and automation.
  • Strong communication skills and a collaborative, team-oriented mindset.


Advantages:

  • Experience with Kafka, Airflow or other data processing tools.
  • Knowledge of Terraform, Pulumi, or other IaC frameworks.
  • Familiarity with Datadog, Prometheus or other observability tools.
  • Experience with AWS (Lambda, EKS, EC2, Step functions, SQS).
  • Working with or building AI-driven tools.

Apply for this position

Click on the button to load the form

More Information