Job Openings Lead Data Engineer (Romania / Ukraine)

About the job Lead Data Engineer (Romania / Ukraine)

Role: Lead Data Engineer

Location: Romania, Ukraine (Remote)

Years of Experience: 5-7 years

Pay: $84,000 - $90,000 PA

Required Skill: Python, SQL,Databricks, Snowflake, ETL, AWS, GCP, Airbyte, Postgres, Kafka, Sisense, CircleCI, Grafana, Kubernetes (EKS)

Language Required: English C1 Level

Job Description

This role emphasizes creating robust, scalable, and secure data lake, ingestion, and distribution systems. As a key leader, you will work across the data stack and collaborate with cross-functional teams, setting high standards in data governance, quality, and security.

Key Responsibilities

  • Data Lake and Pipelines: Design and implement features for ingestion, enrichment, and transformation, ensuring availability, reliability, and scalability.
  • System Design and Project Alignment: Drive system design, creating alignment and feasibility for projects across teams.
  • Full Data Stack: Work across ETL processes, data warehousing, visualization, and cloud infrastructure.
  • Data Management and Governance: Define and implement best practices to ensure data quality, security, and consistency.
  • Optimization: Develop and optimize Spark jobs, notebooks, and pipelines in Databricks.
  • Collaboration and Mentorship: Partner with the Chief Architect, mentor engineers, and support DataOps culture.
  • Stakeholder Communication: Communicate data strategies and plans to stakeholders across the organization.

Required Skills and Experience

  • Data Engineering: 5+ years in data pipeline design and implementation using Databricks and Python (or PySpark).
  • SQL & Visualization: Proficiency in SQL and visualization tools such as Sisense or Power BI.
  • Cloud Platforms: Experience with AWS or GCP, focusing on cloud-native data engineering.
  • ETL and Data Governance: Expertise in ETL processes and data governance principles to maintain quality and consistency.

Must-Have:

  • Python, SQL, Databricks, ETL processes
    AWS or GCP
  • Visualization tools (Sisense or similar)
  • Airbyte or comparable ETL tools
  • Terraform for environment templating, CI/CD tools (CircleCI, GitHub Actions

Nice to Have:

  • Kedro, Kafka, and data mesh design
  • Postgres, Terraform, CircleCI, Grafana
  • Knowledge of microservices and data modeling

Additional Skills

  • Technical: Designing large-scale data systems, SQL/NoSQL databases, and familiarity with streaming services like Kafka.
  • Soft Skills: Strong client-facing abilities, strategic planning, and stakeholder management.

Our Tech Stack

  • Python, SQL, Databricks, Snowflake, Airbyte, Postgres, Kafka, Sisense, CircleCI, Grafana, Kubernetes (EKS)

(Knowledge of deprecated systems like Druid or Datadog is a plus)