Job Openings Senior Data Engineer

About the job Senior Data Engineer

Tech Lead Data Engineering

Position Overview:
As a Tech Lead Data Engineering, you will act as the key technical advisor for stakeholders, take ownership of data engineering projects, manage technical debt, and mentor junior team members. You will contribute to architecture, design, and development of robust data solutions.

Key Responsibilities

Leadership & Collaboration:

  • Serve as the primary technical contact for stakeholders and oversee data project execution.
  • Mentor junior engineers and lead design discussions.

Data Engineering:

  • Develop batch and real-time data pipelines using Databricks, PySpark, and DBT.
  • Clean, transform, and optimize raw data for analytics, ensuring cost and performance efficiency.
  • Design and manage database schemas and optimize storage formats (e.g., Parquet, Delta Lake).
  • Implement data quality checks, lineage documentation, and advanced partitioning/indexing strategies.

Collaboration:

  • Partner with data scientists and analysts to ensure seamless data access for analytics and machine learning.
  • Work with Platform Engineers to align on best practices.

Security & Compliance:

  • Enforce role-based access controls, data encryption, and compliance with data privacy regulations.

Infrastructure & Configuration:

  • Manage Azure cloud infrastructure, including Data Lake, Blob Storage, and Databricks clusters.
  • Build and maintain pipelines using Azure Data Factory.

Documentation & Continuous Improvement:

  • Maintain detailed documentation of pipelines, processes, and tools.
  • Propose and implement optimizations for systems and workflows.

Education & Experience

  • Masters or PhD in Engineering, Computer Science, or related fields.
  • 5+ years of experience in data engineering, particularly in large-scale environments.
  • Proficiency with cloud ecosystems, especially Azure.

Technical Skills

Programming & Tools:

  • Advanced Python (e.g., Pandas, dependency management) and SQL.
  • Strong software engineering practices (e.g., design patterns, clean code, testing).

Data Processing & Storage:

  • Expertise in PySpark, Delta Lake, and storage formats (e.g., Parquet, JSON).
  • Proficiency with Azure Data Factory, Databricks, and database management (SQL and NoSQL).

Data Architecture:

  • Strong skills in schema design, ETL workflows, and modeling (Dimensional, 3NF, Data Vault).

DevOps & Automation:

  • Experience with CI/CD tools (Azure DevOps) and Infrastructure as Code (ARM templates, Terraform preferred).

Security & Monitoring:

  • Knowledge of role-based access, encryption, and monitoring tools (e.g., Azure Monitor).