Job Openings Senior Data Engineer

About the job Senior Data Engineer

Job Summary

The Data Engineer will be responsible for designing, developing, and managing data pipelines and architectures. S/he will ensure that data flows seamlessly from multiple sources into databases, data lakes or warehouses and is processed in a way that supports business analysis and decision-making. The ideal candidate will have strong software engineering skills, experience with data processing frameworks, and the ability to optimize and scale large data systems.

Essential Duties and Responsibilities

  • Design, build, and maintain scalable data pipelines that support data collection, transformation, and integration from various sources.
  • Maintain databases, data warehouses, data lake ensuring data integrity, scalability, and reliability.
  • Implement efficient ETL/ELT (Extract, Transform, Load) processes to automate data collection, cleaning, and transformation tasks.
  • Integrate data from various internal and external sources, including APIs, cloud storage, third-party systems, and databases.
  • Optimize data architectures for performance, scalability, and cost efficiency. Ensure fast and reliable data retrieval.
  • Work closely with the data team to understand data needs and deliver reliable, clean, and accessible data solutions.
  • Implement data validation and monitoring processes to ensure data accuracy and integrity. Identify and troubleshoot data issues and bottlenecks.
  • Manage and optimize cloud-based infrastructure (e.g., AWS, Azure) to support data storage, processing, and retrieval.
  • Automate repetitive tasks and processes using scripting languages (e.g., Python and SQL) and workflows.
  • Document data architectures, pipelines, and workflows that will be understood by the Information Technology (IT) department. Provide technical support and troubleshooting to the team.
  • Demonstrating a strong track record of interacting with the business and technology partners to drive the requirement, technical process into acceptance criteria.

Performance Indicators

Data Pipeline Efficiency and Reliability

Data Quality and Integrity

Data Processing

Data Storage Optimization

Data Compliance & Governance

Scalability of Systems

Cost Efficiency of Data Infrastructure

Time to Resolve Data Issues

Automation and Process Improvement

Collaboration and Communication

System Downtime and Availability

Documentation and Knowledge Sharing

Innovation and Problem-Solving

Learning & Development

Qualifications

  • Degree in Computer Engineering / Data Science / Statistics / Physics or any other related field in IT.
  • Knowledge in Agile methodology (e.g. Scrum & Kanban)
  • Experienced in using cloud computing platforms (Azure)
  • Coding Languages

o Python (Apache Spark)

o SQL

o DAX

File types

o parquet; sql; csv; xlsx; json; txt

Systems / Softwares / Platforms

o Github; Databricks; AWS; PowerBI; Linux

Other Qualifications

o ETL/ELT; Integration; API; Statistics; Code/Performance Optimization; Data Wrangling;

Data Pipelines; Data Modeling; Data Quality; Data Governance

Nice to have but not required

o LS Central; Dynamics 365 (Business Central); Microsoft SQL Server; Azure; Tableau;

Javascript; HTML; CSS; AL

Shift: Dayshift
Setup: Onsite
Location: Libis, Quezon City

By Applying, you give consent to collect, store, and/or process personal and/or sensitive information for the purpose of recruitment and employment may it be internal to Cobden & Carter International and/or to its clients.