About the job Senior Data Engineer
Job Summary
The Data Engineer will be responsible for designing, developing, and managing data pipelines and architectures. S/he will ensure that data flows seamlessly from multiple sources into databases, data lakes or warehouses and is processed in a way that supports business analysis and decision-making. The ideal candidate will have strong software engineering skills, experience with data processing frameworks, and the ability to optimize and scale large data systems.
Essential Duties and Responsibilities
- Design, build, and maintain scalable data pipelines that support data collection, transformation, and integration from various sources.
- Maintain databases, data warehouses, data lake ensuring data integrity, scalability, and reliability.
- Implement efficient ETL/ELT (Extract, Transform, Load) processes to automate data collection, cleaning, and transformation tasks.
- Integrate data from various internal and external sources, including APIs, cloud storage, third-party systems, and databases.
- Optimize data architectures for performance, scalability, and cost efficiency. Ensure fast and reliable data retrieval.
- Work closely with the data team to understand data needs and deliver reliable, clean, and accessible data solutions.
- Implement data validation and monitoring processes to ensure data accuracy and integrity. Identify and troubleshoot data issues and bottlenecks.
- Manage and optimize cloud-based infrastructure (e.g., AWS, Azure) to support data storage, processing, and retrieval.
- Automate repetitive tasks and processes using scripting languages (e.g., Python and SQL) and workflows.
- Document data architectures, pipelines, and workflows that will be understood by the Information Technology (IT) department. Provide technical support and troubleshooting to the team.
- Demonstrating a strong track record of interacting with the business and technology partners to drive the requirement, technical process into acceptance criteria.
Performance Indicators
Data Pipeline Efficiency and Reliability
Data Quality and Integrity
Data Processing
Data Storage Optimization
Data Compliance & Governance
Scalability of Systems
Cost Efficiency of Data Infrastructure
Time to Resolve Data Issues
Automation and Process Improvement
Collaboration and Communication
System Downtime and Availability
Documentation and Knowledge Sharing
Innovation and Problem-Solving
Learning & Development
Qualifications
- Degree in Computer Engineering / Data Science / Statistics / Physics or any other related field in IT.
- Knowledge in Agile methodology (e.g. Scrum & Kanban)
- Experienced in using cloud computing platforms (Azure)
- Coding Languages
o Python (Apache Spark)
o SQL
o DAX
File types
o parquet; sql; csv; xlsx; json; txt
Systems / Softwares / Platforms
o Github; Databricks; AWS; PowerBI; Linux
Other Qualifications
o ETL/ELT; Integration; API; Statistics; Code/Performance Optimization; Data Wrangling;
Data Pipelines; Data Modeling; Data Quality; Data Governance
Nice to have but not required
o LS Central; Dynamics 365 (Business Central); Microsoft SQL Server; Azure; Tableau;
Javascript; HTML; CSS; AL
Shift: Dayshift
Setup: Onsite
Location: Libis, Quezon City
By Applying, you give consent to collect, store, and/or process personal and/or sensitive information for the purpose of recruitment and employment may it be internal to Cobden & Carter International and/or to its clients.