About the job Data Engineer
Job Description:-
Design, develop, and manage ETL workflows using AWS Glue for structured and unstructured data.
Write clean, efficient, and scalable code in Python for data manipulation and automation tasks.
Use R for statistical modeling, data analysis, and generating insights where applicable.
Collaborate with data scientists, analysts, and stakeholders to understand data requirements and deliver high-quality solutions.
Develop and maintain data models, data dictionaries, and technical documentation.
Optimize data pipelines and processes for performance and scalability.
Ensure data quality, integrity, and security across all data sources and pipelines.
Monitor, troubleshoot, and resolve data pipeline issues in production environments.
Required Skills & Qualifications:
4+ years of experience in data engineering or similar roles.
Strong expertise in AWS Glue, including Glue Jobs, Crawlers, and Glue Studio.
Proficiency in Python for data engineering tasks and scripting.
Practical experience with R for data analysis and reporting.
Good understanding of AWS ecosystem including S3, Lambda, Athena, Redshift, etc.
Experience with data modeling, warehousing concepts, and performance tuning.
Familiarity with version control systems like Git.
Strong analytical and problem-solving skills.
Excellent communication and collaboration skills.
Preferred Qualifications:
Experience with DevOps practices and CI/CD pipelines for data engineering workflows.
Familiarity with data governance, privacy, and compliance requirements.
Exposure to big data technologies such as Spark or Hadoop is a plus.
Knowledge of visualization tools (e.g., Tableau, Power BI) is a plus.