About the job Cloud Data Engineer
Job Overview
We are on an exciting journey to build and scale our advanced analytics practice. We are looking for a
Cloud Data Engineer. The suitable candidate should have demonstrated experience in designing and
implementing ETL solutions on-premises and cloud platforms to support Enterprise data warehouse,
Data Lake and advanced analytics capabilities. Success in this role comes from marrying a strong data
engineering background with product and business acumen to deliver scalable data pipeline and BI
solutions that can enable data for Self Service and Advanced Analytics at TE in a simple and standard
manner.
Primary Responsibilities:
- Design and develop ETL solutions using data warehouse design best practices for Next Generation Analytics platform on Big Data.
- Analyze data requirements, complex source data, data models, and determine the best methods in extracting, transforming, and loading the data into the data staging, warehouse, and other system integration projects.
- Document all ETL and data warehouse processes and flows.
- Develop and deploy ETL job workflow with reliable error/exception handling and rollback.
- Design, Develop, Test, Adapt ETL code & jobs to accommodate changes in source data and newbusiness requirements.
- Manage foundational data administration tasks such as scheduling jobs, troubleshooting job errors, identifying issues with job windows, assisting with Database backups and performance tuning.
- Ensure data quality throughout entire ETL process, including audits and of truth. Create or update technical documentation for transition to support teams.
- Develop, contribute, and adhere to the development of standards and sound procedural practices.
Competencies & Experience Required/Desired:
- 3+ years of Data Engineering experience in ETL design, development & optimization using tools like Talend, AWS Glue, etc.
- 3+ years of experience on Big Data technologies like Hadoop, Spark with Scala or Python for developing flows using Batch, Realtime and Streaming processes3+ years of experience in Relational and No-SQL database design, optimization, and performanc with AWS Redshift, AWS Aurora, Oracle & SAP HANA, BW, etc.
- 3+ years of PL/SQL, Complex SQL Tuning, Stored Procedures, Data Warehousing best practices experience2+ years of experience with AWS Services like S3, EMR, Athena, Glue, Lambda, Redshift, Aurora, etc.
- Solid understanding of ETL pipeline and workflow managements tools such as AWS Glue, AWS
- Kinesis, AWS App Flow, AWS Step Functions, etc.
- Proficiency with modern software development methodologies such as Agile, source control CI/CD, project management and issue tracking with JIRA
- Strong problem-solving capabilities. Results oriented. Relies on fact-based logic for decision making.
- Ability to work with multiple projects and work streams at one time. Must be able to deliver results based upon project deadlines
- Willing to flex daily work schedule to allow for time-zone differences for global team communications.
- Strong interpersonal and communication skills
Education
A masters or bachelors degree in computer science, applied mathematics, software engineering, physics, or related quantitative discipline.