Job Openings Cloud Data Engineer

About the job Cloud Data Engineer

Job Overview

We are on an exciting journey to build and scale our advanced analytics practice. We are looking for a

Cloud Data Engineer. The suitable candidate should have demonstrated experience in designing and

implementing ETL solutions on-premises and cloud platforms to support Enterprise data warehouse,

Data Lake and advanced analytics capabilities. Success in this role comes from marrying a strong data

engineering background with product and business acumen to deliver scalable data pipeline and BI

solutions that can enable data for Self Service and Advanced Analytics at TE in a simple and standard

manner.

Primary Responsibilities:

  • Design and develop ETL solutions using data warehouse design best practices for Next Generation Analytics platform on Big Data.
  • Analyze data requirements, complex source data, data models, and determine the best methods in extracting, transforming, and loading the data into the data staging, warehouse, and other system integration projects.
  • Document all ETL and data warehouse processes and flows.
  • Develop and deploy ETL job workflow with reliable error/exception handling and rollback.
  • Design, Develop, Test, Adapt ETL code & jobs to accommodate changes in source data and newbusiness requirements.
  • Manage foundational data administration tasks such as scheduling jobs, troubleshooting job  errors, identifying issues with job windows, assisting with Database backups and performance tuning.
  • Ensure data quality throughout entire ETL process, including audits and of truth. Create or update technical documentation for transition to support teams.
  • Develop, contribute, and adhere to the development of standards and sound procedural practices.

Competencies & Experience Required/Desired:

  • 3+ years of Data Engineering experience in ETL design, development & optimization using tools like Talend, AWS Glue, etc.
  • 3+ years of experience on Big Data technologies like Hadoop, Spark with Scala or Python for developing flows using Batch, Realtime and Streaming processes3+ years of experience in Relational and No-SQL database design, optimization, and performanc with AWS Redshift, AWS Aurora, Oracle & SAP HANA, BW, etc.
  • 3+ years of PL/SQL, Complex SQL Tuning, Stored Procedures, Data Warehousing best practices experience2+ years of experience with AWS Services like S3, EMR, Athena, Glue, Lambda, Redshift, Aurora, etc.
  • Solid understanding of ETL pipeline and workflow managements tools such as AWS Glue, AWS
  •  Kinesis, AWS App Flow, AWS Step Functions, etc.
  • Proficiency with modern software development methodologies such as Agile, source control CI/CD, project management and issue tracking with JIRA
  • Strong problem-solving capabilities. Results oriented. Relies on fact-based logic for decision making.
  • Ability to work with multiple projects and work streams at one time. Must be able to deliver results based upon project deadlines
  • Willing to flex daily work schedule to allow for time-zone differences for global team communications.
  • Strong interpersonal and communication skills

Education

A masters or bachelors degree in computer science, applied mathematics, software engineering, physics, or related quantitative discipline.