Job Openings
XTN-DC44512 | BIG DATA ENGINEER
About the job XTN-DC44512 | BIG DATA ENGINEER
The purpose of this job is to provide technical expertise for research, development and modification of extract, transform, load processes and jobs in support of a Big Data infrastructure of our client.
- Health Insurance/HMO
- Enjoy unlimited MadMax Coffee
- Diverse learning & growth opportunities
- Accessible Cloud HR platform (Sprout)
- Above standard leaves
- Research, develop, document, and modify Big Data Lake processes and jobs per data architecture and modeling requirements; collaborate with Data and Analytics data strategists and data scientists
- Collaborate with business stakeholders to understand data needs including data velocity, veracity, and access patterns
- Provide technical expertise to implement Data and Analytics specifications
- Serve on cross-functional project teams and provide the data and big data perspective on executing key deliverables
- Troubleshoot complex, escalated issues including connection, failed jobs, application errors, server alerts and space thresholds within predefined service level agreements (SLAs)
- Proactively maintain and tune all code according to Big Data and EDW best practices to prevent issues
- Review and ensure appropriate documentation for all new development and modifications of the Big Data Lake processes and jobs
- Perform code and process reviews and oversee testing for solutions developed, and ensure integrity and security of institutional data
- Educate business stakeholders on the usage and benefits of the EDW, Big Data Lake and related technologies
- Mentor and guide less experienced team members and provide feedback on project work
- Model behaviors that support the company’s common purpose; ensure guests and team members are supported at the highest level
- Ensure all activities are in compliance with rules, regulations, policies, and procedures
- Complete other duties as assigned
- Bachelor’s degree in computer science, engineering, information technology, or relatedfield, required.
- Minimum five years of technology operations experience required.
- Strong SQL knowledge and skills required
- Strong knowledge of Relational Databases like Oracle, Postgres or SQL Server required
- String knowledge of relational modeling and features including triggers, stored procedures, and constraints required
- Experience with Apache Spark or Spark-streaming, Message Queue technologies and Python required
- Strong knowledge of enterprise data warehouse (EDW) data models with a focus on Star Schema data modeling techniques required
- Strong knowledge of Amazon Web Services (AWS) or similar Cloud Big Data platform preferred
- Excellent analytical skills and the ability to identify solutions to complex data problems
- Ability to provide excellent customer service
- Excellent written and verbal communication skills
- Willingness to learn and embrace new technologies
- Ability to mentor and motivate a diverse team; ensure team and individualaccountability and performance standards are met
- Ability to prioritize, multitask and manage multiple projects successfully in a fast-pacedand dynamic environment
- Strong organizational skills with attention to detail
- Ability to communicate and interact effectively with different levels of the organizationto negotiate, problem solve, complete projects and influence decision making
- Self-motivated with ability to work both independently and within teams in order toestablish and meet deadlines, goals, and objectives