Job Description
Join our client in embarking on an ambitious data transformation journey using Databricks, guided by best practice data governance and architectural principles. To support this, we are recruiting talented data engineers. As a major UK energy provider, our client is committed to 100% renewable energy and sustainability, focusing on delivering exceptional customer experiences.
It is initially a 3-month contract with potential extension. The role is hybrid, with one day a week based in their Nottingham office, negotiable. It is a full-time role, 37 hours per week.
Accountabilities:
- Develop and maintain scalable, efficient data pipelines within Databricks, evolving them as requirements and technologies change.
- Build and manage an enterprise data model within Databricks.
- Integrate new data sources into the platform using batch and streaming processes, adhering to SLAs.
- Create and maintain documentation for data pipelines and systems, following security and monitoring protocols.
- Ensure data quality and reliability processes are effective to maintain trust in the data.
- Take ownership of complex data engineering projects and develop solutions aligned with business needs.
- Work closely with stakeholders to manage requirements.
- Coach and mentor team members, fostering a culture of innovation and peer review to ensure best practices.
Knowledge and Skills:
- Extensive experience with Python, including advanced concepts like decorators, protocols, functools, context managers, and comprehensions.
- Strong understanding of SQL, database design, and data architecture.
- Experience with Databricks and/or Spark.
- Knowledge of data governance, data cataloguing, data quality principles, and related tools.
- Skilled in data extraction, joining, and aggregation tasks, especially with big data and real-time data using Spark.
- Proficient in data cleansing and transforming data for analysis.
- Understanding of data storage concepts and logical structures like data warehousing.
- Ability to write production-quality, repeatable code for data pipelines, using templating and parameterization.
- Ability to recommend data pipeline designs based on business needs.
- Experience with data migration is a plus.
- Open to new technologies and ways of working.
- Self-motivated, goal-oriented, and proactive.
- Strong troubleshooting skills and problem-solving ability.
- Experience with Git/version control, large legacy codebases, unit and integration testing, CI/CD, and software development best practices.
- Attention to detail and curiosity about data.
- Strong understanding of Linux tooling and concepts.
- Knowledge and experience with AWS is essential.
Note: Successful applicants will undergo pre-employment checks, including a satisfactory DBS check. This vacancy is advertised by Rullion Ltd, an employment business. Rullion is committed to equal opportunities.