Enable job alerts via email!

Streaming Data Platform Engineer

Spyro Soft

Wrocław

On-site

PLN 120,000 - 180,000

Full time

6 days ago
Be an early applicant

Generate a tailored resume in minutes

Land an interview and earn more. Learn more

Start fresh or import an existing resume

Job summary

A leading technology company is seeking a skilled developer to manage our time-series database and develop applications for a cutting-edge project in renewable energy solutions. The role involves configuring database infrastructures, optimizing pipelines, and collaborating with architecture teams to enhance performance. Ideal candidates will have substantial experience with InfluxDB, programming in Python and Java, and a familiarity with various automation and cloud technologies.

Qualifications

  • At least 3 years of experience with InfluxDB.
  • At least 3 years of experience with Python and Java.
  • Familiarity with Ansible, Terraform or Terragrunt.

Responsibilities

  • Configure and operate time-series database infrastructure for telemetry data.
  • Continuously develop and optimize stream-processing pipelines.
  • Automate build and deployment processes in hybrid environments.

Skills

InfluxDB
Python
Java
Ansible
Terraform
TerraGrunt
Apache Kafka
Confluent Cloud
Docker
Kubernetes
ArgoCD
AKS
SCMS
CI/CD
Gitlab-CI
Cloud
Azure

Job description

Requirements:
  • InfluxDB
  • Python/Java
  • Ansible/Terraform/Terragrunt
  • Apache Kafka/Confluent Cloud
  • Docker/Kubernetes/ArgoCD/AKS
  • SCMS
  • CI/CD/Gitlab-CI
  • Cloud
Project description

Our customer is a leading German producer of customized solutions for the self-supply of solar-powered electricity. This includes photovoltaic, energy storage systems as well as cloud technology systems helping individuals to become energetically independent.

Please note that this position involves occasional on-call duty to resolve potential customer-critical incidents.

Main responsibilities:
  • Configure and operate our time-series database infrastructure to store and process telemetry data.
  • Continuously develop and optimize our stream-processing pipelines.
  • Continuously develop and optimize our Python and Java-based applications that are close to the TSDB.
  • Automate build and deployment processes in hybrid environments (e.g., public cloud and self-hosted).
  • Work closely with development teams and architects to enhance performance and features.
  • Python development to adapt to telemetry database components, API design for scalable infrastructure and API development for interfaces from software development projects.
  • Manage, scale, and handle access management for telemetry databases.
Required skills:
  • At least 3 years of experience working with InfluxDB.
  • At least 3 years of experience with programming languages such as Python and Java.
  • Familiarity with Ansible, Terraform or Terragrunt.
  • Knowledge of Apache Kafka or Confluent Cloud.
  • Experience with containers in scalable environments such as Docker, Kubernetes, ArgoCD and AKS.
  • Experience with SCMS and CI/CD systems like Gitlab-CI.
  • Knowledge of cloud environments and services such as Azure.
Recruitment process:
  • Recruitment screening and experience survey (ca. 1 hour)
  • Profile-specific offline assignment (approx. 2–3 evenings of concentrated work, depending on experience)
  • Client technical interview (ca. 1 hour)
  • Client team interview (ca. 1 hour)
  • Offer meeting (ca. 30 mins)
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.