Our project is to develop the data platform, where all the analytical data of the company will be stacked. This is a great opportunity to participate in the launch and operation of a large Kubernetes/Spark/S3 cluster and one of the most interesting BI practices in Eastern Europe.
Responsibilities: * Support and active development of the Data Platform and Hybrid Data Platform (on-prem + Azure Fabric, in progress) * Support for a team of data engineers and analysts
Skills: * Understanding the advantage of GitOps/IaC over manual work * Kubernetes, Helm, ArgoCD, Prometheus, Grafana, Loki, HashiCorp Vault * Apache Spark in Kubernetes, Apache Kafka, Minio/S3, Apache Airflow * Docker (BuildKit), Gitlab, Gitlab CI * Experience with at least one of the most popular programming languages, such as Python, Golang, Java, Scala, etc. Be able writing a code
Will be a plus: * Kerberos, Active directory * Clickhouse * Datahub * Elasticsearch * Experience with any OLAP DB, support, optimize * Security in K8s, HashiCorp Vault, Oauth, OpenID, Keycloak
Will be a significant plus: * Azure, Azure Fabric.
We offer: * The opportunity to work on a large-scale project from scratch * We are not tied to the office, willing to work remotely * Health insurance * Compensation of sports clubs and foreign language schools * Internal training (IT and not only).