Requirements: English
Company: Datumo
Region: Gdansk , Pomeranian Voivodeship
Datumo specializes in providing Big Data and Cloud consulting services to clients from all over the world, primarily in Western Europe, Poland and the USA. Core industries we support include e-commerce, telecommunications and life science. Our team consists of exceptional people whose commitment allows us to conduct highly demanding projects .
Our team members tend to stick around for more than 3 years, and when a project wraps up, we don''t let them go - we embark on a journey to discover exciting new challenges for them. It''s not just a workplace; it''s a community that grows together!
at least 3 years of commercial experience in Big Data
proven record with a selected cloud provider GCP preferred, Azure or AWS
good knowledge of JVM languages - Scala or Java or Kotlin
good knowledge of Python
good knowledge of SQL
experience with BigQuery, Snowflake, Hive or similar distributed datastore
designing and implementing Big Data systems following best practices
ensuring solution quality through automatic tests, CI / CD and code review
English proficiency at B2 level, communicative in Polish
experience in Snowflake/Databricks platform
knowledge of Apache Kafka, Docker and Kubernetes technologies
experience in Machine Learning projects
willingness to share knowledge (conferences,articles,open-source projects)
100% remote work, with workation opportunity
~project switching possible after a certain period
~ Medicover private medical care, co-financing of the Medicover Sport card
~ opportunity to learn English with a native speaker
~ GCP, Azure, Snowflake)
Discover our exemplary projects:
The project integrates data from edge devices into the cloud using Azure services. The platform supports data streaming via either the IoT Edge environment with Java or Python modules, or direct connection using Kafka protocol to Event Hubs. It also facilitates batch data transmission to ADLS. Data transformation from raw telemetry to structured tables is done through Spark jobs in Databricks or data connections and update policies in Azure Data Explorer.
The goal of the project is to improve scalability and performance of the data platform by transitioning over a thousand active pipelines to GCP. The main focus is on rearchitecting existing Spark applications to either Cloud Dataproc or Cloud BigQuery SQL, depending on the Clients requirements and automate it using Cloud Composer.
The project centers on developing and overseeing a data platform for an asset management company focused on ESG investing. The platform, built on Azure cloud, integrates various Azure services for diverse functionalities. The primary task involves implementing and extending complex ETL processes that enrich investment data, using Spark jobs in Scala. Integrations with external data providers, as well as solutions for improving data quality and optimizing cloud resources, have been implemented.
The initiative involves constructing a consumer data platform (CDP) for a major Polish retail company. Datumo actively participates from the projects start, contributing to planning the platforms architecture. Technical interview - 60 minutes
Find out more by visiting our website -