Datumo
Java Developer Want-To-Be
Datumo, Snowflake, Arizona, United States, 85937
Datumo specializes in providing
Big Data and Cloud
consulting services to clients from all over the world, primarily in Western Europe, Poland, and the USA. Core industries we support include
e-commerce, telecommunications, and life science.
Our team consists of
exceptional people
whose commitment allows us to conduct
highly demanding projects.Our team members tend to stick around for more than 3 years, and when a project wraps up, we don't let them go - we embark on a journey to discover exciting new challenges for them. It's not just a workplace; it's a community that grows together!What we expect:
Must-have:
At least 2 years of commercial experience in software development with a strong focus on JavaReadiness to learn and transition into the Big Data domain.Hands-on experience with a selected cloud provider (GCP, Azure).Strong focus on data and databases, with good knowledge of SQL and experience with RDBMS (e.g., MariaDB, Oracle).Experience working with Kubernetes and Docker for deployment.Familiarity with tools for testing (Junit, AssertJ, Mockito), event processing (Kafka, RabbitMQ), and data (ElasticSearch, Redis).Passion for writing clean code and using established design patterns.Understanding of concepts like domain-driven design, test patterns, and common programming principles.Experience in using CI/CD tools.Proven record of collaborating with businesses and corporations.Contribution to internal projects.Knowledge of English at B2 level, communicative Polish.Nice to have:
Experience with data warehouses like Google BigQuery, Databricks, Snowflake, etc.Familiarity with distributed data processing frameworks running on JVMs (e.g., Apache Spark, Flink).Proficiency in JVM programming languages.Experience with Apache Airflow or similar pipeline orchestrators.Experience in Machine Learning projects.Contribution to open-source Big Data tools built with Java.Sharing knowledge by writing articles for the company blog or contributing to open-source projects.What’s on offer:
100% remote work, with workation opportunity20 free daysOnboarding with a dedicated mentorProject switching possible after a certain periodIndividual budget for training and conferencesBenefits: Medicover private medical care, co-financing of the Medicover Sport cardOpportunity to learn English with a native speakerRegular company trips and informal get-togethersDevelopment opportunities in Datumo:Participation in industry conferencesEstablishing Datumo's online brand presenceSupport in obtaining certifications (e.g. GCP, Azure, Snowflake)Involvement in internal initiatives, like building technological roadmapsAccess to internal technological training repositoriesDiscover our exemplary projects:
IoT data ingestion to cloudThe project integrates data from edge devices into the cloud using Azure services. The platform supports data streaming via either the IoT Edge environment with Java or Python modules, or direct connection using Kafka protocol to Event Hubs. It also facilitates batch data transmission to ADLS. Data transformation from raw telemetry to structured tables is done through Spark jobs in Databricks or data connections and update policies in Azure Data Explorer.Petabyte-scale data platform migration to Google CloudThe goal of the project is to improve scalability and performance of the data platform by transitioning over a thousand active pipelines to GCP. The main focus is on rearchitecting existing Spark applications to either Cloud Dataproc or Cloud BigQuery SQL, depending on the Client’s requirements and automate it using Cloud Composer.Data analytics platform for investing companyThe project centers on developing and overseeing a data platform for an asset management company focused on ESG investing. Databricks is the central component. The platform, built on Azure cloud, integrates various Azure services for diverse functionalities. The primary task involves implementing and extending complex ETL processes that enrich investment data, using Spark jobs in Scala. Integrations with external data providers, as well as solutions for improving data quality and optimizing cloud resources, have been implemented.Realtime Consumer Data PlatformThe initiative involves constructing a consumer data platform (CDP) for a major Polish retail company. Datumo actively participates from the project’s start, contributing to planning the platform’s architecture. The CDP is built on Google Cloud Platform (GCP), utilizing services like Pub/Sub, Dataflow, and BigQuery. Open-source tools, including a Kubernetes cluster with Apache Kafka, Apache Airflow, and Apache Flink, are used to meet specific requirements. This combination offers significant possibilities for the platform.If you like what we do and you dream about creating this world with us - don’t wait, apply now!
#J-18808-Ljbffr
Big Data and Cloud
consulting services to clients from all over the world, primarily in Western Europe, Poland, and the USA. Core industries we support include
e-commerce, telecommunications, and life science.
Our team consists of
exceptional people
whose commitment allows us to conduct
highly demanding projects.Our team members tend to stick around for more than 3 years, and when a project wraps up, we don't let them go - we embark on a journey to discover exciting new challenges for them. It's not just a workplace; it's a community that grows together!What we expect:
Must-have:
At least 2 years of commercial experience in software development with a strong focus on JavaReadiness to learn and transition into the Big Data domain.Hands-on experience with a selected cloud provider (GCP, Azure).Strong focus on data and databases, with good knowledge of SQL and experience with RDBMS (e.g., MariaDB, Oracle).Experience working with Kubernetes and Docker for deployment.Familiarity with tools for testing (Junit, AssertJ, Mockito), event processing (Kafka, RabbitMQ), and data (ElasticSearch, Redis).Passion for writing clean code and using established design patterns.Understanding of concepts like domain-driven design, test patterns, and common programming principles.Experience in using CI/CD tools.Proven record of collaborating with businesses and corporations.Contribution to internal projects.Knowledge of English at B2 level, communicative Polish.Nice to have:
Experience with data warehouses like Google BigQuery, Databricks, Snowflake, etc.Familiarity with distributed data processing frameworks running on JVMs (e.g., Apache Spark, Flink).Proficiency in JVM programming languages.Experience with Apache Airflow or similar pipeline orchestrators.Experience in Machine Learning projects.Contribution to open-source Big Data tools built with Java.Sharing knowledge by writing articles for the company blog or contributing to open-source projects.What’s on offer:
100% remote work, with workation opportunity20 free daysOnboarding with a dedicated mentorProject switching possible after a certain periodIndividual budget for training and conferencesBenefits: Medicover private medical care, co-financing of the Medicover Sport cardOpportunity to learn English with a native speakerRegular company trips and informal get-togethersDevelopment opportunities in Datumo:Participation in industry conferencesEstablishing Datumo's online brand presenceSupport in obtaining certifications (e.g. GCP, Azure, Snowflake)Involvement in internal initiatives, like building technological roadmapsAccess to internal technological training repositoriesDiscover our exemplary projects:
IoT data ingestion to cloudThe project integrates data from edge devices into the cloud using Azure services. The platform supports data streaming via either the IoT Edge environment with Java or Python modules, or direct connection using Kafka protocol to Event Hubs. It also facilitates batch data transmission to ADLS. Data transformation from raw telemetry to structured tables is done through Spark jobs in Databricks or data connections and update policies in Azure Data Explorer.Petabyte-scale data platform migration to Google CloudThe goal of the project is to improve scalability and performance of the data platform by transitioning over a thousand active pipelines to GCP. The main focus is on rearchitecting existing Spark applications to either Cloud Dataproc or Cloud BigQuery SQL, depending on the Client’s requirements and automate it using Cloud Composer.Data analytics platform for investing companyThe project centers on developing and overseeing a data platform for an asset management company focused on ESG investing. Databricks is the central component. The platform, built on Azure cloud, integrates various Azure services for diverse functionalities. The primary task involves implementing and extending complex ETL processes that enrich investment data, using Spark jobs in Scala. Integrations with external data providers, as well as solutions for improving data quality and optimizing cloud resources, have been implemented.Realtime Consumer Data PlatformThe initiative involves constructing a consumer data platform (CDP) for a major Polish retail company. Datumo actively participates from the project’s start, contributing to planning the platform’s architecture. The CDP is built on Google Cloud Platform (GCP), utilizing services like Pub/Sub, Dataflow, and BigQuery. Open-source tools, including a Kubernetes cluster with Apache Kafka, Apache Airflow, and Apache Flink, are used to meet specific requirements. This combination offers significant possibilities for the platform.If you like what we do and you dream about creating this world with us - don’t wait, apply now!
#J-18808-Ljbffr