Software Engineer (Java)

Job description

About us

Who we are? We are Big Data experts, working with international clients, creating and leading innovative projects related to the Big Data environment. We offer tailor-made solutions. It does not matter whether we are talking about building a Data Lake, conducting training in the field of data management, or performing detailed Big Data analysis. We don’t just focus on one technology, instead, we specialize in a whole range of open-source and public cloud tools. Our team brings together over 130 specialists in their fields. We have participated in dozens of conferences, and written countless amounts of code, we are the organizers of Big Data Tech Summit Warsaw, the largest Polish conference related to Big Data topics. We run webinars, share knowledge on blogs, create whitepapers and more. Why? Because we believe that Big Data is an indispensable future of business.
Thanks to that, we always select the most optimal Big Data solutions.


We are working on the development of a modern open-source tool Marquez for Data Lineage and Data Discovery. Marquez is a data dog-like tool for data engineers & data scientists to track, monitor, and search their datasets and the pipelines that process them. More and more companies already use Marquez and even more need something like this to handle the chaos in their production-grade data lakes with hundreds or thousands of datasets.


As a Senior Software Engineer, you will develop an open-source code and be a part of the community that can shift data engineering to a new era. Public Github environment and discussions open to the rest of the world limit the number of meetings to a minimum.


  • Lead implementation of OpenLineage integrations for data tools

  • Write simple, understandable, testable, and efficient code with a focus on solid technical stability and a great user experience

Technologies used:
  • Java, Scala, optionally Python
  • Open Source and proprietary data processing engines and tools, like: Apache Flink, Apache Spark, Apache Beam, Apache Iceberg, Delta Lake, Trino, Snowflake, BigQuery, Airflow etc.
  • Github


  • Very good knowledge of Java
  • Strong knowledge of Apache Flink
  • Good knowledge of SQL
  • Experience working with databases or data processing systems
  • Knowledge of the big data landscape
  • Ability to work independently and focus on impactful features
Nice to have:
  • Experience as a contributor to an open-source community
  • Ability to clearly communicate over text: work with a team spanning multiple time zones

We offer
  • Salary : 150-200 PLN net + VAT/h B2B (depending on knowledge and experience)

  • 100% remote work

  • Elastic working hours

  • Possibility to work from the office located in the heart of Warsaw

  • Opportunity to learn and develop with the best Big Data specialists in Poland

  • International projects

  • Possibility of conducting workshops and training

  • Clear career path and certifications

  • Co-financing sport card

  • Co-financing health care

  • All equipment needed for work