Description du poste:

The opportunity

We are looking for an experienced data engineer who will help us process, gather, and analyze large-scale datasets. You will be working in the Datalab team that is responsible for designing, developing, and deploying machine learning & deep learning solutions in our products. The core of the work will be focused on ensuring the availability of appropriate datasets to build sound, robust and updated models. You’ll also make sure that the performance and stability of these models are appropriately monitored and measured. Besides tasks associated with dataset management, an important part of the work concerns the development of appropriate approaches and tools to ensure the professional management of our models in production.

Your day-to-day at Egnyte

  • Working closely within a team of machine learning engineers to apply innovative solutions to large unstructured datasets
  • Sourcing, analyzing, and processing data from various sources (e.g., BigQuery and BigTable) to build cohesive datasets
  • Developing tools to manage our datasets, data pipelines, and other artifacts created during the whole model lifecycle
  • Supporting the deployment of machine learning models on our infrastructure, including containerization, instrumentation, and versioning
  • Supporting the whole lifecycle of our machine learning models, including gathering data for (re)training, A/B testing, deployment, monitoring, retraining, and redeployments
  • Working closely within a distributed team to analyze and apply innovative solutions over billions of documents
  • Communicating your approach and results to a wider audience through articles and presentations

About you

  • Documented success with data engineering at scale in a SaaS or Cloud environment
  • Ability to implement efficient data import, cleansing, and transformation functions on large scale
  • Experience with large datasets and distributed computing, especially with the Google Cloud Platform
  • Good knowledge of any major cloud provider environment (such as GCP, AWS, Azure)
  • Excellence in SQL
  • Fluency in Python, Docker, Kubernetes
  • Solid English skills to effectively communicate with other team members

Bonus skills

  • Sound knowledge of machine learning & deep learning
  • Experience with advanced analytical modeling and statistical forecasting techniques
  • Good understanding of No–SQL and Graph databases
  • Knowledge of Java, Scala or Golang programming languages
  • Familiarity with Kubeflow
  • Advanced communication skills, especially with regards to knowledge transfer

The offer

  • Salary from 12.000 to 16.000 PLN Net + VAT (B2B) depending on skills and experience
  • 100% remote work possible
  • Stock options
  • Your own Egnyte account with lifetime access
  • 4000 PLN Gross conference budget per person and additional 4 training days each year to boost your skills
  • Apple or Dell (your call) equipment with a Linux distro of your choice or Windows
  • Flexible working hours or remote work days if you need them
  • MultiKafeteria: you can choose a MultiSport card or gift cards every month
  • Private medical healthcare
  • In-house English classes
  • Team initiatives and events


Python SQL Docker Big Data Kubernetes


Other Development
Poland, Poznań
Type d'emploi
À distance
Type de contrat
Expérience en années
1 - 11  Années
Niveau d'anglais
Avancé C1-C2
Salaire Net
16 000 PLN