Data Engineer Databricks

4 tygodni temu


Warszawa, mazowieckie, Polska Addepto Pełny etat 15 zł - 120 zł

Addepto is a leading AI consulting and data engineering company that builds scalable, ROI-focused AI solutions for some of the world’s largest enterprises and pioneering startups, including Rolls Royce, Continental, Porsche, ABB, and WGU. With our exclusive focus on Artificial Intelligence and Big Data, we help organizations unlock the full potential of their data through systems designed for measurable business impact and long-term growth.


Beyond client projects, we have developed our own product offerings born from real-life client insights and challenges. We are also actively releasing open-source solutions to the community, transforming practical experience into tools that benefit the broader AI ecosystem. This commitment to scalable innovation, proven ROI delivery, and knowledge sharing has earned us recognition by Forbes as one of the top 10 AI consulting companies worldwide.


As a Data Engineer, you will have the exciting opportunity to work with a team of technology experts on challenging projects across various industries, leveraging cutting-edge technologies. Here are some of the projects we are seeking talented individuals to join:

  • Design and development of a universal data platform for global aerospace companies. This Azure and Databricks powered initiative combines diverse enterprise and public data sources. The data platform is at the early stages of the development, covering design of architecture and processes as well as giving freedom for technology selection.

  • Data Platform Transformation for energy management association body. This project addressed critical data management challenges, boosting user adoption, performance, and data integrity. The team is implementing a comprehensive data catalog, leveraging Databricks and Apache Spark/PySpark, for simplified data access and governance. Secure integration solutions and enhanced data quality monitoring, utilizing Delta Live Table tests, established trust in the platform. The intermediate result is a user-friendly, secure, and data-driven platform, serving as a basis for further development of ML components.

  • Design of the data transformation and following data ops pipelines for global car manufacturer. This project aims to build a data processing system for both real-time streaming and batch data. We’ll handle data for business uses like process monitoring, analysis, and reporting, while also exploring LLMs for chatbots and data analysis. Key tasks include data cleaning, normalization, and optimizing the data model for performance and accuracy.


Your main responsibilities:

  • Design scalable data processing pipelines for streaming and batch processing using Big Data technologies like Databricks, Airflow and/or Dagster.

  • Contribute to the development of CI/CD and MLOps processes.

  • Develop applications to aggregate, process, and analyze data from diverse sources.

  • Collaborate with the Data Science team on Machine Learning projects, including text/image analysis and predictive model building.

  • Develop and organize data transformations using Databricks/DBT and Apache Airflow.

  • Translate business requirements into technical solutions and ensure optimal performance and quality.


What you’ll need to succeed in this role:

  • At least 3 years of commercial experience implementing, developing, or maintaining Big Data systems.

  • Strong programming skills in Python: writing a clean code, OOP design.

  • Experience in designing and implementing data governance and data management processes.

  • Familiarity with Big Data technologies like Airflow or Dagster, Databricks, Spark and DBT.

  • Experience implementing and deploying solutions in cloud environments (with a preference for Azure).

  • Knowledge of how to build and deploy Power BI reports and dashboards for data visualization.

  • Excellent understanding of dimensional data and data modeling techniques.

  • Excellent communication skills and consulting experience with direct interaction with clients.

  • Ability to work independently and take ownership of project deliverables.

  • Master’s or Ph.D. in Computer Science, Data Science, Mathematics, Physics, or a related field.


Discover our perks & benefits:

  • Work in a supportive team of passionate enthusiasts of AI & Big Data.

  • Engage with top-tier global enterprises and cutting-edge startups on international projects.

  • Enjoy flexible work arrangements, allowing you to work remotely or from modern offices and coworking spaces.

  • Accelerate your professional growth through career paths, knowledge-sharing initiatives, language classes, and sponsored training or conferences, including a partnership with Databricks, which offers industry-leading training materials and certifications.

  • Choose from various employment options: B2B, employment contracts, or contracts of mandate.

  • Make use of 20 fully paid days off available for B2B contractors and individuals under contracts of mandate.

  • Participate in team-building events and utilize the integration budget.

  • Celebrate work anniversaries, birthdays, and milestones.

  • Access medical and sports packages, eye care, and well-being support services, including psychotherapy and coaching.

  • Get full work equipment for optimal productivity, including a laptop and other necessary devices.

  • With our backing, you can boost your personal brand by speaking at conferences, writing for our blog, or participating in meetups.

  • Experience a smooth onboarding with a dedicated buddy, and start your journey in our friendly, supportive, and autonomous culture.


Are you interested in Addepto and would like to join us?


Get in touch We are looking forward to receiving your application. Would you like to know more about us?

Visit our website (career page) and social media (Facebook, LinkedIn, Instagram).


  • Data Engineer Databricks

    3 tygodni temu


    Warszawa, mazowieckie, Polska 7N Pełny etat 28 zł - 560 zł

    Data Engineer (Databricks)We are looking for an experienced Data Engineer for our client in the pharmaceutical industry. The role focuses on developing and professionalizing data foundations within the organization.Work mode: 100% remote TasksDesign, develop, test, and maintain robust data pipelines and ETL processes.Implement and manage CI/CD pipelines to...


  • Warszawa, mazowieckie, Polska Upvanta sp. z o.o. Pełny etat 26 zł - 880 zł

    Databricks Data Engineer / ArchitectWe’re looking for an experienced Senior Data Engineer to join our team. If you’re passionate about data, modern platforms, and solving complex problems at scale, this opportunity is for you.What you’ll do:You’ll work with cutting-edge data platforms, design and maintain large-scale data pipelines, and collaborate...


  • Warszawa, mazowieckie, Polska Capgemini Polska Pełny etat

    Azure Databricks Data EngineerAt Capgemini Invent, we believe difference drives change. As inventive transformation consultants, we blend our strategic, creative and scientific capabilities, collaborating closely with clients to deliver cutting-edge solutions. Join us to drive transformation tailored to our client's challenges of today and tomorrow. Informed...


  • Warszawa, mazowieckie, Polska KMD Poland Pełny etat 25 zł - 200 zł

    Location: Warsaw (Inflancka 4A) or Remote Work (Poland) B2B Contract, Targeted Salary: 150 - 170 PLN Net/Hour #Python #ApacheSpark #Databricks #MSSQL #Git #CI/CD #Docker #Azure #Kubernetes Are you ready to join our international team as a Data Engineer with Databricks? We shall tell you why you should What products do we develop? KMD Elements is a...


  • Warszawa, mazowieckie, mazowieckie, Polska Capgemini Polska Pełny etat

    Azure Databricks Data EngineerMiejsce pracy: WarszawaTechnologies we useExpectedDatabricksDelta LakeSparkSQLPythonOptionalMLflowPower BIAbout the projectAt Capgemini Invent, we believe difference drives change. As inventive transformation consultants, we blend our strategic, creative and scientific capabilities, collaborating closely with clients to deliver...

  • Senior Data Engineer

    2 tygodni temu


    Warszawa, mazowieckie, mazowieckie, Polska CRESTT sp. z o.o. Pełny etat

    Senior Data Engineer (Databricks, Python)Miejsce pracy: WarszawaTechnologie, których używamyWymaganeDatabricksPySparkSQLGitAirflowdbtKafkaFlinkFabric Data FactoryPythonMicrosoft AzureO projekcieDołącz do zespołu Specjalistów, który tworzy rozwiązania w obszarze Data Lakehouse, Business Intelligence i Advanced Analytics w środowiskach chmurowych....


  • Warszawa, mazowieckie, Polska emagine Polska Pełny etat

    General Information:Industry: HealthcareRate: up to 150 PLN/hLocation: 100% remoteThis position seeks a skilled Data Engineer with a minimum of 3 years of experience specializing in Azure Data Services and Databricks. The ideal candidate will have a strong proficiency in SQL and Python, along with a solid understanding of data modeling and architectural...


  • Warszawa, mazowieckie, mazowieckie, Polska KMD Poland Sp. z o.o. Pełny etat

    Data Engineer with DatabricksMiejsce pracy: WarszawaTechnologies we useExpectedPythonApache SparkDatabricksGitAbout the project#Python #ApacheSpark #Databricks #MSSQL #Git #CI/CD #Docker #Azure #KubernetesAre you ready to join our international team as a Data Engineer with Databricks? We shall tell you why you shouldWhat products do we develop?KMD Elements...


  • Warszawa, mazowieckie, Polska KMD Poland Pełny etat 26 zł - 880 zł

    Location: Warsaw (Inflancka 4A) or Remote Work (Poland) B2B Contract, Targeted Salary: 160 - 180 PLN Net/Hour #Python #ApacheSpark #Databricks #MSSQL #Git #CI/CD #Docker #Azure #Kubernetes Are you ready to join our international team as a Senior Data Engineer with Databricks? We shall tell you why you should What products do we develop? KMD Elements is a...

  • Databricks Architect

    1 tydzień temu


    Warszawa, mazowieckie, Polska Remodevs Pełny etat 30 zł - 240 zł

    We are looking for an experienced Databricks Architect to join our team and lead the design and implementation of advanced data solutions.Your Responsibilities:Lead the design and implementation of scalable data architectures on Databricks.Define best practices for ETL/ELT pipelines, data modeling, and Medallion architectures (Bronze, Silver, Gold).Establish...