JR-124399 Lead Databricks Engineer
We are seeking a Lead Databricks Engineer to drive the design, development, and optimization of data pipelines and analytics solutions on the Databricks Lakehouse platform. This role is ideal for a hands-on technical leader who is passionate about big data technologies, cloud computing and enabling business insights through scalable data architectures.
Locations:
• Serbia
• Albania
• Bosnia and Herzegovina
• Montenegro
• North Macedonia
• Ukraine
• Georgia
• Argentina
• Brazil
• Estonia
• Latvia
• Lithuania
• Finland
• Romania
• Hungary
• Slovakia
• Slovenia
• Czech Republic
Requirements:
• Bachelors or Master's degree in Computer Science, Information Technology, Engineering or related field;
• 5+ years of experience in Data Engineering or Big Data roles;
• 2+ years of hands-on experience with Databricks, Spark (PySpark or Scala) and Delta Lake;
• Strong knowledge of cloud platforms (AWS, Azure or GCP) and modern data architectures (Lakehouse, Data Mesh, etc.);
• Proficiency in SQL, Python and distributed data processing;
• Experience with CI/CD, version control (Git) and DataOps in a data environment;
• Deep understanding of data governance, cataloging and security concepts;
• Experience leading a team or a project, acting as a team/tech lead of a project.
Nice to Have:
• Databricks (Pro, Architect) certification - Experience with machine learning pipelines and MLOps in Databricks;
• Exposure to streaming technologies (Kafka, Spark Structured Streaming);
• Knowledge of DBT, Airflow or other similar transformation and orchestration tools.
Other skills:
• English excellent written and verbal communication skills;
• Ability to work in a global multi-cultural and multi-national company;
• Ability to lead conversations with both technical and business representatives;
• Proven ability to work both independently and as a part of an international project team.
Job Responsibilities:
• Lead end-to-end development of data pipelines, ETL/ELT processes and batch/streaming solutions using Databricks and Apache Spark;
• Design and implement Lakehouse architectures that align with business and technical requirements;
• Collaborate with data scientists, analysts and engineers to deliver high-performance data products and ML features;
• Define and enforce coding standards, best practices and performance tuning strategies across Databricks notebooks and jobs;
• Optimize data models in Delta Lake and implement data governance standards using Unity Catalog;
• Manage integration of data sources across cloud platforms (e.g. AWS, Azure, GCP) using native and third-party connectors;
• Contribute to and lead technical reviews, architecture sessions and mentoring of less experienced engineers
• Automate infrastructure deployment with tools like Terraform, Databricks CLI or others;
• Ensure data platform solutions are secure, compliant, and scalable across global business units.
What We Offer:
• Competitive salary;
• 100% remote opportunity;
• Opportunities for professional growth and advancement;
• A collaborative and innovative work environment;
• 20 days of paid vacation, 15 paid days of sick leave with a doctors note, and 5 days of paid sick leave without a doctors note;
• Medical insurance coverage for employees, with optional family coverage at corporate rates;
• Support for participation in professional development opportunities (webinars, conferences, trainings, etc.);
• Regular team-building activities and bi-annual company-wide events;
• Flexible work environment (in-office, remote, or hybrid depending on preferences and manager approval).
Job ID: JR -124399
Apply tot his job
Apply To this Job