We are looking for a Senior Data Engineer with strong expertise in Databricks to join our growing team. You will be responsible for designing and scaling the data infrastructure that powers Leakmited’s AI and data-driven products. This role is ideal for someone who thrives at the intersection of big data engineering, cloud platforms, and applied AI.
What you will do
Architect, build, and maintain scalable data pipelines and ETL processes using Databricks.
Optimize data workflows for performance, reliability, and cost-efficiency.
Collaborate closely with Data Scientists and ML Engineers to deploy and operationalize LLMs and other ML models into production systems.
Manage data quality, governance, and security across multiple sources and environments.
Support the integration of structured, semi-structured, and unstructured data into Leakmited’s platforms.
Contribute to the strategic evolution of Leakmited’s data architecture.
5+ years of experience in Data Engineering, Big Data, or related fields.
Hands-on expertise with Databricks (PySpark, Delta Lake, MLflow, SQL).
Solid experience with cloud platforms (Azure, AWS, or GCP).
Strong programming skills in Python and SQL.
Experience with workflow orchestration tools (Airflow, Prefect, dbt, etc.).
Knowledge of data modeling, storage optimization, and performance tuning.
A collaborative mindset and a passion for building reliable data systems that serve real-world impact.
Strong Bonus: exposure to ML/AI workflows, APIs, or real-time streaming pipelines.