- Start: ASAP
Hey there!
We’re looking for Databricks Architects for our US-based client. The work focuses on areas like migration, data ingestion, and optimization of Databricks-based solutions. The client has an ongoing need for specialists - their projects tend to be short-term (with a high chance of extension), and since the demand is steady, they’re usually able to offer new assignments after each one wraps up.
Right now, they’re looking for experts to join an AI/ML healthcare project. The platform is analyzing both text data and images generated by medical devices (X-rays, MRI, etc.). The gathered data will be migrated to a Databricks-powered cloud database. The platform will manage the full data lifecycle and include built-in features for compliance, auditing, cohort creation, and model reuse. The ultimate goal is to address issues in current data systems (e.g., scattered sources, manual processes, insufficient security).
We're looking for engineers fluent in Python, with strong cloud experience and solid knowledge of Databricks and Apache Spark. The projects are mostly for US-based companies - in most cases, working hours require only a slight overlap (like 10:00–18:00 CET), but we’re flexible on the schedule.
Scope:
Ensuring secure data storage
Processing and indexing DICOM data
Validating data, building processing pipelines, creating and sharing cohorts
Planning and executing database migrations
Collaborating closely with a cross-functional team (data engineers, data scientists, clinical informaticians, support teams, etc.)
- ️ 8+ years of experience in data engineering or related roles
- ️ Strong hands-on experience with Databricks and Apache Spark
- ️ Excellent knowledge of Python
- ️ Experience with cloud migrations
- ️ Hands-on experience with AWS (especially Amazon S3)
- ️ Previous work on AI/ML projects
- ️ Strong communication & teamwork skills
- ️ A proactive mindset and independence
- ️ English good enough for smooth daily communication
- ️ Experience with Azure (Data Factory, Synapse, Logic Apps, Data Lake)
- ️ Skills in designing/optimizing data pipelines using DBT, SSIS, TimeXtender or similar (ETL/ELT tools)
- ️ Experience with big data or NoSQL platforms (e.g. Redshift, Hadoop, EMR, Google Data)
What do we offer and how do we work?
Open communication throughout the recruitment process and beyond – we care about transparency
We keep our hiring process simple and candidate-friendly
We work remote-first – travel is kept to a minimum
We offer private medical care (Medicover) and Multisport card for contractors
How to apply?
Just send us your application via the form!