Hybrid Data Engineer (Databricks) chez Coherent solutions
Coherent solutions · Americas, États-Unis d'Amérique · Hybrid
- Bureau à Americas
Project Description
We are looking for an experienced Data Engineer with deep expertise in Databricks to join our advanced analytics and data engineering team. The ideal candidate will play a key role in designing, building, and optimizing large-scale data solutions on the Databricks platform, supporting business intelligence, advanced analytics, and machine learning initiatives. You will collaborate with cross-functional teams to deliver robust, scalable, and high-performance data pipelines and architectures.
Technologies
- Databricks (including Spark, Delta Lake, MLflow)
- Python/Scala
- SQL
- ETL concepts
- Distributed data processing
- Data warehousing
- Cloud Platforms & Storage
What You'll Do
- Lead the design, development, and deployment of scalable data pipelines and ETL processes using Databricks (Spark, Delta Lake, MLflow);
- Architect and implement data lakehouse solutions, ensuring data quality, governance, and security;
- Optimize data workflows for performance and cost efficiency on Databricks and cloud platforms (Azure, AWS, or GCP);
- Collaborate with data scientists, analysts, and business stakeholders to understand data requirements and deliver actionable insights;
- Mentor and guide junior engineers, promoting best practices in data engineering and Databricks usage;
- Develop and maintain documentation, data models, and technical standards;
- Monitor, troubleshoot, and resolve issues in production data pipelines and environments;
- Stay current with emerging trends and technologies in data engineering and Databricks ecosystem;
Job Requirements
- Bachelor’s or Master’s degree in Computer Science, Engineering, Mathematics, or a related field;
- 5+ years of experience in data engineering, with at least 2 years of hands-on experience with Databricks (including Spark, Delta Lake, and MLflow);
- Strong proficiency in Python and/or Scala for data processing;
- Deep understanding of distributed data processing, data warehousing, and ETL concepts;
- Experience with cloud data platforms (Azure Data Lake, AWS S3, or Google Cloud Storage);
- Solid knowledge of SQL and experience with large-scale relational and NoSQL databases;
- Familiarity with CI/CD, DevOps, and infrastructure-as-code practices for data engineering;
- Experience with data governance, security, and compliance in cloud environments;
- Excellent problem-solving, communication, and leadership skills;
- English: Upper Intermediate level or higher;
What Do We Offer
The global benefits package includes:
- Technical and non-technical training for professional and personal growth;
- Internal conferences and meetups to learn from industry experts;
- Support and mentorship from an experienced employee to help you professional grow and development;
- Internal startup incubator;
- Health insurance;
- English courses;
- Sports activities to promote a healthy lifestyle;
- Flexible work options, including remote and hybrid opportunities;
- Referral program for bringing in new talent;
- Work anniversary program and additional vacation days.