Firmenlogo

Senior Data Engineer at Avacone

Avacone · Basel, · Hybrid

Apply Now

Description

The Opportunity

We are supporting a major data platform transformation within a banking environment, moving from a legacy SQL Server and SSIS-based setup to a modern, scalable architecture built on dbt, Dagster, and OpenShift.

This role is not about maintaining existing systems. It is about rebuilding a critical data platform from the ground up, with direct impact on risk, trading PnL, and core financial data flows.

We are looking for a hands-on Senior Data Engineer who can take ownership of complex migration workstreams and deliver reliably in a regulated, high-stakes environment.  

What You Will Do

You will play a central role in the end-to-end migration and modernisation of the data platform.

Platform Transformation

  • Translate legacy ETL logic from SSIS and stored procedures into modern ELT pipelines using dbt
  • Implement Data Vault 2.0 structures including Raw Vault and Business Vault
  • Build datamarts and curated datasets for downstream analytics and reporting


Orchestration & Infrastructure

  • Design and operate workflows using Dagster, including scheduling, dependencies, and recovery mechanisms
  • Deploy and run data workloads on OpenShift / Kubernetes environments


Event-Driven Data Processing

  • Enable near real-time data processing using Kafka-triggered pipelines
  • Integrate with upstream data lake environments and external data providers


Data Quality & Validation

  • Establish robust data validation and reconciliation processes
  • Implement automated testing and monitoring using dbt


Operational Ownership

  • Support production pipelines and resolve incidents when required
  • Create clear documentation and ensure operational readiness
  • Continuously improve performance, reliability, and maintainability

What You Will Do

You will play a central role in the end-to-end migration and modernisation of the data platform.


Platform Transformation

  • Translate legacy ETL logic from SSIS and stored procedures into modern ELT pipelines using dbt
  • Implement Data Vault 2.0 structures including Raw Vault and Business Vault
  • Build datamarts and curated datasets for downstream analytics and reporting


Orchestration & Infrastructure

  • Design and operate workflows using Dagster, including scheduling, dependencies, and recovery mechanisms
  • Deploy and run data workloads on OpenShift / Kubernetes environments


Event-Driven Data Processing

  • Enable near real-time data processing using Kafka-triggered pipelines
  • Integrate with upstream data lake environments and external data providers


Data Quality & Validation

  • Establish robust data validation and reconciliation processes
  • Implement automated testing and monitoring using dbt


Operational Ownership

  • Support production pipelines and resolve incidents when required
  • Create clear documentation and ensure operational readiness
  • Continuously improve performance, reliability, and maintainability

Requirements

What You Bring

Technical Expertise

  • Strong experience with SQL Server and T-SQL, including performance optimisation
  • Proven hands-on experience with dbt in production environments
  • Solid experience with workflow orchestration tools, ideally Dagster
  • Practical knowledge of Data Vault 2.0 modelling concepts
  • Experience working with container platforms such as OpenShift or Kubernetes
  • Familiarity with event-driven architectures and Kafka


Domain Experience

  • Experience working with financial data, ideally in banking or trading environments
  • Understanding of risk and PnL data structures is a strong advantage


Working Style

  • Strong ownership mindset with the ability to work independently
  • Structured, pragmatic, and delivery-focused
  • Comfortable operating in complex and regulated environments
  • Clear communicator across both technical and business stakeholders

What Success Looks Like

Within the first months, you will have:

  • Delivered initial Data Vault structures and migrated datasets into the new platform
  • Established stable, event-driven pipelines
  • Ensured data consistency and validation between legacy and new systems
  • Contributed to a production-ready, scalable data platform
Apply Now

Other home office and work from home jobs