About the job
We are seeking a skilled Data Engineer to contribute to a major enterprise real-time data transformation program in Brussels. The project centers on event-driven architecture and will give you the opportunity to work closely with distributed engineering teams across Europe on a mission-critical cloud initiative.
Key Responsibilities
- Architect and implement real-time data pipelines leveraging Azure services (Azure Data Factory, Azure Synapse, Azure Event Hubs, Azure Stream Analytics, Cosmos DB).
- Design and optimize scalable infrastructure for caching, integration, and API-driven workloads.
- Build and manage CI/CD pipelines in Azure DevOps, with Infrastructure-as-Code delivered via Terraform or Bicep.
- Develop and maintain services in Python (FastAPI) for API and microservice integration.
- Implement monitoring and observability with Azure Monitor, Application Insights, and Log Analytics.
- Perform system performance testing and produce technical documentation.
- Collaborate with architects, developers, and operations teams across Europe.
Core Skills & Experience
- Proven hands-on experience with Azure cloud services, especially Data Factory, Synapse, Event Hubs, and Stream Analytics (Cosmos DB a plus).
- Strong knowledge of event-driven architectures and low-latency data processing.
- Proficiency in Python, with experience developing APIs using FastAPI.
- Practical experience with Azure DevOps CI/CD and Terraform or Bicep.
- Knowledge of monitoring and observability in the Azure ecosystem.
- Strong collaboration skills and ability to document and present technical solutions.
Solliciteren