Key Responsibilities:
Data Pipeline Development & ETL
- Design, develop, and manage data pipelines on Azure Databricks for data import, pseudonymization, cleansing, and export to AWS S3 (Parquet format)
- Ensure workflows meet high standards for quality, reliability, and performance
Infrastructure Optimization
- Focus on stabilizing and optimizing the infrastructure, with an initial emphasis on improvements before transitioning to new feature development
Monitoring & Alerting
- Implement monitoring using Grafana, Prometheus, and Dynatrace, ensuring data quality and system reliability
Operational Support & Maintenance
- Rotate in monitoring MongoDB Atlas, Countly, and other data pipelines within the team
- Manage access rights, deployments, and address production issues, including regular updates and improvements
- Availability for on-call shifts outside regular office hours to independently resolve technical challenges (operational tasks are rotated among team members)
Required Skills & Qualifications
- Core Tech Stack: Databricks on Azure (Spark, Python, SQL) with solid hands-on experience with Azure Data Factory
- Database Knowledge: PostgreSQL and foundational MongoDB
- Infrastructure: Kubernetes, Docker, and Terraform, plus Git and Jenkins for CI/CD
- Monitoring Tools: Grafana, Prometheus, Dynatrace
- Additional Skills: Familiarity with AWS S3 for data export, flexible project handling, and strong team collaboration
If you’re a proactive Data Engineer with a strong background in Databricks on Azure, we’d love to meet you!
We offer
At Avangarde Software you will feel like home and will be surrounded by fun, caring and supportive colleagues. We encourage a healthy work-life balance adjustable to your preference!
Our promise! We will:
- Teach you
- Learn from you
- Find and boost your strong points
- Put you in the right context to succeed
- Grow together with you