Job Description
MAIN RESPONSABILITIES
• Incident Response
o Understand problems from a user perspective and communicate to clearly
understand the issue.
o Reproduce bugs or issues that users are facing.
o Apply root cause analysis to quickly and efficiently
o Find the root cause of the problem, patch it, test it, and communicate with the
end user.
o Write postmortems summarizing every step of resolution and helping the team
to track all issues.
o Monitor existing flows and infrastructure and perform the same tasks when
discovering bugs/issues through monitoring and alerting.
• Maintenance
o Monitor flows and infrastructure to identify potential issues.
o Adapt configurations to keep flows and infrastructures working as expected,
keeping the operations without incident.
• Database Optimization
o Track costs and time of processing through dedicated dashboards.
o Alert people who query tables the wrong way, involving high costs.
o Track down jobs, views, and tables that are running inefficiently and occur
either high costs or low speed of execution.
o Optimize jobs, queries, and tables to optimize both costs and speed of
execution.
• Infrastructure Management
o Manage infrastructure through Terraform.
o Share and propose good practices.
o Decommission useless infrastructures such as services, tables, or virtual
machines.
• Deployments
o Track future deployments with a Data Architect and participate in Deployment
Reviews.
o Share and propose good practices of deployment.
o Accompany Data Engineers during the entire process of deployments.
o Accompany Data Engineers in the following period of active monitoring.
o Ensure diligent application of deployment process, logging, and monitoring
strategy.
o Take over newly deployed flows in the run process.
REQUESTED HARD SKILLS
• Google Cloud Platform: General knowledge of the platform and various services,
and at least one year of experience with GCP.
• Apache Airflow: At least two years of experience with the Airflow orchestrator,
experience with Google Composer is a plus.
• Google BigQuery: Extensive experience (at least 4 years) with GBQ, know how to
optimize tables and queries, and able to design database architecture.
• Terraform: At least two years of experience with Terraform, and know good
practices of GitOps.
• Apache Spark: this is an optional expertise we would value. Some of our pipelines
use pySpark.
• Additional Knowledge and Experience that are a Plus:
o Pub/Sub
o Kafka
o Azure Analysis Services
o Google Cloud Storage optimizatio
How the process will look like
Your teammates will gather all requirements within our organization. Then, once priority has been discussed, you will decide as a team on the best solutions and architecture to meet these needs. In continuous increments and continuous communication between the team and stakeholders, you’re part of making data play an even more important (and understood) part withing Brand New Day.
EUR 38K - 71K *
Thank you for submitting your application. We will contact you shortly!