Agoda - Senior Data Platform Engineer — StarRocks
Upload My Resume
Drop here or click to browse · Tap to choose · PDF, DOCX, DOC, RTF, TXT
Requirements
• 5+ years of experience in data engineering / database engineering / DBA / platform engineering, with at least several years owning production OLAP/BI databases at scale. • Strong expertise with StarRocks (or similar modern MPP / columnar systems such as Vertica, ClickHouse, Presto/Trino, Druid, Doris, Impala) and solid SQL Server experience, including: • Query execution internals and tuning (plans, statistics, partitioning, indexing, materialized views). • Capacity planning, storage layout, and backup/restore. • Proven experience operating mission‑critical database or BI services in production: • On‑call incident response, debugging across layers, and post‑mortem/root‑cause analysis. • Designing and maintaining SLIs/SLOs, dashboards, and alerting for latency/availability/error budgets. • Solid understanding of data warehouse and ETL concepts (core DW tables like Fact-Dimension, MSSQL Replication/CDC, Staging patterns, data freshness & consistency, DQ processes). • Hands‑on experience with Linux and Windows in production environments (patching, hardening, troubleshooting resource issues). • Experience with infrastructure automation and CI/CD: • Pipelines and deployment automation (GitLab CI, Ansible/AWX, ArgoCD, or similar). • Git‑based workflows for database objects. • Strong communication and collaboration skills: • Able to represent BI‑Infra DBA as the technical point of contact in cross‑team design and incident discussions at IC4 level. • Comfortable explaining complex trade‑offs to both technical and non‑technical stakeholders. • Experience with StarRocks or similar modern OLAP engines, including: • Designing schemas and materialized views for BI workloads. • Operating clusters across multiple environments (DR, performance/capacity testing, cost tuning). • Hands‑on work with Superset, Metabase, Tableau, Power BI Report Server as data consumers — understanding how visualization patterns translate into backend workload and SLOs. • Exposure to Hadoop / Data Lakehouse ecosystems (e.g., Impala, Spark, Hive) and strategies for moving workloads from DW to lake/next‑gen engines. • Experience with observability stacks (Prometheus, Grafana, Hedwig/NOC dashboards and alerting), SRE practices, and especially using AI/LLM-based tools to help with log analysis, anomaly detection, or auto-remediation for database/BI engines — this will be a strong plus. • Prior technical leadership at IC4 level or equivalent, driving multi‑quarter cross‑team projects and raising engineering standards for reliability, performance, and maintainability. • Please review our Hiring Process Guidelines before your interview — click here to learn how interviewing at Agoda works.
Responsibilities
• Design and evolve Agoda’s BI data warehouse architecture across SQL Server DW, Vertica, StarRocks, and related services, focusing on availability, scalability, and cost efficiency. • Lead migration and modernization projects such as: • Migrating workloads off MSSQL DW to more sustainable platforms (Production standby, data lake, StarRocks, etc.). • Decommissioning legacy components as part of the long‑term DW strategy. • Define and implement data and query engine standards (indexing/partitioning patterns, materialized views, retention strategies, governance and access patterns) for BI platforms. • Define SLIs/SLOs for key services (StarRocks, Vertica, Tableau, Superset/Metabase, MSSQL DW) and ensure they are instrumented, monitored, and reviewed regularly. • Lead incident response for BI‑Infra DBA‑owned platforms: diagnose cross‑system issues (Query Engine - Database, VIz tools, Compute, Storage, Network), post‑mortem/root‑cause analysis, and follow‑up improvements. • Partner with BI‑Infra Ops/Dev/Viz and DBDev/DBOps on end‑to‑end reliability — from ingestion pipelines to visualization queries and user traffic patterns. • Perform deep query and workload tuning for high‑impact systems (e.g., finance, supply, whitelabel analytics, experimentation, DQ), reducing CPU/memory/IO and improving latency at scale. • Optimize resource consumption and license/capacity planning for Query Engine – Database and BI tools. • Collaborate with infra and product teams to evaluate and adopt new query engines / database technologies where they materially improve reliability, performance, or cost. • Own upgrade and patching strategies for database and BI stacks (OS, engine, drivers), using canary/blue‑green and rollback plans to minimize downtime. • Improve and extend CI/CD and deployment automation for database objects, jobs, and configuration (GitLab, AWX/Ansible, ArgoCD, etc.). • Build and maintain observability for BI-Infra DBA services (Grafana, Prometheus/SLO dashboards, alerting, Hedwig/NOC integrations) to detect and act on issues proactively. • Implement and maintain secure access patterns for DW and BI platforms (service accounts, RBAC, gMSA, password rotation, SSO/OKTA integration, DR strategy). • Partner with security and data governance teams to remediate vulnerabilities within SLAs, enforce data residency constraints, and ensure that upcoming workloads (e.g., whitelabels) are compliant by design. • Act as technical lead for BI‑Infra DBA initiatives, collaborating closely with BI‑Infra Ops/Dev/Viz, Database Operation/Development, Infra, and BI product teams. • Mentor and coach junior engineers on database performance, reliability, and platform engineering best practices; help raise the bar for design reviews, runbooks, and incident hygiene. • Contribute to knowledge sharing - Documents [Confluence, Agoda Internal Document Framework], internal talks, runbooks and help shape the technical roadmap for BI‑Infra DBA.
Similar Jobs
No credit card. Takes 10 seconds.