
Principal Data Engineer
- Singapore
- Permanent
- Full-time
70% compound annual growth in Revenue over the last 5 years. Sleek has been recognised by The Financial Times, The Straits Times, Forbes and LinkedIn as one of the fastest growing companies in Asia.Backed by world-class investors, we are on track to be one of the few cash flow positive, tech-enabled unicorns based out of Singapore.Requirements:We are looking for a Principal Data Engineer that is excited about the below Mission and Outcomes over the next 6-12 months.Mission:Work closely with cross-functional teams to translate our business vision into impactful data solutions. Drive the alignment of data architecture requirements with strategic goals, ensuring each solution not only meets analytical needs but also advances our core objectives. Be pivotal in bridging the gap between business insights and technical execution by tackling complex challenges in data integration, modeling, and security, and by setting the stage for exceptional data performance and insights. Shape the data roadmap, influence design decisions, and empower our team to deliver innovative, scalable, high-quality data solutions every day.Outcomes:1. Architecture & Design:
- Define the overall greenfield data architecture (batch + streaming) using GCP - BigQuery..
- Establish best practices for ingestion, transformation, data quality, and governance.
- Lead the design and implementation of ETL/ELT pipelines:
- Ingestion: Datastream, Pub/Sub, Dataflow, Airbyte, Fivetran, Rivery
- Storage & Compute: BigQuery, GCS
- Transformations: dbt, Cloud Composer (Airflow), Dagster
- Ensure data quality and reliability with dbt tests, Great Expectations/Soda, and monitoring.
- Implement Dataplex & Data Catalog for metadata, lineage, and discoverability.
- Define IAM policies, row/column-level security, DLP strategies, and compliance controls.
- Define and enforce SLAs, SLOs, and SLIs for pipelines and data products.
- Implement observability tooling:
- Cloud-native: Cloud Monitoring, Logging, Error Reporting, Cloud Trace.
- Third-party (nice-to-have): Monte Carlo, Datafold, Databand, Bigeye.
- Build alerting and incident response playbooks for data failures and anomalies.
- Ensure pipeline resilience (idempotency, retries, backfills, incremental loads).
- Establish disaster recovery and high availability strategies (multi-region storage, backup/restore policies).
- Partner with BI/analytics teams to deliver governed self-service through Looker, Looker Studio, and other tools.
- Support squad-level data product ownership with clear contracts and SLAs.
- Mentor a small data engineering team; set coding, CI/CD, and operational standards.
- Collaborate with squads, product managers, and leadership to deliver trusted data.
- 10+ years experience in data engineering, architecture, or platform roles.
- Strong expertise in GCP data stack: BigQuery, GCS, Dataplex, Data Catalog, Pub/Sub, Dataflow.
- Hands-on experience building ETL/ELT pipelines with dbt + orchestration (Composer/Airflow/Dagster).
- Deep knowledge of data modeling, warehousing, partitioning/clustering strategies.
- Experience with monitoring, reliability engineering, and observability for data systems.
- Familiarity with data governance, lineage, and security policies (IAM, DLP, encryption).
- Strong SQL skills and solid knowledge of Python for data engineering.
- Experience with Snowflake, Databricks, AWS (Redshift, Glue, Athena), or Azure Synapse.
- Knowledge of open-source catalogs (DataHub, Amundsen, OpenMetadata).
- Background in streaming systems (Kafka, Kinesis, Flink, Beam).
- Exposure to data observability tools (Monte Carlo, Bigeye, Datafold, Databand).
- Prior work with Looker, Hex, or other BI/analytics tools.
- Startup or scale-up experience (fast-moving, resource-constrained environments).