Sutherland Logo

Sutherland

Databricks & GCP Data Platform Architect

Reposted 13 Hours Ago
Be an Early Applicant
In-Office or Remote
Hiring Remotely in A Tsin, Salavan
Senior level
In-Office or Remote
Hiring Remotely in A Tsin, Salavan
Senior level
Design and implement scalable Lakehouse solutions on GCP, build data pipelines, configure Databricks, and ensure data governance and compliance.
The summary above was generated by AI
Company Description

About Sutherland

Artificial Intelligence. Automation. Cloud engineering. Advanced analytics. For business leaders, these are key factors of success. For us, they’re our core expertise. We work with iconic brands worldwide. We bring them a unique value proposition through market-leading technology and business process excellence.

We’ve created over 200 unique inventions under several patents across AI and other critical technologies. Leveraging our advanced products and platforms, we drive digital transformation, optimize critical business operations, reinvent experiences, and pioneer new solutions, all provided through a seamless “as a service” model.

For each company, we provide new keys for their businesses, the people they work with, and the customers they serve. We tailor proven and rapid formulas, to fit their unique DNA. We bring together human expertise and artificial intelligence to develop digital chemistry. This unlocks new possibilities, transformative outcomes and enduring relationships.

Sutherland
Unlocking digital performance. Delivering measurable results.

 

Job Description

We are looking for a hands-on Databricks & GCP Data Platform Architect who will design and personally implement scalable Lakehouse solutions on Google Cloud Platform (GCP).

This role requires deep technical involvement, including building pipelines, configuring Databricks, and troubleshooting production issues, in addition to architecture ownership.

Key Responsibilities

1. Architecture & Hands-on Implementation

  • Design end-to-end Databricks Lakehouse architecture on GCP
  • Hands-on implementation of:
    • Databricks workspaces, clusters, jobs, and workflows
    • Delta Lake–based Bronze / Silver / Gold data layers
    • Batch and streaming pipelines using Spark and Databricks
  • Create reference implementations and reusable frameworks for teams
  • Actively participate in coding, reviews, and production deployments

2. Data Engineering (Hands-on)

  • Build and optimize Spark jobs and Databricks notebooks
  • Implement ingestion pipelines from:
    • Databases and enterprise applications
    • Streaming sources (Pub/Sub, Kafka)
    • External and SaaS systems
  • Perform performance tuning and cost optimization
  • Troubleshoot pipeline failures and production issues directly

3. Security, Governance & Compliance

  • Implement (not just define) governance using Unity Catalog
  • Configure access control integrated with GCP IAM
  • Set up secure networking (VPC, private endpoints)
  • Enable audit logging, lineage, and data classification
  • Work closely with security teams to operationalize standards

4. DevOps, Automation & Operations (Hands-on)

  • Build CI/CD pipelines for Databricks notebooks, jobs, and configs
  • Implement Infrastructure as Code using Terraform
  • Set up monitoring, alerting, and operational dashboards
  • Participate in production support, root-cause analysis, and fixes
  • Drive hands-on cost optimization initiatives

5. Stakeholder Collaboration

  • Translate business requirements into implemented solutions
  • Guide and mentor data engineers through code-level support
  • Conduct architecture and code reviews
  • Act as a technical owner from design through production

Required Skills & Experience

Must Have

  • Strong hands-on experience with Databricks (Apache Spark)
  • Proven experience building and deploying Lakehouse architectures
  • Hands-on experience with GCP, including:
    • Google Cloud Storage (GCS)
    • BigQuery
    • Pub/Sub
    • IAM & VPC basics
  • Experience implementing batch and streaming pipelines
  • Strong troubleshooting and production support skills

Good to Have

  • Unity Catalog, Delta Live Tables
  • CI/CD, Git, Terraform
  • MLflow, Vertex AI exposure
  • Multi-cloud Databricks experience (Azure / AWS)

Qualifications

  • 8–12 years of experience in data engineering / data platforms
  • 3+ years in a hands-on architect or senior technical lead role

Additional Information

All your information will be kept confidential according to EEO guidelines.

Similar Jobs

Yesterday
Easy Apply
Remote or Hybrid
Easy Apply
195K-278K Annually
Senior level
195K-278K Annually
Senior level
Artificial Intelligence • Cloud • Computer Vision • Hardware • Internet of Things • Software
The Core Account Executive will engage with strategic enterprise accounts, manage the sales process from prospecting to closing, and build executive relationships. Responsibilities also include demonstrating solution-based sales techniques and aligning with Samsara's cultural principles.
13 Hours Ago
In-Office or Remote
Senior level
Senior level
Artificial Intelligence • Analytics
The Disaster Recovery Engineer will implement and support DR solutions in data centers and multi-cloud environments, focusing on Linux systems, backup/recovery operations, and troubleshooting.
Top Skills: AWSAzureBackupCloudLinuxNasOciRecoveryReplicationSan
13 Hours Ago
In-Office or Remote
141K-175K Annually
Mid level
141K-175K Annually
Mid level
Information Technology • Consulting
The Account Executive will develop new business relationships, manage accounts, and promote ServiceNow solutions, focusing on the public sector in California.
Top Skills: Servicenow

What you need to know about the Colorado Tech Scene

With a business-friendly climate and research universities like CU Boulder and Colorado State, Colorado has made a name for itself as a startup ecosystem. The state boasts a skilled workforce and high quality of life thanks to its affordable housing, vibrant cultural scene and unparalleled opportunities for outdoor recreation. Colorado is also home to the National Renewable Energy Laboratory, helping cement its status as a hub for renewable energy innovation.

Key Facts About Colorado Tech

  • Number of Tech Workers: 260,000; 8.5% of overall workforce (2024 CompTIA survey)
  • Major Tech Employers: Lockheed Martin, Century Link, Comcast, BAE Systems, Level 3
  • Key Industries: Software, artificial intelligence, aerospace, e-commerce, fintech, healthtech
  • Funding Landscape: $4.9 billion in VC funding in 2024 (Pitchbook)
  • Notable Investors: Access Venture Partners, Ridgeline Ventures, Techstars, Blackhorn Ventures
  • Research Centers and Universities: Colorado School of Mines, University of Colorado Boulder, University of Denver, Colorado State University, Mesa Laboratory, Space Science Institute, National Center for Atmospheric Research, National Renewable Energy Laboratory, Gottlieb Institute

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account