Skip to main content

Job Description

   Back

Architect - Databricks

10-03-2026 17:41:20

Job_303330

10 - 12 years

  • Chennai, Tamil Nadu, India (CHN)

We are looking for an experienced Data Solution Architect to design, architect, and implement large-scale data platforms and analytics solutions. The ideal candidate will have strong hands-on experience in Databricks, Azure, and Google Cloud Platform (GCP), along with the ability to create animated/visual data stories for technical and business audiences.

Candidate will play a key role in defining data strategy, optimizing data pipelines, ensuring platform scalability, and developing reusable frameworks. ⸻

Required Skills & Experience

• 10–12+ years overall in data engineering, analytics, or architecture roles.

• 2–4+ years strong hands-on experience in Databricks (mandatory).

• Deep experience with Azure and GCP cloud ecosystems.

• Strong expertise in Spark, PySpark, SQL, Python.

• Proficiency in data modeling, ETL frameworks, and distributed data systems.

• Experience with streaming technologies (Kafka, Event Hub, Pub/Sub).

• Ability to design and present animated architecture and data flows.

• Strong communication and stakeholder management skills.


Key Responsibilities

1. Architecture & Solution Design

• Design end-to-end data lakehouse, data warehouse, and analytics architectures using Databricks, Azure, and GCP.

• Build scalable and cost-optimized architectures for ingestion, transformation, streaming, and ML workloads.

• Define and enforce architecture standards, patterns, and governance models.

• Translate business requirements into logical and physical data models.


2. Databricks Expertise

• Architect and optimize Databricks Lakehouse solutions using Delta Lake, Unity Catalog, Databricks SQL, MLflow, Auto Loader, and DLT.

• Implement advanced transformations using PySpark, Spark SQL, and notebooks.

• Design job orchestration using Databricks Workflows or other orchestration tools.


3. Cloud Platform Ownership

Azure

• Azure Data Factory, Azure Data Lake Storage (ADLS), Azure Synapse, Azure Functions, Event Hub, Azure DevOps, Azure Kubernetes Service (AKS).

• Architecture of secure and scalable data systems on Azure.

GCP

• BigQuery, Cloud Storage, Dataflow, Pub/Sub, Dataproc, Cloud Composer.

• Hands-on experience designing lakehouse/analytics solutions on GCP.


4. Data Engineering & Integration

• Lead development of batch and real-time data pipelines using Spark, ADF, Dataflow, or Databricks Workflows.

• Implement ETL/ELT frameworks, CI/CD, and reusable components.


6. Security & Governance

• Implement data governance frameworks including Unity Catalog, encryption, and compliance.

• Ensure best practices for cost optimization, monitoring, and performance tuning.