Remote Opportunity – Data Architect (Databricks) – Certification Required

Role: Data Architect (Databricks)
Location: Remote
Experience: 15+ Years (with 4–5+ years in Databricks)
Certification: Databricks Certification (Mandatory)
Key Responsibilities:
  • Lead migration from on-prem/legacy platforms to Databricks Lakehouse.
  • Architect and optimize batch and streaming pipelines for AI/ML & GenAI.
  • Re-engineer Spark workloads with Delta Lake and Unity Catalog.
  • Integrate ML models into operational pipelines for real-time insights.
  • Design orchestration using Airflow or Databricks Workflows with CI/CD.
  • Implement governance, security, and data quality frameworks.
  • Collaborate with data scientists, ML engineers, and business teams.
Required Skills:
  • 15+ years in data engineering/architecture, with 4–5+ years in Databricks.
  • Hands-on with Databricks, Spark, Delta Lake, Unity Catalog.
  • Strong knowledge of Airflow, Databricks Jobs, Kafka, streaming frameworks.
  • Proven experience with AI/ML pipeline integration (MLflow or equivalent).
  • Excellent communication and stakeholder engagement.
Preferred Skills:
  • Experience with CI/CD, Terraform, DevOps practices.
  • Exposure to MLOps, Feature Store, GenAI/LLM integration.





If you are interested, please reply with your updated resume.

Regards,
Raiyyan
raiyyank061@gmail.com

:

:
:
:
    
🔔 Get our daily C2C jobs / Hotlist notifications on 

WHATSAPP              TELEGRAM                  LINKEDIN
   

Leave a Reply

Your email address will not be published. Required fields are marked *