Berkley

Databricks Data Engineer

Location Name VA, Manassas
ID
2026-13782
Date Posted
15 hours ago(3/25/2026 9:15 AM)
Company
Berkley Alternative Markets Tech
Primary Location
US-VA-Manassas
Loc2
MO, Chesterfield - 14755 North Outer Forty
Category
Information Technology

Company Details

Berkley_AltMarketsTech_Logo_Color

 

We started in early 2019 as a small group of technologists with a passion for making insurance better. Today we are working with a team of industry experts who run five different insurance brands and collectively control $1 billion in annual premiums.

 

We believe in an idea and execution meritocracy. In other words, a place where the best ideas win and the people who deliver the most value get the most opportunities.

 

As we grow our team, we are looking for inquisitive, entrepreneurial people who are excited to reimagine the insurance industry.

 

Insurance is too complex. Help us make it better.

Responsibilities

This position requires on-site work Monday–Thursday at either our Manassas, VA or Chesterfield, MO location. 

 

 

The Databricks Data Engineer will help design, build, deploy, and maintain scalable and production grade data pipelines in modern cloud environments, enabling analytics, AI, ML, and decision advantage at scale.  This role will work with cutting-edge tools like Databricks, Delta Lake, PySpark, and AI/BI genie to transform raw data into actionable insights.  As a hands-on Databricks Data Engineer with deep expertise in Azure Databricks and MLOps, this role will have the opportunity to migrate and translate legacy SSIS ETL logic into scalable, cloud-native data pipelines in Databricks. This role will partner with data engineers, data scientists, and product manager to design features, train/evaluate models, and deploy them to production using MLflow, Databricks and Workflows—with rigorous observability, governance (Unity Catalog), and CI/CD automation.

 

Data Pipeline Engineering

 

  • Design, build, and maintain high-performance, scalable ETL/ELT pipelines using Azure Databricks, Delta Lake, and PySpark.
  • Convert and modernize existing SSIS package logic into cloud-native Databricks pipelines using PySpark notebooks, Delta Live Tables (DLT), and Databricks Workflows.
  • Implement reliable batch and streaming pipelines with robust data quality and validation frameworks.
  • Optimize pipeline performance using Photon, efficient file formats, partitioning, Z-ordering, and caching strategies.

 

Lakehouse Platform Development

 

  • Develop and manage datasets within Delta Lake, ensuring ACID reliability, schema evolution, versioning, and time travel.
  • Architect feature-rich data layers including:
    • Bronze (raw ingestion)
    • Silver (validated, conformed)
    • Gold (analytics-ready and ML-ready)
  • Implement data governance using Unity Catalog for fine-grained access control, lineage, auditability, and metadata management.

 

 

MLOps & ML-Enabled Data Pipelines

 

  • Partner with data scientists and data engineers to create feature pipelines, model training pipelines, and production scoring pipelines.
  • Deploy and operationalize models using MLflow, Databricks Model Registry, and Databricks Workflows.
  • Use Databricks built-in AI SQL functions such as ai_query, ai_forecast, ai_analyze_sentiment to generate actionable insight from large amount of unstructured or structured raw data
  • Implement monitoring for:
    • Pipeline failures
    • Data/feature drift
    • Model performance degradation
    • Operational SLAs/SLIs/SLOs
  • Build automated CI/CD workflows using GitHub Actions or Azure DevOps for notebook deployment, pipeline testing, and environment promotion.

 

Data Platform, Data Security & Data Governance

 

  • Collaborate with data engineers to design reliable data products on Delta Lake; leverage Delta Live Tables (DLT) for declarative pipelines when applicable.
  • Enforce Unity Catalog for lineage, permissions, and audit; manage secrets, tokens, and keys securely (e.g., Databricks secrets, Key Vault/Secrets Manager).

 

Collaboration & Leadership

 

  • Work closely with cross-functional teams: data engineering, data scientist, product manager, and business stakeholders.
  • Serve as a Databricks SME—championing best practices, code standards, governance, and reusable frameworks.
  • Document architecture, workflows, data models, runbooks, and operational procedures.

 

Qualifications

  • Minimum of 3 years of experience in Databricks, PySpark notebooks, Python, DevOps, software development, and data engineering.
  • Certified Databricks Data Engineer Associate or Professional is a plus.

 

Skills & Competencies

  • Proficient in designing, building, deploying, and maintaining high-performance, scalable ETL/ELT pipelines using Azure Databricks, Delta Lake, and PySpark Notebook.
  • Proficient in building, deploying, and operating production ML models such as supervised, unsupervised, and anomaly detection, including techniques for imbalanced datasets
  • Proficient with ML engineering and MLOps, including model versioning, CI/CD for ML, monitoring, drift detection, and automated retraining
  • Proficiency in Python including Pandas and PySpark Dataframes
  • Expert level of SQL skills including Stored Procedure, experience with SSIS, SSRS, Power BI is a plus.
  • Proficient with cloud data engineering platforms, such as Azure, Databricks, Spark, or SQL, and batch and streaming pipelines
  • Familiar with Databricks AI Built-In Functions such as AI_Query, AI_Gen, AI_Classify, AI_Forecast, AI_Analyze_Sentiment, able to use them to extract actionable insights from large amount of unstructured or structured raw data
  • Experience with Python and ML frameworks, such as PyTorch or TensorFlow
  • Experience improving data quality, lineage, and observability in enterprise data environments and operationalizing rules and model-driven scoring for prioritization, routing, or case selection
  • Experience with predictive analytics, machine learning and artificial intelligence desired.

Education

  • A Bachelor’s degree in Computer Science, Management Information Systems, Engineering, Math, Physics, or a related quantitative field is required (4-year degree). Master’s degree preferred
  • Experience in the commercial insurance industry is a plus.

 

Additional Company Details

The Company is an equal employment opportunity employer.

We do not accept any unsolicited resumes from external recruiting firms.

The company offers a competitive compensation plan and robust benefits package for full time regular employees.
Base salary & Benefits include Health, dental, vision, life, disability, wellness, paid time off, 401(k) and profit-sharing plans.

The actual salary for this position will be determined by a number of factors, including the scope, complexity and location of the role; the skills, education, training, credentials and experience of the candidate; and other conditions of employment.

Additional Requirements

• Ability to travel locally and nationally up to 5% of the time

Sponsorship Details

Sponsorship not Offered for this Role

Options

Sorry the Share function is not working properly at this moment. Please refresh the page and try again later.
Share on your newsfeed

Connect With Us!

Not ready to apply? Connect with us for general consideration.