Data & Analytics
Sr Databricks Data Engineer(Strong Testing Experience)
Quadrant Technologies
Full-Time
Senior
$125k – $146k/yr
Remote
Posted 2d ago
Tech Stack
PythonGoAWSSparkRedshift
Job Description
**Role Summary:**
We are seeking a Senior Databricks Data Engineer (QA Testing) with strong hands on experience building and optimizing data pipelines on Databricks. This role focuses on implementation, performance, and reliability while working within a well defined architectural framework.
**Key Responsibilities:**
* Develop and maintain data pipelines on Databricks using medallion architecture principles
* Build, optimize, and maintain Databricks notebooks for ingestion, transformation, and analytics
* Implement data flows across boundary systems ensuring reliability, scalability, and data quality
* Apply performance tuning techniques including query optimization, cluster sizing, and Delta Lake optimization
* Work with Unity Catalog to manage schemas, permissions, and data access controls
* Collaborate closely with architects to implement approved data designs and standards
* Troubleshoot pipeline failures, data issues, and performance bottlenecks
* Ensure code quality, documentation, and operational readiness of data solutions
**Required Qualifications:**
* Strong hands on experience with Databricks and Delta Lake
* Strong hands\-on experience with AWS services such as S3, EC2, Lambda, Glue, and Redshift.
* Working experience with IBM DataStage for ETL development, data integration, and data transformation.
* Proven ability to design metadata\-driven test harnesses that scale — parameterized test configs that can run against 500 jobs today and 80K tomorrow without rearchitecting
* Experience with data validation tooling (Great Expectations, Deequ, or custom PySpark validation libraries) for row count, schema, transformation logic, and output comparison checks
* Performance and throughput benchmarking — rows/sec, latency comparisons between legacy (DataStage) and target (Databricks) pipelines
* Experience with Iceberg Tables
* Proven experience implementing medallion architecture in real world use cases
* Experience building notebooks using Python, SQL, or Scala
* Solid understanding of data integration across multiple systems
* Strong performance tuning and debugging skills in Databricks environments
* Experience working in enterprise scale data platforms
**Scale \& Automation Mindset**
* Experience designing test coverage strategies for large\-scale migrations (thousands of jobs), including prioritization by complexity tier (Low/Medium/High) and risk scoring
* Ability to build reusable test templates that align with ingestion patterns — Full Load, Incremental/CDC, File Ingest, API — so each converted job inherits its test suite from the framework
* Comfort with GraphFrames or lineage\-based dependency analysis to sequence integration tests based on upstream/downstream job relationships
**Nice to Have:**
* Databricks certification
* Experience working with governed data platforms using Unity Catalog
Pay: $60\.00 \- $70\.00 per hour
Experience:
* Databricks: 5 years (Required)
* DataStage: 5 years (Preferred)
* Metadata: 5 years (Preferred)
* AWS: 5 years (Preferred)
* Data Validation: 5 years (Preferred)
* Deeq: 1 year (Preferred)
* Great Expectations: 1 year (Preferred)
* Pyspark: 5 years (Required)
* Iceberg: 3 years (Preferred)
Work Location: Remote
Get jobs like this in your inbox
Join thousands of digital nomads getting the best remote jobs delivered weekly. Free, no spam.
Similar Jobs
Staff Data Scientist, Growth Analytics
Foodsmart
Full-Time$170k – $200kRemote
2d ago
ReactExpressPythonGo+3
Data Engineer Consultant
DAS42
Full-Time$95k – $125kRemote
3d ago
ExpressPythonGoAWS+4
Lead Data Engineer
Redwood Logistics
Full-Time$120k – $140kRemote
3d ago
PythonGoMySQLAWS+6
Sr AI Engineer
Coterie Insurance
Full-Time$115k – $180kRemoteVisa
3d ago
FastAPIPythonGoAWS+6