4864- Senior Data Engineer ( contractor through Deel)
LOCATIONS : ARGENTINA / URUGUAY / MEXICO ( REMOTE)
- Only resumes in English and candidates based and authorized to work in the listed locations will be considered for the role.
At Compass, we’re on a mission to help everyone find their place in the world. Since 2012, we’ve been transforming the real estate industry with our end-to-end technology platform, empowering residential real estate agents to deliver outstanding service to their clients. Our culture thrives on interpersonal connectivity, collaborative impact, and bold, innovative solutions.
Data is the foundation of the Compass technologies. Our team is responsible for architecting, building, and maintaining a unified, scalable and cost effective analytics platform, including data lake, data warehouse, data pipelines, and data operational tooling to support data stakeholders across the company.
As a data engineer, you will be responsible for building, optimizing and maintaining scalable data pipelines using distributed computing on Cloud . You are a data expert who understands and optimizes data systems from the ground up. You will collaborate with data analysts and scientists to support data initiatives, and ensure consistent, optimal data delivery architecture for ongoing projects. You must be self-directed and comfortable supporting the data needs of multiple teams, systems and products.
What you will do :
Data Architecture & Pipeline Development : Design, implement, and maintain scalable, secure, and high-performance data architectures and pipelines for both real-time and batch processing. This includes building and maintaining Delta Lake pipelines (medallion patterns), writing scalable ETL logic with PySpark & Python , and operating and maintaining Apache Airflow DAGs and Databricks jobs . You'll also be responsible for deploying infrastructure as code using Terraform or CloudFormation .Data Quality & Governance : Implement a robust data quality framework with automated checks, monitoring, and alerting. Ensure proper data governance, PII handling, and data lineage are in place. You will also own S3 data layouts and policies , enforcing lifecycle, retention, and access controls.Data Operations & DevOps : Provide on-call support for data operations, monitor systems, and resolve data quality and operational issues. Automate deployments and manual processes, maintaining and extending deployment scripts and CI / CD pipelines (CircleCI, Makefile) . You will also participate in on-call rotation duties and incident response.Collaboration : Work closely with data analysts, data scientists, and other stakeholders to understand data requirements and deliver solutions that align with business goals. You will also collaborate with infrastructure teams to onboard new data sources and downstream consumers.Technical Leadership : Write modular, testable code and maintain technical documentation and SOPs. You will also be a key resource for troubleshooting and optimizing data workflows.What we are looking for :
Education : Bachelor's or Master's degree in Computer Science, Information Technology, or a related engineering field.Fluent English spoken and written is a mustExperience : 4+ years of solid professional experience building and maintaining production data pipelines and ETL tooling.Core Programming : Strong Python skills including testing (pytest), packaging, and scripting for automation.Big Data Frameworks : Advanced knowledge of and hands-on experience with PySpark and Apache Spark for distributed data processing and performance tuning.Cloud & Platform Expertise :Expert-level experience with Apache Airflow (DAG patterns, custom operators, scheduling, and troubleshooting).Hands-on Databricks experience (job configuration, cluster sizing, and REST API).Practical experience with AWS S3, IAM , and common S3 data patterns.Database & Data Lake :Expert SQL skills for complex transformations and data validation.Deep knowledge of Delta Lake / Lakehouse architecture, including table management, partitioning, schema evolution, and vacuum / optimize processes.Experience with both relational and non-relational databases.DevOps : Experience with version control systems ( Git ) and CI / CD for data pipelines.Desirable Skills
Experience with Infrastructure as Code (IaC) tools like Terraform and / or CloudFormation .Familiarity with other AWS data services such as Glue, Athena, and EMR .Experience integrating external sources like Salesforce, Stitch, or Kafka .Knowledge of data governance and prior experience with PII handling.Familiarity with monitoring and observability tools like Prometheus, Grafana, CloudWatch, or Sumo.Experience with machine learning workflows and LLMs for data mapping or PII masking.Proven expertise in stakeholder management and communication.Compass offers flexible work environment and benefits package :
100% remote work (Flexible around US time zones)USD fixed monthly compensation (through Deel)Paid local national holidaysAutonomy in managing delivery of your assignments and time offMac equipment provided