Position: Senior Data Pipeline Engineer

Location: 100% Remote, 40-hour Contract Position

About Clever

At Clever Real Estate, we envision a world where every person makes real estate moves that bring them closer to their life's goals. To achieve that vision, we connect people with the most trustworthy advice and the best solutions for every step of their real estate journey.

We've built the leading online education platform in real estate and are rated the #1 real estate company on TrustPilot — and our solutions have helped people save over $160 million on real estate fees. Clever is venture-backed, profitable, and still evolving. If you're looking for a challenging role with deep ownership, smart co-workers, and real-world impact– read on to learn more!

About the Role [100% Remote]

Clever is seeking a highly skilled Senior Data Pipeline Engineer to join our team. This role involves designing, developing, and maintaining robust data pipelines for efficient extraction, transformation, and loading (ETL/ELT) of data from various sources into our data ecosystem, which includes internal, B2B, and consumer-facing tools. This position requires a strong understanding of data engineering principles, distributed computing systems, and experience with big data technologies.

Key Responsibilities:

  • Design and Maintenance: Build and maintain scalable and reliable data pipelines to ingest, process, and transform large volumes of data from diverse sources.
  • Collaboration: Work with cross-functional teams, including data scientists, software engineers, product owners, and business analysts, to understand data requirements and translate them into technical solutions.
  • Optimization: Optimize data pipelines for performance, scalability, and cost-efficiency, ensuring timely delivery of high-quality data to downstream systems and applications.
  • Quality Assurance: Implement data quality checks, error handling mechanisms, and monitoring solutions to ensure the integrity and reliability of data pipelines.
  • Innovation: Stay current with emerging technologies and industry trends in data engineering and apply best practices to continuously improve our data pipeline infrastructure.
  • Leadership: Provide technical leadership and mentorship to junior members of the data engineering team, fostering a culture of collaboration, innovation, and excellence.

What to Expect

First 30 Days:

  • Gain a comprehensive understanding of all the data sources we utilize.
  • Begin developing high-level models and diagrams to map out where the data is coming from and where it is stored.
  • Create diagrams of data flow and storage architecture.

First 90 Days:

  • Ensure the delivery of content products on a weekly basis.
  • Complete ~50% of the data modeling and implementation tasks.
  • Establish a strong collaborative relationship with Product Managers, actively planning for the next 6 to 12 months.
  • Research, vet, and select appropriate data build tools and data warehouse solutions (e.g., Redshift, Snowflake, BigQuery).

First 6 Months:

  • Develop and release a minimum of 4 to 5 content series [6 is above and beyond!].
  • Build a functional data pipeline that addresses the needs and organizes Clever's data effectively.


  • Bachelor's or Master's degree in Computer Science, Engineering, or a related field, or equivalent experience.
  • 5+ years of experience in data engineering, with a focus on building and maintaining data pipelines.
  • Proficiency in object-oriented programming languages and experience with frameworks/libraries for data processing (e.g., Apache Spark, Apache Beam).
  • Experience with high volume data warehouses (e.g., Redshift, Snowflake, BigQuery).
  • Strong understanding of distributed computing concepts and experience with big data technologies (e.g., Apache Kafka, Apache Flink).
  • Experience with cloud platforms, particularly AWS, and related services such as AWS Glue and AWS Data Pipeline.
  • Familiarity with data pipeline orchestration tools such as Apache Airflow for workflow management and scheduling.
  • Experience with modern data modeling tools such as DBT (Data Build Tool) for transforming data into analytics-friendly formats.
  • Excellent problem-solving skills, attention to detail, and ability to work effectively in a fast-paced environment.
  • Strong communication skills and ability to collaborate with cross-functional teams to deliver high-quality solutions.

Technologies We Use

  • Directus
  • Google Colab
  • Jira for issue tracking
  • Github Actions, BitRise, AWS, and Heroku for infrastructure and CI/CD integration


We are looking for a full-time Senior Data Pipeline Engineer who can work ~40 hours per week. Clever Real Estate operates on a flexible work schedule with core business hours from 9 AM to 5 PM CT, Monday through Friday. We accommodate variations in schedules as long as team members have a significant overlap (~6 hours) with the core business hours.

Equal Employment Opportunity Employer Statement:

Clever Real Estate provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state or local laws.

Senior Data Pipeline Engineer

Solicitar ahora
Back to search page