Senior Data Engineer

November 25

Apply Now
Logo of People Data Labs

People Data Labs

API • B2B • SaaS

People Data Labs is a data infrastructure company that provides large-scale person and company datasets via developer-friendly APIs and data enrichment services. Their platform offers identity resolution, contact and firmographic enrichment, and data delivery tools used by sales, marketing, recruiting, and risk teams. They sell primarily to businesses through a SaaS/API model and emphasize data quality, coverage, and privacy/compliance capabilities.

51 - 200 employees

Founded 2015

🔌 API

🤝 B2B

☁️ SaaS

📋 Description

• Build infrastructure for ingestion, transformation, and loading an exponentially increasing volume of data from a variety of sources using Spark, SQL, AWS, and Databricks • Building an organic entity resolution framework capable of correctly merging hundreds of billions of individual entities into a number of clean, consumable datasets. • Developing CI/CD pipelines and anomaly detection systems capable of continuously improving the quality of data we're pushing into production. • Dreaming up solutions to largely undefined data engineering and data science problems.

🎯 Requirements

• 5-7+ years of industry experience with clear examples of strategic technical problem-solving and implementation • Strong software development fundamentals. • Experience with Python • Expertise with Apache Spark (Java, Scala, and/or Python-based) • Experience with SQL • Experience building scalable data processing systems (e.g., cleaning, transformation) from the ground up. • Experience using developer-oriented data pipeline and workflow orchestration (e.g., Airflow (preferred), dbt, dagster or similar) • Knowledge of modern data design and storage patterns (e.g., incremental updating, partitioning and segmentation, rebuilds and backfills) • Experience working in Databricks (including delta live tables, data lakehouse patterns, etc.) • Experience with cloud computing services (AWS (preferred), GCP, Azure or similar) • Experience with data warehousing (e.g., Databricks, Snowflake, Redshift, BigQuery, or similar) • Understanding of modern data storage formats and tools (e.g., parquet, ORC, Avro, Delta Lake)

🏖️ Benefits

• Stock • Competitive Salaries • Unlimited paid time off • Medical, dental, & vision insurance • Health, fitness, and office stipends • The permanent ability to work wherever and however you want

Apply Now

Similar Jobs

November 25

Data Engineer designing and maintaining data pipelines and reporting systems at CorServ Holdings Inc. Building data structures and delivering actionable solutions for business needs.

Java

JavaScript

NoSQL

PHP

Python

SQL

November 25

Nift

11 - 50

Data Engineer responsible for building and scaling data infrastructure for Nift’s marketing platform. Collaborate with data science and product teams to enhance data solutions.

Airflow

Amazon Redshift

AWS

Cloud

Docker

ETL

Kafka

Kubernetes

PySpark

Spark

SQL

Terraform

November 25

Senior Data Architect managing scalable data architecture for financial insights and analytics at Curinos. Leading the development of data strategy and ensuring data governance across reporting functions.

Azure

ETL

PySpark

Python

Scala

SQL

Unity

November 25

Guidehouse

10,000+ employees

Data Migration Lead overseeing the full lifecycle of data migration for EHR implementation. Collaborating with stakeholders to ensure data integrity and compliance in healthcare technology.

Cloud

ETL

Oracle

November 25

GenAI Data Engineer designing and scaling AI data systems powering Dynatron’s intelligent SaaS platform. Collaborate on AI infrastructure and machine learning capabilities for automotive services.

Airflow

Amazon Redshift

AWS

Cloud

Python

SQL

Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or support@remoterocketship.com