Data Engineer, Python, PySpark, AWS Glue, Amazon Athena, SQL, Apache Airflow

Job not on LinkedIn

September 14

Apply Now
Logo of Wizdaa

Wizdaa

Recruitment • HR Tech • SaaS

Wizdaa is a company that provides access to top-tier remote developers, specializing in helping startups build their dream development teams in the U. S. time zone. They offer a meticulous six-stage human and AI screening process to ensure access to the top 1% of engineering talent. Wizdaa's services include managing hiring processes, onboarding, payroll, benefits, and taxes, allowing startups to focus on core business matters. Known for competitive rates averaging $30/hour, Wizdaa emphasizes cultural fit, technical excellence, and English fluency among their developers. The company prides itself on delivering tailored, cost-effective solutions that maximize startups' runway and success. They also offer insights into leveraging AI tools and aligning remote teams with U. S. time zones to boost productivity.

📋 Description

• Build, optimize, and scale data pipelines and infrastructure using Python, TypeScript, Apache Airflow, PySpark, AWS Glue, and Snowflake. • Design, operationalize, and monitor ingest and transformation workflows: DAGs, alerting, retries, SLAs, lineage, and cost controls. • Collaborate with platform and AI/ML teams to automate ingestion, validation, and real-time compute workflows; work toward a feature store. • Integrate pipeline health and metrics into engineering dashboards for full visibility and observability. • Model data and implement efficient, scalable transformations in Snowflake and PostgreSQL. • Build reusable frameworks and connectors to standardize internal data publishing and consumption.

🎯 Requirements

• 4+ years of production data engineering experience. • Deep, hands-on experience with Apache Airflow, AWS Glue, PySpark, and Python-based data pipelines. • Strong SQL skills and experience operating PostgreSQL in live environments. • Solid understanding of cloud-native data workflows (AWS preferred) and pipeline observability (metrics, logging, tracing, alerting). • Proven experience owning pipelines end-to-end: design, implementation, testing, deployment, monitoring, and iteration. • Experience with Snowflake performance tuning and cost optimization (warehouses, partitions, clustering, query profiling) (preferred). • Real-time or near-real-time processing experience (streaming ingestion, incremental models, CDC) (preferred). • Hands-on experience with a backend TypeScript framework (e.g., NestJS) is a strong plus. • Experience with data quality frameworks, contract testing, or schema management (e.g., Great Expectations, dbt tests, OpenAPI/Protobuf/Avro). • Background in building internal developer platforms or data platform components (connectors, SDKs, CI/CD for data) (preferred). • Work hours aligned with EST or PT time zone.

🏖️ Benefits

• Fully remote position • Compensation paid in USD • Work hours aligned with EST or PT time zone

Apply Now

Similar Jobs

September 13

Full Stack Developer building React Electron desktop apps for a US data science startup. Shipping features and integrating real-time data with Node.js/Python backends.

Cloud

Electron

Firebase

JavaScript

Node.js

Python

React

September 11

PrideLogic

11 - 50

Data Engineer building, optimizing, and operating AWS/PySpark/Airflow pipelines and Snowflake transformations. Collaborates with platform and AI/ML teams on ingestion, observability, and reusable connectors.

Airflow

Apache

AWS

Cloud

Node.js

Postgres

PySpark

Python

SQL

TypeScript

September 11

Python developer building data pipelines and automations for an Amazon-focused e-commerce company. Connect APIs, web scraping, Google Sheets, and deploy automations in AWS.

AWS

EC2

Pandas

Python

Selenium

August 28

Senior Backend Developer building Python/Django and Go services for a cybersecurity company. Designing scalable APIs, optimizing performance, and mentoring engineers.

AWS

BigQuery

Cloud

Cyber Security

Django

ElasticSearch

Google Cloud Platform

Linux

MySQL

Postgres

Python

Redis

Go

August 25

Lead backend engineer building scalable asset monitoring services for Motive's fleet operations platform.

AWS

Azure

Cassandra

Cloud

Distributed Systems

DynamoDB

Google Cloud Platform

Java

NoSQL

Postgres

Redis

Ruby

Ruby on Rails

Go

Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or support@remoterocketship.com