Data Engineer – Python, PySpark, AWS Glue, Amazon Athena, SQL, Apache Airflow

Job not on LinkedIn

September 10

Apply Now
Logo of PrideLogic

PrideLogic

PrideLogic is a company that does not currently have detailed information available as their website is under construction. More details may be provided in the future once updates are made available on their site.

11 - 50 employees

📋 Description

• Build, optimize, and scale data pipelines and infrastructure using Python, TypeScript, Apache Airflow, PySpark, AWS Glue, and Snowflake. • Design, operationalize, and monitor ingest and transformation workflows: DAGs, alerting, retries, SLAs, lineage, and cost controls. • Collaborate with platform and AI/ML teams to automate ingestion, validation, and real-time compute workflows; work toward a feature store. • Integrate pipeline health and metrics into engineering dashboards for full visibility and observability. • Model data and implement efficient, scalable transformations in Snowflake and PostgreSQL. • Build reusable frameworks and connectors to standardize internal data publishing and consumption.

🎯 Requirements

• 4+ years of production data engineering experience. • Deep, hands-on experience with Apache Airflow, AWS Glue, PySpark, and Python-based data pipelines. • Strong SQL skills and experience operating PostgreSQL in live environments. • Solid understanding of cloud-native data workflows (AWS preferred) and pipeline observability (metrics, logging, tracing, alerting). • Proven experience owning pipelines end-to-end: design, implementation, testing, deployment, monitoring, and iteration. • Experience with Snowflake performance tuning (warehouses, partitions, clustering, query profiling) and cost optimization. • Real-time or near-real-time processing experience (e.g., streaming ingestion, incremental models, CDC). • Hands-on experience with a backend TypeScript framework (e.g., NestJS) is a strong plus. • Experience with data quality frameworks, contract testing, or schema management (e.g., Great Expectations, dbt tests, OpenAPI/Protobuf/Avro). • Background in building internal developer platforms or data platform components (connectors, SDKs, CI/CD for data).

Apply Now

Similar Jobs

September 10

GFT Technologies

10,000+ employees

🔒 Cybersecurity

📋 Compliance

☁️ SaaS

Senior Golang engineer building AWS microservices for GFT's BatchOnCloud platform. Implement COBOL/PySpark compatibility and support batch troubleshooting.

🇧🇷 Brazil – Remote

⏰ Full Time

🟠 Senior

🔙 Backend Engineer

🗣️🇧🇷🇵🇹 Portuguese Required

AWS

PySpark

Go

September 10

GFT Technologies

10,000+ employees

🔒 Cybersecurity

📋 Compliance

☁️ SaaS

Senior .NET Backend Developer at GFT leading microservices architecture and technical teams. Focus on .NET Core, cloud, containers, and dev practices.

🇧🇷 Brazil – Remote

⏰ Full Time

🟡 Mid-level

🟠 Senior

🔙 Backend Engineer

🗣️🇧🇷🇵🇹 Portuguese Required

September 10

GFT Technologies

10,000+ employees

🔒 Cybersecurity

📋 Compliance

☁️ SaaS

Develop Kotlin/Java microservices, CI/CD, AWS, observability; lead architecture and mentor teams at GFT.

🇧🇷 Brazil – Remote

⏰ Full Time

🟡 Mid-level

🟠 Senior

🔙 Backend Engineer

🗣️🇧🇷🇵🇹 Portuguese Required

September 9

Wizdaa

11 - 50

🎯 Recruiter

👥 HR Tech

☁️ SaaS

Build and operate ETL pipelines with PySpark, Airflow, and AWS Glue; model data in Snowflake and PostgreSQL for analytics and ML.

🇧🇷 Brazil – Remote

⏰ Full Time

🟡 Mid-level

🟠 Senior

🔙 Backend Engineer

September 9

CI&T

5001 - 10000

🤖 Artificial Intelligence

☁️ SaaS

Senior Java Developer building scalable Java microservices for benefits-sector client at CI&T. Focus on APIs, architecture, cloud, and automated testing.

🇧🇷 Brazil – Remote

💰 $5.5M Venture Round on 2014-04

⏰ Full Time

🟠 Senior

🔙 Backend Engineer

🗣️🇧🇷🇵🇹 Portuguese Required

Developed by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or support@remoterocketship.com