Staff Data Engineer, Spark

Job not on LinkedIn

September 18

Apply Now
Logo of Wizard

Wizard

Artificial Intelligence • eCommerce • Retail

Wizard AI is a company that provides a tailored shopping experience powered by artificial intelligence. It offers a unique text-based service that curates product recommendations across the internet, utilizing AI models that understand user preferences to predict needs and make personalized suggestions. Wizard AI simplifies shopping by finding, ordering, and tracking products, as well as managing returns through SMS communication. This service is designed to save time and improve the convenience of online shopping by efficiently juggling product data, customer reviews, and other digital content to make informed recommendations for customers.

51 - 200 employees

🤖 Artificial Intelligence

🛍️ eCommerce

🛒 Retail

💰 $50M Series A on 2021-10

📋 Description

• Design and evolve scalable, distributed data infrastructure across cloud platforms • Build and maintain real time and batch data processing pipelines supporting analytics and AI/ML workloads • Develop and manage integrations with third party e-commerce platforms to expand our data ecosystem • Ensure data availability, reliability and quality through monitoring and automated auditing • Partner with engineering, AI and product teams on data solutions for business critical needs • Mentor and support data engineers, establishing best practices and code quality standards

🎯 Requirements

• 5+ years of software development and data engineering experience with demonstrated ownership of production grade data infrastructure • Bachelor's degree in Computer Science or a related field, or equivalent practical experience. • Deep expertise scaling Spark in production (Databricks, EMR, etc) • Strong understanding of distributed computing and modern data modeling for scalable systems • Proficient in Python with experience implementing software engineering best practices • Hands-on experience with both relational (MySQL / PostgreSQL) and NoSQL (MongoDB, DynamoDB, Cassandra) databases • Strong communicator with experience influencing cross functional stakeholders

Apply Now

Similar Jobs

September 16

Principal Data Engineer leading design and implementation of Trella Health's multi-tenant data lake handling PHI. Driving platform adoption, mentoring team, and maintaining critical pipelines.

Apache

AWS

Azure

Cloud

Docker

Google Cloud Platform

Kafka

Kubernetes

Python

Spark

SQL

Terraform

September 12

Staff DataOps Engineer applying DevOps principles to GoodLeap's data platform; drive automation, observability, and reliability.

Airflow

Cloud

Terraform

September 6

Data Architect designing enterprise Databricks Lakehouse for Live Nation Entertainment. Lead architecture, modeling, governance, and mentorship across data platform.

AWS

Azure

Cloud

ETL

Hadoop

Kafka

NoSQL

Spark

August 24

Data Engineer role focused on SAP BW solutions development. Collaborate with teams and optimize data warehousing processes in this remote position.

Cloud

ETL

SQL

August 22

Lead design and implementation of scalable AWS/Azure data lakes, warehouses, and streaming platforms for Tkxel. Mentor engineers and ensure governance and security.

Airflow

Apache

AWS

Azure

Cloud

Docker

ETL

Informatica

Kafka

Kubernetes

Node.js

NoSQL

Spark

Tableau

Terraform

Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or support@remoterocketship.com