
eCommerce • Marketing • AI
Rithum is a comprehensive e-commerce platform designed to empower brands, retailers, and suppliers to effectively launch and scale their businesses. Offering a variety of solutions including multichannel marketing, fulfillment management, and AI-driven supplier discovery, Rithum helps users optimize their online presence and streamline operations across a vast network of marketplaces. With a focus on flexibility and efficiency, Rithum aims to transform commerce by creating profitable and engaging shopping experiences.
501 - 1000 employees
Founded 1997
🛍️ eCommerce
September 17

eCommerce • Marketing • AI
Rithum is a comprehensive e-commerce platform designed to empower brands, retailers, and suppliers to effectively launch and scale their businesses. Offering a variety of solutions including multichannel marketing, fulfillment management, and AI-driven supplier discovery, Rithum helps users optimize their online presence and streamline operations across a vast network of marketplaces. With a focus on flexibility and efficiency, Rithum aims to transform commerce by creating profitable and engaging shopping experiences.
501 - 1000 employees
Founded 1997
🛍️ eCommerce
• Design and implement scalable ETL/ELT workflows for batch and streaming data using AWS primitives (S3, Kinesis, Glue, Redshift, Athena) • Architect and maintain cloud-native data platforms with automated ingestion, transformation, and governance using DBT, Apache Spark, Delta Lake, Airflow, Databricks • Work with Product, BI, Support, Data Scientists and Engineers to support data needs and resolve technical challenges • Optimize data lake/lakehouse infrastructure to support AI workloads and large-scale analytics • Ensure data quality, lineage, and observability and develop/enforce data governance and privacy protections • Partner with Data Scientists to optimize pipelines for model training, inference, and continuous learning • Build self-healing data pipelines with AI-driven error detection, root cause analysis, and automated remediation • Implement intelligent data lineage tracking and AI-assisted data discovery systems with natural language interfaces • Leverage AI coding assistants to accelerate development, generate complex SQL, and optimize data pipeline code • Develop data quality monitoring (anomaly detection, profiling) and ML-driven pipeline orchestration • Generate and maintain living documentation and participate in full software development lifecycle • Mentor junior engineers, lead tool evaluation and adoption, and drive innovation in data architecture • Participate in on-call rotation as needed
• 3+ years of experience in data engineering, including building and maintaining large-scale data pipelines • Extensive experience in SQL RDBMS (SQLServer or similar) with dimensional modeling using star schema, and foundational data warehousing concepts • Hands-on experience with AWS services such as Redshift, Athena, S3, Kinesis, Lambda, Glue • Experience with DBT, Databricks or similar data platform tooling • Experience working with structured and unstructured data and implementing data quality frameworks • Demonstrated experience using AI coding tools (GitHub Copilot, Cursor, or similar), with understanding of prompt engineering • Understanding of AI/ML concepts and data requirements, including feature stores, model versioning, and real-time inference pipelines • Excellent communication and collaboration skills • Preferred: Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field • Preferred: Experience in a SaaS or e-commerce environment with AI/ML products • Preferred: Knowledge of stream processing frameworks like Kafka, Flink, or Spark Structured Streaming • Preferred: Familiarity with LLMOps and AI model deployment patterns in data infrastructure • Preferred: Experience with AI-powered data tools such as automated data catalogs, intelligent monitoring systems, or AI-assisted query optimization • Preferred: Experience with containerization and orchestration tools like Docker and Kubernetes • Willingness to travel up to 10%
• Medical, Dental and Psychology benefits • Life insurance and disability benefits • Competitive time off package with 25 Days of PTO, 13 Company-Paid Holidays, 2 Wellness days and 1 Paid Volunteer Day • Voucher program for Transportation, Meals & Childcare • Work from the Madrid co-working space, if desired • Remote Working Stipend: €40/month automatically applied in payroll • Access to wellbeing tools such as the Calm App and an Employee Assistance Program • Professional development stipend and learning and development offerings • Charitable contribution match per team member • Industry-competitive compensation and total rewards benefits • Remote-first working conditions and generous time off
Apply NowAugust 20
Mid Data Engineer building data architecture and storage solutions for Volkswagen Group Services. Lead technical data strategy and implement cloud-based data platforms.
August 12
Join Prima’s Engineering team to bridge ML/data science with engineering. Build data products and pipelines for motor insurance growth.
August 7
Data Engineer developing Azure and Big Data solutions for a global IT company. Collaborates in a skilled development team with focus on CI/CD and data integrity.
July 30
Seeking a Senior Data Engineer to architect scalable data pipelines using Google Cloud Platform and Databricks.
July 11
Join Volkswagen as a Mid Data Engineer in Barcelona to enhance enterprise data quality, integrity, and accessibility.