Data Engineer Master

Job not on LinkedIn

Yesterday

Apply Now
Logo of CI&T

CI&T

Artificial Intelligence • Cloud Services • SaaS

CI&T is a global tech transformation specialist focusing on helping organizations navigate their technology journey. With services spanning from application modernization and cloud solutions to AI-driven data analytics and customer experience, CI&T empowers businesses to accelerate their growth and maximize operational efficiency. The company emphasizes digital product design, strategy consulting, and immersive experiences, ensuring a robust support system for enterprises in various industries.

5001 - 10000 employees

Founded 1995

🤖 Artificial Intelligence

☁️ SaaS

💰 $5.5M Venture Round on 2014-04

📋 Description

• Define, architect, and implement scalable data platforms and end-to-end ELT pipelines aligned with modern Lakehouse principles. • Work closely with cross-functional teams across the US, Colombia, and Brazil to ensure that our data ecosystem is reliable, future-proof, and aligned with enterprise architecture standards. • Present architectural recommendations and ensure alignment across distributed teams.

🎯 Requirements

• Expert level SQL, with demonstrated ability to optimize, refactor, and validate large-scale transformations. • Advanced Python (or similar) for automation, orchestration, and pipeline development. • Hands-on architecture and engineering experience with Snowflake, including performance tuning, security, data governance, dynamic tables, and workload management. • Advanced dbt expertise, including transformation logic, testing, documentation, deployment patterns, and CI/CD integration. • Proven production experience with Data Vault 2.0, including Hubs, Links, Satellites, PIT tables, multi-active satellites, and Business Vault patterns. • Experience with AutomateDV or equivalent frameworks is a strong asset. • Deep understanding of Data Lakehouse architectures, including medallion zone structures, incremental ingestion, and open table formats (Iceberg, Delta, Hudi is a plus). • Solid foundation in data modeling best practices, including normalized models, dimensional modeling, historization, and scalable enterprise patterns. • Ability to translate complex business requirements into robust, extensible architectural designs. • Experience orchestrating ELT/ETL workflows using Airflow, including DAG design, dependency strategies, and dynamic task generation. • Familiarity with modern orchestration frameworks such as Prefect, Dagster, or AWS Glue. • Comfort with CI/CD pipelines using GitHub Actions or similar tools, integrating dbt testing and Snowflake deployments. • Understanding of infrastructure automation, configuration-as-code, and environment management. • Nice to Have: Experience with data observability platforms (Monte Carlo, Datafold, Great Expectations). • Knowledge of Docker or Kubernetes for reproducibility and scalable deployments. • Familiarity with Kafka, AMQP, or other message brokers and event-driven architectures. • Experience working with REST/GraphQL APIs, streaming ingestion (Kinesis, Firehose), or real-time processing. • Experience supporting hybrid architectures, multi-cloud designs, or enterprise Lakehouse strategies.

🏖️ Benefits

• - Premium Healthcare • - Meal voucher • - Maternity and Parental leaves • - Mobile services subsidy • - Sick pay-Life insurance • - CI&T University • - Colombian Holidays • - Paid Vacations

Apply Now

Similar Jobs

Yesterday

Senior Software Engineer managing high-performance data pipelines for satellite observation at Slingshot Aerospace. Collaborating on scalable cloud-native systems with a focus on reliability.

Apache

AWS

Cloud

Docker

ETL

Kafka

Kubernetes

Linux

MongoDB

Postgres

Python

RabbitMQ

Spark

Yesterday

Senior Data Engineer implementing scalable database solutions for a healthcare technology company. Collaborating with teams to enhance product features and data services.

Azure

ElasticSearch

MongoDB

Node.js

NoSQL

Python

RDBMS

SQL

2 days ago

Senior Data Engineer managing and optimizing data systems within the GCP stack for Second Nature. Leading technical projects and collaborating with cross-functional teams to enhance data insights.

BigQuery

Cloud

Docker

Google Cloud Platform

Python

SQL

2 days ago

Senior Data Engineer managing and transforming data pipeline at Experian. Utilizing DBT and Snowflake for product development and research in fraud detection.

AWS

Cloud

Python

SQL

Terraform

2 days ago

Data Architect designing a data foundation to support AI/ML applications at Leidos. Responsible for creating big data systems and ensuring data access and documentation.

ETL

TypeScript

Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or support@remoterocketship.com