Senior Data Engineer

Job not on LinkedIn

November 18

🗣️🇧🇷🇵🇹 Portuguese Required

Apply Now
Logo of CI&T

CI&T

Artificial Intelligence • Cloud Services • SaaS

CI&T is a global tech transformation specialist focusing on helping organizations navigate their technology journey. With services spanning from application modernization and cloud solutions to AI-driven data analytics and customer experience, CI&T empowers businesses to accelerate their growth and maximize operational efficiency. The company emphasizes digital product design, strategy consulting, and immersive experiences, ensuring a robust support system for enterprises in various industries.

5001 - 10000 employees

Founded 1995

🤖 Artificial Intelligence

☁️ SaaS

💰 $5.5M Venture Round on 2014-04

📋 Description

• Your mission will be to serve as the hands-on pillar of the Data Governance core, materializing the blueprint into reusable components. • Implement pipelines, services and automations for metadata/lineage, Data Quality-as-Code, Policy-as-Code and the Semantic Layer with observability and performance on GCP. • Build metadata/lineage pipelines and connectors (batch/streaming) with versioning and end-to-end traceability. • Implement Data Quality-as-Code frameworks: tests, monitors, dataset SLOs/SLIs, alerting and auto-remediation. • Implement Policy-as-Code and controls in GCP: classification (Cloud DLP), masking/tokenization, enforcement (policy tags, IAM Conditions, row- and column-level security). • Model and publish the Semantic Layer (Looker semantic model), and create data contracts and validators within CI/CD. • Develop reusable packages/templates, technical documentation and runbooks. • Integrate with corporate platforms and APIs, ensuring performance, reliability and cost-efficiency. • Instrument observability (logs, metrics, tracing) and participate in incident response, promoting continuous improvement.

🎯 Requirements

• 5+ years in data engineering with strong experience in Python, SQL and PySpark. • Hands-on experience with batch and streaming ETL/ELT on GCP (Dataflow/Beam, Pub/Sub, Dataproc/Spark, Datastream) and orchestration with Composer (Airflow). • Data Quality/Observability (Great Expectations, Soda, Monte Carlo) and metadata/lineage tools (Data Catalog, DataHub/OpenMetadata, OpenLineage). • Policy-as-Code on GCP: Cloud DLP, policy tags, IAM/IAM Conditions; security and privacy (LGPD). • Git, CI/CD (Cloud Build), IaC (Terraform); performance tuning and troubleshooting in BigQuery/Spark. • Strong communication and collaboration skills with product, security and compliance teams.

🏖️ Benefits

• Health and dental plans • Meal and food vouchers • Childcare assistance • Extended parental leave • Partnerships with gyms and health and wellness professionals via Wellhub (Gympass) TotalPass • Profit-sharing (PLR) • Life insurance • Continuous learning platform (CI&T University) • Discount club • Free online platform dedicated to promoting physical and mental health and well-being • Prenatal and responsible parenthood course • Partnerships with online course platforms • Language learning platform • And many others

Apply Now

Similar Jobs

November 18

Data Engineer role in a global technology firm, focused on automation and data analysis for business intelligence. Engaging in data modeling, analysis, and process improvement at LATAM level.

Cloud

Python

SQL

November 17

Data Engineer developing scalable ETL/ELT pipelines in GCP and Databricks for client projects. Responsible for collaborating and implementing data governance and quality standards.

Airflow

BigQuery

Cloud

ETL

Google Cloud Platform

Numpy

Pandas

PySpark

Python

November 17

EY

10,000+ employees

💸 Finance

Senior Data Engineer at EY focusing on data migration and pipeline optimization in Brazil. Collaborating with diverse teams in a global environment.

🗣️🇧🇷🇵🇹 Portuguese Required

PySpark

Python

November 15

Engenheiro de Dados Sênior projetando e implementando soluções de ingestão e transformação no Azure. Trabalhando em parceria com equipes para disponibilizar dados confiáveis e de qualidade.

🗣️🇧🇷🇵🇹 Portuguese Required

Airflow

Azure

ETL

Pandas

PySpark

Python

Spark

SQL

Vault

November 14

Data Engineer responsible for constructing and optimizing data pipelines using Pentaho Data Integration and REST APIs. Collaborating closely with Automation Architect and BI teams to support ongoing automation projects.

🗣️🇧🇷🇵🇹 Portuguese Required

AWS

Azure

Cloud

ETL

Google Cloud Platform

Python

SQL

Built by Lior Neu-ner. I'd love to hear your feedback — Get in touch via DM or support@remoterocketship.com