
Artificial Intelligence • Cloud Services • SaaS
CI&T is a global tech transformation specialist focusing on helping organizations navigate their technology journey. With services spanning from application modernization and cloud solutions to AI-driven data analytics and customer experience, CI&T empowers businesses to accelerate their growth and maximize operational efficiency. The company emphasizes digital product design, strategy consulting, and immersive experiences, ensuring a robust support system for enterprises in various industries.
November 18
🗣️🇧🇷🇵🇹 Portuguese Required

Artificial Intelligence • Cloud Services • SaaS
CI&T is a global tech transformation specialist focusing on helping organizations navigate their technology journey. With services spanning from application modernization and cloud solutions to AI-driven data analytics and customer experience, CI&T empowers businesses to accelerate their growth and maximize operational efficiency. The company emphasizes digital product design, strategy consulting, and immersive experiences, ensuring a robust support system for enterprises in various industries.
• Your mission will be to serve as the hands-on pillar of the Data Governance core, materializing the blueprint into reusable components. • Implement pipelines, services and automations for metadata/lineage, Data Quality-as-Code, Policy-as-Code and the Semantic Layer with observability and performance on GCP. • Build metadata/lineage pipelines and connectors (batch/streaming) with versioning and end-to-end traceability. • Implement Data Quality-as-Code frameworks: tests, monitors, dataset SLOs/SLIs, alerting and auto-remediation. • Implement Policy-as-Code and controls in GCP: classification (Cloud DLP), masking/tokenization, enforcement (policy tags, IAM Conditions, row- and column-level security). • Model and publish the Semantic Layer (Looker semantic model), and create data contracts and validators within CI/CD. • Develop reusable packages/templates, technical documentation and runbooks. • Integrate with corporate platforms and APIs, ensuring performance, reliability and cost-efficiency. • Instrument observability (logs, metrics, tracing) and participate in incident response, promoting continuous improvement.
• 5+ years in data engineering with strong experience in Python, SQL and PySpark. • Hands-on experience with batch and streaming ETL/ELT on GCP (Dataflow/Beam, Pub/Sub, Dataproc/Spark, Datastream) and orchestration with Composer (Airflow). • Data Quality/Observability (Great Expectations, Soda, Monte Carlo) and metadata/lineage tools (Data Catalog, DataHub/OpenMetadata, OpenLineage). • Policy-as-Code on GCP: Cloud DLP, policy tags, IAM/IAM Conditions; security and privacy (LGPD). • Git, CI/CD (Cloud Build), IaC (Terraform); performance tuning and troubleshooting in BigQuery/Spark. • Strong communication and collaboration skills with product, security and compliance teams.
• Health and dental plans • Meal and food vouchers • Childcare assistance • Extended parental leave • Partnerships with gyms and health and wellness professionals via Wellhub (Gympass) TotalPass • Profit-sharing (PLR) • Life insurance • Continuous learning platform (CI&T University) • Discount club • Free online platform dedicated to promoting physical and mental health and well-being • Prenatal and responsible parenthood course • Partnerships with online course platforms • Language learning platform • And many others
Apply NowNovember 18
Data Engineer role in a global technology firm, focused on automation and data analysis for business intelligence. Engaging in data modeling, analysis, and process improvement at LATAM level.
Cloud
Python
SQL
November 17
Data Engineer developing scalable ETL/ELT pipelines in GCP and Databricks for client projects. Responsible for collaborating and implementing data governance and quality standards.
Airflow
BigQuery
Cloud
ETL
Google Cloud Platform
Numpy
Pandas
PySpark
Python
November 17
Senior Data Engineer at EY focusing on data migration and pipeline optimization in Brazil. Collaborating with diverse teams in a global environment.
🗣️🇧🇷🇵🇹 Portuguese Required
PySpark
Python
November 15
Engenheiro de Dados Sênior projetando e implementando soluções de ingestão e transformação no Azure. Trabalhando em parceria com equipes para disponibilizar dados confiáveis e de qualidade.
🗣️🇧🇷🇵🇹 Portuguese Required
Airflow
Azure
ETL
Pandas
PySpark
Python
Spark
SQL
Vault
November 14
Data Engineer responsible for constructing and optimizing data pipelines using Pentaho Data Integration and REST APIs. Collaborating closely with Automation Architect and BI teams to support ongoing automation projects.
🗣️🇧🇷🇵🇹 Portuguese Required
AWS
Azure
Cloud
ETL
Google Cloud Platform
Python
SQL