
API • Artificial Intelligence • Cloud Solutions
Leega is a leading technology solutions provider in Latin America, specializing in data analytics and cloud solutions. As the first company in the region certified by Google Cloud for Data Analytics, Leega offers a range of services including application development, machine learning, and risk management analytics. The firm partners with major cloud services such as AWS and Microsoft Azure to help businesses enhance their data management and transition effectively to the cloud, ultimately driving digital transformation and innovation.
November 12
🗣️🇧🇷🇵🇹 Portuguese Required

API • Artificial Intelligence • Cloud Solutions
Leega is a leading technology solutions provider in Latin America, specializing in data analytics and cloud solutions. As the first company in the region certified by Google Cloud for Data Analytics, Leega offers a range of services including application development, machine learning, and risk management analytics. The firm partners with major cloud services such as AWS and Microsoft Azure to help businesses enhance their data management and transition effectively to the cloud, ultimately driving digital transformation and innovation.
• Understand business needs and ML/AI requirements: Collaborate with stakeholders and Senior Data Scientists to understand business problems and translate them into testable hypotheses and Machine Learning projects. • Participate in data- and AI-driven discussions: Contribute initial analyses, insights, and model proposals to address data and AI challenges. • Act as an interface between Technology and Data Science teams: Work closely with Data Engineers and Architects to ensure data availability and efficient model training. • Data preparation and exploration in Azure Databricks: Perform cleaning, transformation, standardization, and enrichment of raw data using the processing power of Azure Databricks and Spark. • Conduct Exploratory Data Analysis (EDA) to understand distributions, identify outliers, and select relevant features (Feature Engineering). • Develop and validate ML models: Build, train, and validate predictive models (regression, classification, clustering) using data science frameworks such as Scikit-Learn, TensorFlow, or PyTorch. • Document and present results and model performance metrics (e.g., AUC, F1-Score). • Support operationalization (MLOps): Assist in transitioning developed models to production environments, using tools like MLflow for experiment and model tracking and management in Azure Databricks. • Monitor data and model quality: Track data quality, integrity and drift, and monitor model performance in production, proposing adjustments when necessary.
• Experience with CI/CD tools and automation of machine learning workflows. • Data observability. • Familiarity with data monitoring and validation tools. • Pipeline orchestration. • Knowledge of SQL. • Data Lakes and Data Warehouses: • Ability to manage Data Lakes for AI data preparation, with a focus on Azure Databricks. • Feature storage. • Azure + Databricks experience. • Ability to collaborate with other Data Scientists to prepare datasets for AI models, ensuring high quality and performance. • Implementation of automated pipelines for model training, validation, deployment and monitoring. • Ability to define and manage feature repositories for AI models (Feature Engineering). • Knowledge of LGPD/GDPR to ensure regulatory compliance in data handling for AI. • Experience working with agile methodologies (Scrum/Kanban). • Ability to understand business requirements and translate them into technical AI solutions. • Focus on Azure Databricks: Practical knowledge and hands-on experience using Azure Databricks for data processing and analysis, including the use of notebooks (Python/Spark) and cluster management. • Understanding of the Feature Store concept and experience consuming data from Data Lakes (preferably Azure Data Lake Storage) and structuring data for model consumption. • Languages: Python (Required) for data manipulation (Pandas, NumPy) and model development. • Practical experience with major libraries and frameworks such as Scikit-Learn, TensorFlow/Keras or PyTorch and libraries like XGBoost/LightGBM. • Basic knowledge of MLflow for tracking and versioning models and experiments in the Databricks environment. • Proficiency in SQL/Spark SQL for querying and manipulating large-scale data.
• 🏥 Health insurance (Porto Seguro) • 🦷 Dental plan (Porto Seguro) • 💰 Profit Sharing (PLR) • 👶 Childcare assistance • 🍽️ Alelo Meal and Food Allowance • 💻 Home office allowance • 📚 Partnerships with educational institutions • 🚀 Support for certifications, including cloud • 🎁 Livelo points • 🏋️♂️ TotalPass • 🧘♂️ Mindself
Apply NowNovember 9
Junior Data Scientist developing analytics solutions for gaming industry. Collaborating on data pipelines and machine learning models to enhance player engagement and product innovation.
April 26
Remote data scientist position focused on predictive models and large data sets. Requires expertise in Python and machine learning libraries.
🗣️🇧🇷🇵🇹 Portuguese Required