Infrastructure Engineer

Job not on LinkedIn

October 23

Apply Now
Logo of Cobalt AI

Cobalt AI

Artificial Intelligence • Security • Enterprise

Cobalt AI is a company that specializes in AI-powered security monitoring solutions. They offer a suite of products designed for proactive security management, including the Cobalt Monitoring Intelligence platform, the Cobalt Command Center, and Cobalt Security Robots. These products integrate artificial intelligence with human verification to provide real-time monitoring of security systems, such as surveillance cameras and alarms, and enhance security operations with features like threat intelligence and travel risk monitoring. Cobalt AI's solutions aim to improve incident detection accuracy, reduce false positives, and automate repetitive security tasks, allowing organizations to enhance their security while saving costs.

📋 Description

• Take ownership of building, deploying, and monitoring highly available Kubernetes-based infrastructure. • Develop and scale the orchestration platform for our global fleet of AI-enabled cameras and Edge Processors ensuring robust, secure, and reliable operations in diverse network environments. • Take ownership of the entire edge device lifecycle, from automated provisioning and secure software deployment to remote monitoring, diagnostics, and telemetry collection at scale. • Apply infrastructure-as-code best practices to declaratively manage infrastructure and deployments via Terraform and Argo CD. • Identify and champion Developer Experience (DevX) improvements across our technologies and processes. • Manage CI/CD pipelines. • Develop observability solutions to ensure performance and reliability targets are met as scale and complexity increase. • Support integration efforts with Access Control and Video Management systems. • Critically assess and address infrastructure security risks. • Lead by example, establishing and maintaining best practices. • Gather input from multiple teams to ensure high-quality outcomes. • Deploy changes with immediate impacts on our product, worldwide.

🎯 Requirements

• Demonstrated expertise in Linux, Networking, and Distributed systems. • Deep understanding of backend development best practices. • Hands-on experience managing highly available deployments with container orchestration and declarative IaC. • Knowledge of monitoring and observability tools and best practices. • Experience optimizing cloud provider spends. • Hands-on experience with hybrid cloud/edge architectures, including device identity, secure communication patterns, and fleet orchestration at scale. • Eagerness to work with, mentor, and learn from peers through code reviews, design documents, and pair programming. • Authorization to work in the United States.

🏖️ Benefits

• Positive, people-first culture • Work-life balance • Professional development investment

Apply Now

Similar Jobs

October 23

Build out Tempo’s infrastructure stack for a blockchain design partner focused on stablecoins and payments. Ensure efficient processes and deployment within engineering teams.

Cloud

Grafana

Kubernetes

Linux

Prometheus

Python

Rust

Terraform

Go

October 21

Senior Infrastructure Engineer at OpenTeams building scalable AI/ML infrastructure using tools like Kubernetes. Collaborating remotely and contributing to open-source MLOps projects.

AWS

Azure

Cloud

Google Cloud Platform

Kubernetes

Python

Terraform

Go

October 21

Descript

51 - 200

Infrastructure Engineer enhancing product performance and reliability at Descript. Collaborating with engineering leadership to improve core infrastructure and mentoring engineers.

Cloud

Linux

October 18

Infrastructure Operations Engineer for Voltage Park's 24/7 team ensuring stability and performance of AI infrastructure. Collaborating with other teams to deliver high-performance environments.

Ansible

AWS

Kubernetes

Linux

NFS

Prometheus

Python

Terraform

Go

October 17

Data Infrastructure Engineer designing and maintaining data systems that power AI model training. Building scalable ingestion pipelines and collaborating with researchers for diverse datasets.

Airflow

AWS

Azure

Cloud

Distributed Systems

ETL

Google Cloud Platform

Java

Python

Ray

Scala

Spark

SQL