Braintrust Uptime Monitor
The end-to-end platform for building world-class AI apps.
Last 30 Days Performance
Average Uptime
100%
Based on 30-day monitoring period
Average Response Time
209.17ms
Mean response time across all checks
Daily Status Overview
Hover for detailsHistorical Performance
Dec-2025
99.78% uptime
Monthly Uptime
99.78%
Monthly Response Time
202ms
Daily Status Breakdown
Nov-2025
100% uptime
Monthly Uptime
100%
Monthly Response Time
238ms
Daily Status Breakdown
Oct-2025
100% uptime
Monthly Uptime
100%
Monthly Response Time
265ms
Daily Status Breakdown
Sep-2025
100% uptime
Monthly Uptime
100%
Monthly Response Time
265ms
Daily Status Breakdown
Aug-2025
99.85% uptime
Monthly Uptime
99.85%
Monthly Response Time
287ms
Daily Status Breakdown
Jul-2025
99.87% uptime
Monthly Uptime
99.87%
Monthly Response Time
330ms
Daily Status Breakdown
Jun-2025
100% uptime
Monthly Uptime
100%
Monthly Response Time
349ms
Daily Status Breakdown
May-2025
99.93% uptime
Monthly Uptime
99.93%
Monthly Response Time
363ms
Daily Status Breakdown
Apr-2025
99.65% uptime
Monthly Uptime
99.65%
Monthly Response Time
431ms
Daily Status Breakdown
Related Uptime Monitors
Explore uptime status for similar tools that also have monitoring enabled.
-
OperationalNeuralTrust
Secure, test, & scale LLMs
NeuralTrust offers a unified platform for securing, testing, monitoring, and scaling Large Language Model (LLM) applications, ensuring robust security, regulatory compliance, and operational control for enterprises.
Last checked: 3 hours ago View Status -
OperationalHumanloop
The LLM evals platform for enterprises to ship and scale AI with confidence
Humanloop is an enterprise-grade platform that provides tools for LLM evaluation, prompt management, and AI observability, enabling teams to develop, evaluate, and deploy trustworthy AI applications.
Last checked: 3 hours ago View Status -
OperationalLangtail
The low-code platform for testing AI apps
Langtail is a comprehensive testing platform that enables teams to test and debug LLM-powered applications with a spreadsheet-like interface, offering security features and integration with major LLM providers.
Last checked: 3 hours ago View Status -
OperationalBenchLLM
The best way to evaluate LLM-powered apps
BenchLLM is a tool for evaluating LLM-powered applications. It allows users to build test suites, generate quality reports, and choose between automated, interactive, or custom evaluation strategies.
Last checked: 3 hours ago View Status -
OperationalGentrace
Intuitive evals for intelligent applications
Gentrace is an LLM evaluation platform designed for AI teams to test and automate evaluations of generative AI products and agents. It facilitates collaborative development and ensures high-quality LLM applications.
Last checked: 3 hours ago View Status -
OperationalOpenLIT
Open Source Platform for AI Engineering
OpenLIT is an open-source observability platform designed to streamline AI development workflows, particularly for Generative AI and LLMs, offering features like prompt management, performance tracking, and secure secrets management.
Last checked: 3 hours ago View Status