LlamaEdge Uptime Monitor
The easiest, smallest and fastest local LLM runtime and API server.
Last 30 Days Performance
Average Uptime
100%
Based on 30-day monitoring period
Average Response Time
75.5ms
Mean response time across all checks
Daily Status Overview
Hover for detailsHistorical Performance
Dec-2025
100% uptime
Monthly Uptime
100%
Monthly Response Time
84ms
Daily Status Breakdown
Nov-2025
99.86% uptime
Monthly Uptime
99.86%
Monthly Response Time
86ms
Daily Status Breakdown
Oct-2025
100% uptime
Monthly Uptime
100%
Monthly Response Time
84ms
Daily Status Breakdown
Sep-2025
100% uptime
Monthly Uptime
100%
Monthly Response Time
86ms
Daily Status Breakdown
Aug-2025
99.74% uptime
Monthly Uptime
99.74%
Monthly Response Time
98ms
Daily Status Breakdown
Jul-2025
100% uptime
Monthly Uptime
100%
Monthly Response Time
87ms
Daily Status Breakdown
Jun-2025
99.86% uptime
Monthly Uptime
99.86%
Monthly Response Time
94ms
Daily Status Breakdown
May-2025
99.92% uptime
Monthly Uptime
99.92%
Monthly Response Time
88ms
Daily Status Breakdown
Related Uptime Monitors
Explore uptime status for similar tools that also have monitoring enabled.
-
OperationalWasmEdge
Fast, lightweight, portable, and OpenAI compatible WebAssembly runtime for edge AI and LLM inference
WasmEdge is a cloud-native WebAssembly runtime that enables fast, lightweight, and secure AI inference and LLM applications on the edge with native GPU support and OpenAI compatibility.
Last checked: 2 hours ago View Status -
OperationalOllama
Get up and running with large language models locally
Ollama is a platform that enables users to run powerful language models like Llama 3.3, DeepSeek-R1, Phi-4, Mistral, and Gemma 2 on their local machines.
Last checked: 14 hours ago View Status -
Operationallm-studio.me
Local LLM Running & Download Platform
LM Studio is a user-friendly desktop application that allows users to run various large language models (LLMs) locally and offline, including Llama 2, PN3, Falcon, Mistral, StarCoder, and GEMMA models from Hugging Face.
Last checked: 1 hour ago View Status -
OperationalWebLLM
High-Performance In-Browser LLM Inference Engine
WebLLM enables running large language models (LLMs) directly within a web browser using WebGPU for hardware acceleration, reducing server costs and enhancing privacy.
Last checked: 2 hours ago View Status -
OperationalLora
Integrate local LLM with one line of code.
Lora provides an SDK for integrating a fine-tuned, mobile-optimized local Large Language Model (LLM) into applications with minimal setup, offering GPT-4o-mini level performance.
Last checked: 12 hours ago View Status -
OperationalLM Studio
Discover, download, and run local LLMs on your computer
LM Studio is a desktop application that allows users to run Large Language Models (LLMs) locally and offline, supporting various architectures including Llama, Mistral, Phi, Gemma, DeepSeek, and Qwen 2.5.
Last checked: 1 hour ago View Status