Open-source LLM development tools - AI tools

  • Laminar
    Laminar The AI engineering platform for LLM products

    Laminar is an open-source platform that enables developers to trace, evaluate, label, and analyze Large Language Model (LLM) applications with minimal code integration.

    • Freemium
    • From 25$
  • Hegel AI
    Hegel AI Developer Platform for Large Language Model (LLM) Applications

    Hegel AI provides a developer platform for building, monitoring, and improving large language model (LLM) applications, featuring tools for experimentation, evaluation, and feedback integration.

    • Contact for Pricing
  • Flowise
    Flowise Build LLM Apps Easily - Open Source Low-Code Tool for LLM Orchestration

    Flowise is an open-source low-code platform that enables developers to build customized LLM orchestration flows and AI agents through a drag-and-drop interface.

    • Freemium
    • From 35$
  • Langfuse
    Langfuse Open Source LLM Engineering Platform

    Langfuse provides an open-source platform for tracing, evaluating, and managing prompts to debug and improve LLM applications.

    • Freemium
    • From 59$
  • Missing Studio
    Missing Studio An open-source AI studio for rapid development and robust deployment of production-ready generative AI.

    Missing Studio is an open-source AI platform designed for developers to build and deploy generative AI applications. It offers tools for managing LLMs, optimizing performance, and ensuring reliability.

    • Free
  • PromptsLabs
    PromptsLabs A Library of Prompts for Testing LLMs

    PromptsLabs is a community-driven platform providing copy-paste prompts to test the performance of new LLMs. Explore and contribute to a growing collection of prompts.

    • Free
  • phoenix.arize.com
    phoenix.arize.com Open-source LLM tracing and evaluation

    Phoenix accelerates AI development with powerful insights, allowing seamless evaluation, experimentation, and optimization of AI applications in real time.

    • Freemium
  • BenchLLM
    BenchLLM The best way to evaluate LLM-powered apps

    BenchLLM is a tool for evaluating LLM-powered applications. It allows users to build test suites, generate quality reports, and choose between automated, interactive, or custom evaluation strategies.

    • Other
  • Agenta
    Agenta End-to-End LLM Engineering Platform

    Agenta is an LLM engineering platform offering tools for prompt engineering, versioning, evaluation, and observability in a single, collaborative environment.

    • Freemium
    • From 49$
  • EleutherAI
    EleutherAI Empowering Open-Source Artificial Intelligence Research

    EleutherAI is a research institute focused on advancing and democratizing open-source AI, particularly in language modeling, interpretability, and alignment. They train, release, and evaluate powerful open-source LLMs.

    • Free
  • LM Studio
    LM Studio Discover, download, and run local LLMs on your computer

    LM Studio is a desktop application that allows users to run Large Language Models (LLMs) locally and offline, supporting various architectures including Llama, Mistral, Phi, Gemma, DeepSeek, and Qwen 2.5.

    • Free
  • LMQL
    LMQL A programming language for LLMs.

    LMQL is a programming language designed for large language models, offering robust and modular prompting with types, templates, and constraints.

    • Free
  • docs.litellm.ai
    docs.litellm.ai Unified Interface for Accessing 100+ LLMs

    LiteLLM provides a simplified and standardized way to interact with over 100 large language models (LLMs) using a consistent OpenAI-compatible input/output format.

    • Free
  • Rig
    Rig Build Modular and Scalable LLM Applications in Rust

    Rig is a Rust-based framework for building modular and scalable LLM applications. It offers a unified LLM interface, Rust-powered performance, and advanced AI workflow abstractions.

    • Free
  • Keywords AI
    Keywords AI LLM monitoring for AI startups

    Keywords AI is a comprehensive developer platform for LLM applications, offering monitoring, debugging, and deployment tools. It serves as a Datadog-like solution specifically designed for LLM applications.

    • Freemium
    • From 7$
  • LLMStack
    LLMStack Open-source platform to build AI Agents, workflows and applications with your data

    LLMStack is an open-source development platform that enables users to build AI agents, workflows, and applications by integrating various model providers and custom data sources.

    • Other
  • aider
    aider AI Pair Programming in Your Terminal

    Aider is a command-line tool that enables pair programming with LLMs to edit code in your local git repository. It supports various LLMs and offers top-tier performance on software engineering benchmarks.

    • Free
  • Gentrace
    Gentrace Intuitive evals for intelligent applications

    Gentrace is an LLM evaluation platform designed for AI teams to test and automate evaluations of generative AI products and agents. It facilitates collaborative development and ensures high-quality LLM applications.

    • Usage Based
  • Unify
    Unify Build AI Your Way

    Unify provides tools to build, test, and optimize LLM pipelines with custom interfaces and a unified API for accessing all models across providers.

    • Freemium
    • From 40$
  • ModelBench
    ModelBench No-Code LLM Evaluations

    ModelBench enables teams to rapidly deploy AI solutions with no-code LLM evaluations. It allows users to compare over 180 models, design and benchmark prompts, and trace LLM runs, accelerating AI development.

    • Free Trial
    • From 49$
  • LMSYS Org
    LMSYS Org Developing open, accessible, and scalable large model systems

    LMSYS Org is a leading organization dedicated to developing and evaluating large language models and systems, offering open-source tools and frameworks for AI research and implementation.

    • Free
  • docs.mistral.ai
    docs.mistral.ai Building the Best Open Source Models

    Mistral AI is a research lab developing state-of-the-art open-source and commercial Large Language Models (LLMs) for developers and enterprises.

    • Other
  • Neural Magic
    Neural Magic Deploy Open-Source LLMs to Production with Maximum Efficiency

    Neural Magic offers enterprise inference server solutions to streamline AI model deployment, maximizing computational efficiency and reducing costs on both GPU and CPU infrastructure.

    • Contact for Pricing
  • Langbase
    Langbase The most powerful serverless platform for building AI products

    Langbase is a serverless AI developer platform that enables developers to build, deploy, and manage AI products with composable infrastructure, featuring BaseAI - the first Web AI Framework.

    • Freemium
    • From 20$
  • Rubra
    Rubra Open-weight, tool-calling LLMs

    Rubra provides a collection of open-weight large language models (LLMs) enhanced with tool-calling capabilities, ideal for building AI agents.

    • Free
  • PromptMage
    PromptMage A Python framework for simplified LLM-based application development

    PromptMage is a Python framework that streamlines the development of complex, multi-step applications powered by Large Language Models (LLMs), offering version control, testing capabilities, and automated API generation.

    • Other
  • Rhesis AI
    Rhesis AI Open-source test generation SDK for LLM applications

    Rhesis AI offers an open-source SDK to generate comprehensive, context-specific test sets for LLM applications, enhancing AI evaluation, reliability, and compliance.

    • Freemium
  • OpenRouter
    OpenRouter A unified interface for LLMs

    OpenRouter provides a unified interface for accessing and comparing various Large Language Models (LLMs), offering users the ability to find optimal models and pricing for their specific prompts.

    • Usage Based
  • Inductor
    Inductor Streamline Production-Ready LLM Applications

    Inductor enables developers to rapidly prototype, evaluate, and improve LLM applications, ensuring high-quality app delivery.

    • Freemium
  • W&B Weave
    W&B Weave A Framework for Developing and Deploying LLM-Based Applications

    Weights & Biases (W&B) Weave is a comprehensive framework designed for tracking, experimenting with, evaluating, deploying, and enhancing LLM-based applications.

    • Other
  • Didn't find tool you were looking for?

    Be as detailed as possible for better results
    EliteAi.tools logo

    Elite AI Tools

    EliteAi.tools is the premier AI tools directory, exclusively featuring high-quality, useful, and thoroughly tested tools. Discover the perfect AI tool for your task using our AI-powered search engine.

    Subscribe to our newsletter

    Subscribe to our weekly newsletter and stay updated with the latest high-quality AI tools delivered straight to your inbox.

    © 2025 EliteAi.tools. All Rights Reserved.