back to top
HomeSoftwareAI ToolsLibreChat: Top Open-Source ChatGPT Alternative for Self-Hosting AI Models Like GPT-OSS, LLaMA,...

LibreChat: Top Open-Source ChatGPT Alternative for Self-Hosting AI Models Like GPT-OSS, LLaMA, Mistral & More

- Advertisement -

File Information

FileDetails
NameLibreChat
VersionLatest (Stable)
Format.zip
LicenseMIT
PlatformWindows, macOS, Linux
UI TypeWeb UI (accessible via browser)
Size300MB (may vary by build)

Description

LibreChat is a game-changer in the world of AI chat interfaces. Designed with inspiration from OpenAI’s ChatGPT and supercharged with cutting-edge enhancements, LibreChat offers a modern, clean & highly customizable interface to run your own LLMs. Whether you’re a developer, researcher, or just someone who wants full control over their AI assistant experience. LibreChat gives you everything you need, without the need for third-party subscriptions or cloud lock-in.

Unlike other platforms, LibreChat is completely open source and community-powered. It supports a wide range of models — from cloud-based APIs like OpenAI, Claude, or Gemini, to fully local options like Ollama, LM Studio & even your own private GPT OSS instances.

The beauty of LibreChat lies in its flexibility. You can run it locally or deploy it on a private server, rebrand it entirely, build your own agents, integrate advanced tools, or simply use it as a personal assistant. With support for voice chat, multimodal inputs, image generation, markdown, conversation forking, file uploads, web search and much more, it’s truly one of the most powerful chat interfaces available today. It comes with the benefits of following:

  • Fully rebrandable: Customize logo, UI, name & links
  • Own your data: Nothing is sent to external servers
  • All-in-one hub: Use multiple models in one UI
  • Model flexibility: From OpenAI to local GGUF files
  • Multi-user support: Secure logins for teams or personal use

LibreChat is perfect for individuals or teams who want the power of AI, without the privacy compromises or vendor limitations.

Features of LibreChat

  • UI & UX Inspired by ChatGPT: Clean, modern, intuitive
  • AI Model Selection: Choose from OpenAI, Claude, Google, Vertex AI, Anthropic, LM Studio, Ollama & more
  • Custom Endpoints: Use your own OpenAI compatible API
  • File Upload & Analysis: Chat with PDFs, images, and more
  • Multilingual UI: 20+ languages supported
  • Speech Recognition & Audio Response: Hands-free interaction
  • Secure Auth: OAuth2, LDAP, email login, multi-user access
  • Agents & Tools: Custom assistants, tool integration
  • Image Generation: DALL-E, Stable Diffusion, GPT-Image
  • Reasoning UI: Built for advanced CoT models like DeepSeek
  • Import/Export Chats: Markdown, screenshots, JSON
  • Presets: Save and reuse prompts with custom models
  • Web Search Integration: Live internet access for your models
  • Full Open Source: Fork, extend, or contribute freely

Screenshots

System Requirements

PlatformRequirements
WindowsNode.js (v18+),MongoDB, npm , Git
macOSNode.js (v18+), MongoDB, npm, Git
LinuxNode.js (v18+), MongoDB, npm, Git

How to Install LibreChat

Before you start: Scroll to the bottom of this page & download the latest .zip file of LibreChat.

Step-by-Step Installation Guide

1. Extract the .zip File

After downloading, unzip the file to any location you prefer:
Example: Documents/LibreChat


2. Open Terminal / CMD

Navigate into the LibreChat folder:

cd LibreChat

3. Create Your .env File

Copy the sample environment file:

On Linux/macOS:

cp .env.example .env

On Windows:

copy .env.example .env

4. Configure MongoDB

Open the .env file and set the MONGO_URI to your own MongoDB connection string.

If you don’t have MongoDB installed, you can use MongoDB Atlas or install MongoDB locally.


5. Install All Dependencies

Install all required packages with:

npm ci

6. Build the Frontend

Compile the frontend UI:

npm run frontend

7. Start the Backend Server

Launch the backend server:

npm run backend

8. Open LibreChat in Browser

Go to:

http://localhost:3080/

LibreChat will now be accessible with GPT-OSS and other models if configured properly!


9. Updating LibreChat in the Future

To pull the latest changes and rebuild:

git pull
npm ci
npm run frontend
npm run backend

How to Add GPT-OSS or Other Local AI Models to LibreChat

LibreChat supports a variety of local models like:

  • GPT-OSS
  • LLaMA / LLaMA 3
  • Gemma
  • Mistral
  • OpenHermes
  • WizardCoder, and more…

Here’s how to integrate them:

librechat.yaml : Tell LibreChat how to connect to Ollama

- name: "Ollama"
apiKey: "ollama" # just a label; no actual key needed for local models
baseURL: http://localhost:11434/v1/ # "http://host.docker.internal:11434/v1/" if using Docker
models:
default: [
"llama3", # change to any model you want (like gpt-oss, mistral, gemma, etc.)
"mistral",
"gemma",
"phi3"
]
fetch: false # true = fetch available models dynamically from Ollama (optional)
titleConvo: true
titleModel: "llama3"
summarize: false
summaryModel: "llama3"
forcePrompt: false
modelDisplayLabel: "Ollama"

Key things:

  • baseURL: host.docker.internal is needed only if LibreChat is in Docker.
  • models.default: You can manually list the models you want, like "gpt-oss", "mistral", "gemma", etc.
  • fetch: true: If set, LibreChat will automatically fetch all models available in Ollama (if the name starts with ollama).

2. docker-compose.override.yaml : Mount your config and run Ollama

services:
api:
volumes:
- type: bind
source: ./librechat.yaml
target: /app/librechat.yaml

ollama:
image: ollama/ollama:latest
ports:
- "11434:11434"
volumes:
- ./ollama:/root/.ollama # where your models and data live
deploy:
resources:
reservations:
devices:
- driver: nvidia
capabilities: [compute, utility] # for GPU acceleration (optional)

Tip: If you’re not using Docker, you don’t need to worry about the docker-compose.override.yaml. Just make sure Ollama is running locally on localhost:11434.

  1. Restart LibreChat:
npm run backend
  1. Now go to Settings > Models and you’ll see your custom model ready to chat!

For More Information about model addition and configurations, Read Official Documentation of LibreChat


LibreChat: The Ultimate Self-Hosted Open Source AI Chat Platform

- Advertisment -
YOU MAY ALSO LIKE
Parallel Code – Run Multiple AI Coding Agents with Git Worktree Isolation

Parallel Code – Run Multiple AI Coding Agents with Git Worktree Isolation

0
Running multiple AI coding agents is powerful. It is also messy. Put them on the same branch and they overwrite each other. Split them across terminals and you forget which one is doing what. You can manually create feature branches and worktrees, but after the third task you start feeling like a part-time git administrator. Parallel Code handles that part for you. Create a task and the app: Creates a new branch from main Sets up a separate git worktree Symlinks node_modules and other ignored directories Launches the selected AI agent inside that worktree Each task lives in its own isolated environment. Five agents can work on five features in the same repo at the same
Linkora Open-Source Link Manager Desktop App

Linkora Desktop: Open-Source Link Manager App

0
Linkora Desktop is built for users who want full control over their bookmarks without relying on proprietary cloud services. All links are stored locally using SQLite, ensuring fast performance and privacy by default. You can create unlimited folders and subfolders, organize links with tags, and switch between multiple layout views for better visual management. If you need cross-device access, Linkora offers an optional self-hosted sync server, meaning you control where your data lives. With browser extension support and Android integration, Linkora makes saving and organizing links effortless across platforms.
Velo Fast & AI-Enhanced Desktop Email Client with Offline Support

Velo: Fast & AI-Enhanced Desktop Email Client with Offline Support

0
Velo is designed for speed, privacy, and efficiency. Your emails are stored locally in an SQLite database, so you stay in full control of your data without any external servers or hidden trackers. Its keyboard-first interface allows fast inbox management, while built-in AI features help you summarize threads, compose smart replies, and search naturally. If you want a fully offline client or prefer connecting to Gmail or IMAP/SMTP accounts, Velo adapts to your workflow.

Don’t miss any Tech Story

Subscribe To Firethering NewsLetter

You Can Unsubscribe Anytime! Read more in our privacy policy