2025-09-11 16:50:05 +08:00
# AnythingLLM
2025-04-26 12:03:23 +08:00
2025-09-11 16:50:05 +08:00
[AnythingLLM ](https://github.com/Mintplex-Labs/anything-llm ) is a full-stack application that enables you to turn any document, resource, or piece of content into context that any LLM can use as references during chatting.
2025-04-26 12:03:23 +08:00
It allows you to deploy a large language model (LLM) server with vLLM as the backend, which exposes OpenAI-compatible endpoints.
## Prerequisites
2025-09-11 16:50:05 +08:00
Set up the vLLM environment:
```bash
pip install vllm
```
2025-04-26 12:03:23 +08:00
## Deploy
2025-09-11 16:50:05 +08:00
1. Start the vLLM server with a supported chat-completion model, for example:
2025-04-26 12:03:23 +08:00
2025-09-11 16:50:05 +08:00
```bash
vllm serve Qwen/Qwen1.5-32B-Chat-AWQ --max-model-len 4096
```
1. Download and install [AnythingLLM Desktop ](https://anythingllm.com/desktop ).
1. Configure the AI provider:
- At the bottom, click the 🔧 wrench icon -> **Open settings ** -> **AI Providers ** -> **LLM ** .
- Enter the following values:
- LLM Provider: Generic OpenAI
- Base URL: `http://{vllm server host}:{vllm server port}/v1`
- Chat Model Name: `Qwen/Qwen1.5-32B-Chat-AWQ`
2025-04-26 12:03:23 +08:00
2025-09-11 16:50:05 +08:00

2025-04-26 12:03:23 +08:00
2025-09-11 16:50:05 +08:00
1. Create a workspace:
2025-04-26 12:03:23 +08:00
2025-09-11 16:50:05 +08:00
1. At the bottom, click the ↺ back icon and back to workspaces.
1. Create a workspace (e.g., `vllm` ) and start chatting.
2025-04-26 12:03:23 +08:00
2025-09-11 16:50:05 +08:00

2025-04-26 12:03:23 +08:00
2025-09-11 16:50:05 +08:00
1. Add a document.
2025-04-26 12:03:23 +08:00
2025-09-11 16:50:05 +08:00
1. Click the 📎 attachment icon.
1. Upload a document.
1. Select and move the document into your workspace.
1. Save and embed it.
2025-04-26 12:03:23 +08:00
2025-09-11 16:50:05 +08:00

2025-04-26 12:03:23 +08:00
2025-09-11 16:50:05 +08:00
1. Chat using your document as context.
2025-04-26 12:03:23 +08:00
2025-09-11 16:50:05 +08:00
