项目概述
Web UI for working with large language models
项目地址
https://github.com/sedwards2009/llm-multitool
项目页面预览
关键指标
- Stars:38
- 主要语言:Go
- License:MIT License
- 最近更新:2024-06-13T21:01:56Z
- 默认分支:main
本站高速下载(国内可用)
- 源码压缩包下载:点击下载(本站镜像)
- SHA256:300b218964e290f91df659f58d93c2270bbbad9f7aa09941d6e1210a0482ed12
安装部署要点(README 精选)
OpenAI Configuration
The following example backend.yaml file shows how to connect to OpenAI’s ChatGPT model. You need to generate your own API token on OpenAI’s website to use in the file.
- name: OpenAI
api_token: "sk-FaKeOpEnAiToKeN7Nll3FAKzZET3BlbkFJLz8Oume19ZeAjGh3rabc"
models:
- gpt-3.5-turbo
- gpt-4
The name field can be any name you like, but it is best to keep it short.
api_token holds the value of the token you generated at OpenAI.
If you don’t want to copy your token directly into your configuration file, you can omit the api_token file and replace it with api_token_from with a value naming a text file from which to read the read token from. The file path is relative to the backend.yaml file.
models is a list of model to permit. OpenAI have many different models and varieties, but only a handful of the the LLMs are useful for use with llm-multitool.
Running
If you have built llm-multitool from source then the executable will be in backend/llm-multitool and you should have written a minial backend.yaml file. Start up llm-multitool with:
backend/llm-multitool -c backend.yaml
This will start the server and it will listen on address 127.0.0.1 port 5050 by default.
Open your browser on http://127.0.0.1:5050 to use the llm-multitool UI.
常用命令(从 README 提取)
(未提取到命令块)
通用部署说明(适用于大多数项目)
- 下载源码并阅读 README
- 安装依赖(pip/npm/yarn 等)
- 配置环境变量(API Key、模型路径、数据库等)
- 启动服务并测试访问
- 上线建议:Nginx 反代 + HTTPS + 进程守护(systemd / pm2)
免责声明与版权说明
本文仅做开源项目整理与教程索引,源码版权归原作者所有,请遵循对应 License 合规使用。








暂无评论内容