llama-factory || AutoDL平台 ||启动web界面

报错如下:

bash 复制代码
root@autodl-container-d83e478b47-3def8c49:~/LLaMA-Factory# llamafactory-cli webui
* Running on local URL:  http://0.0.0.0:7860

Could not create share link. Missing file: /root/miniconda3/lib/python3.10/site-packages/gradio/frpc_linux_amd64_v0.3. 

Please check your internet connection. This can happen if your antivirus software blocks the download of this file. You can install manually by following these steps: 

1. Download this file: https://cdn-media.huggingface.co/frpc-gradio-0.3/frpc_linux_amd64
2. Rename the downloaded file to: frpc_linux_amd64_v0.3
3. Move the file to this location: /root/miniconda3/lib/python3.10/site-packages/gradio

解决办法:

bash 复制代码
cp frpc_linux_amd64_v0.3 /root/miniconda3/lib/python3.10/site-packages/gradio
bash 复制代码
cd /root/miniconda3/lib/python3.10/site-packages/gradio

提高权限

bash 复制代码
 chmod +x frpc_linux_amd64_v0.3

重新执行

bash 复制代码
llamafactory-cli webui
bash 复制代码
 CUDA_VISIBLE_DEVICES=0 llamafactory-cli webchat   --model_name_or_path /root/autodl-fs/Qwen2.5-1.5B-Instruct   --template qwen

结果:

bash 复制代码
[INFO|2025-03-02 23:15:05] llamafactory.model.model_utils.attention:157 >> Using torch SDPA for faster training and inference.
[INFO|2025-03-02 23:15:05] llamafactory.model.loader:157 >> all params: 1,543,714,304
* Running on local URL:  http://0.0.0.0:7860
* Running on public URL: https://35d22b023607f1702a.gradio.live

This share link expires in 72 hours. For free permanent hosting and GPU upgrades, run `gradio deploy` from the terminal in the working directory to deploy to Hugging Face Spaces (https://huggingface.co/spaces)

下面的链接就是可访问的动态链接

相关推荐
liliangcsdn14 小时前
mac测试ollama llamaindex
数据仓库·人工智能·prompt·llama
茫茫人海一粒沙18 小时前
使用 LLaMA 3 8B 微调一个 Reward Model:从入门到实践
llama
liliangcsdn3 天前
mac llama_index agent算术式子计算示例
人工智能·python·macos·llama
许愿与你永世安宁4 天前
RAG(检索增强生成)里的文档管理
数据库·人工智能·gpt·oracle·llama·rag
许愿与你永世安宁7 天前
基于Llama的RAG 3种模型配置方法
人工智能·python·自然语言处理·json·github·llama·faiss
至善迎风7 天前
本地部署 Kimi K2 全指南(llama.cpp、vLLM、Docker 三法)
docker·容器·llama·kimi
阿斯卡码7 天前
安装 llama-cpp-python 的CPU和GPU方法
开发语言·python·llama
学不会就看8 天前
llama-factory快速开始
llama
NullPointerExpection11 天前
LLM大语言模型不适合统计算数,可以让大模型根据数据自己建表、插入数据、编写查询sql统计
数据库·人工智能·sql·算法·llm·llama·工作流
MUTA️11 天前
Llama系列:Llama1, Llama2,Llama3内容概述
llama