
This section describes how to deploy the Vicuna model for local inference using the FastChat framework.
You can obtain the Vicuna model weights following the official FastChat instructions:
The provided init.sh script initializes the Vicuna server components.
Below is the reference setup:
export CUDA_VISIBLE_DEVICES=4,5
python3 -m fastchat.serve.controller
python3 -m fastchat.serve.model_worker --model-path /data/xxx/vicuna-13b-v1.5
python3 -m fastchat.serve.openai_api_server --host localhost --port 8000
You can install all required Python packages using:
pip install -r requirements.txt
We currently provide the core implementation of the Causal Graph under src, including:
full_graph.py),swm_dataset.py, label.json), andtrain.py).