a bit more Vicuna context

This commit is contained in:
Carl Pearson
2023-05-08 13:01:06 -06:00
parent 31cb3ac354
commit f758e62fa3

View File

@@ -39,7 +39,11 @@ categories = []
+++
Vicuna-13B is an LLM based off of the LLaMa model.
Vicuna-13B is an LLM chatbot based off of the LLaMa model.
It's authors claim it achieves 90% of the quality of ChatGPT in a "fun and non-scientific" evaluation.
You can rent some cloud hardware and experiment with Vicuna 13B yourself!
Using CPU-only is slow (couple tokens per second), but fast enough for yout to get an idea of what to expect.
## Set up your Cloud Instance
@@ -125,7 +129,7 @@ I got an error about regenerating protobuf functions if I used protobuf > 3.20.
## Apply the vicuna deltas
FastChat has done the work of getting a little chat interface set up.
[FastChat](https://github.com/lm-sys/FastChat) has done the work of getting a little chat interface set up.
We'll use their package to download the deltas and apply them as well.
```