From f758e62fa3eb936a42357dbc4592c53edcc6b5fa Mon Sep 17 00:00:00 2001 From: Carl Pearson Date: Mon, 8 May 2023 13:01:06 -0600 Subject: [PATCH] a bit more Vicuna context --- content/post/20230506_vicuna/index.md | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) diff --git a/content/post/20230506_vicuna/index.md b/content/post/20230506_vicuna/index.md index 1b00a82..9f6fc61 100644 --- a/content/post/20230506_vicuna/index.md +++ b/content/post/20230506_vicuna/index.md @@ -39,7 +39,11 @@ categories = [] +++ -Vicuna-13B is an LLM based off of the LLaMa model. +Vicuna-13B is an LLM chatbot based off of the LLaMa model. +It's authors claim it achieves 90% of the quality of ChatGPT in a "fun and non-scientific" evaluation. + +You can rent some cloud hardware and experiment with Vicuna 13B yourself! +Using CPU-only is slow (couple tokens per second), but fast enough for yout to get an idea of what to expect. ## Set up your Cloud Instance @@ -125,7 +129,7 @@ I got an error about regenerating protobuf functions if I used protobuf > 3.20. ## Apply the vicuna deltas -FastChat has done the work of getting a little chat interface set up. +[FastChat](https://github.com/lm-sys/FastChat) has done the work of getting a little chat interface set up. We'll use their package to download the deltas and apply them as well. ```