Pārlūkot izejas kodu

Update README.md

Eric J. Wang 3 gadi atpakaļ
vecāks
revīzija
8dc0f614c6
1 mainītis faili ar 7 papildinājumiem un 4 dzēšanām
  1. 7 4
      README.md

+ 7 - 4
README.md

@@ -1,6 +1,6 @@
 ## 🦙🌲🤏 Alpaca-LoRA: Low-Rank LLaMA Instruct-Tuning
 
-**Try the pretrained model out on Colab [here](https://colab.research.google.com/drive/1eWAmesrW99p7e1nah5bipn0zikMb8XYC)!** _If you have problems with short outputs or very long outputs, please redownload the weights (`force_download=True`) and pull the latest version of the code._
+**Try the pretrained model out on Colab [here](https://colab.research.google.com/drive/1eWAmesrW99p7e1nah5bipn0zikMb8XYC)!**
 
 This repository contains code for reproducing the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) results using [low-rank adaptation (LoRA)](https://arxiv.org/pdf/2106.09685.pdf).
 We provide an Instruct model of similar quality to `text-davinci-003` that can run [on a Raspberry Pi](https://twitter.com/miolini/status/1634982361757790209) (for research),
@@ -14,6 +14,8 @@ as well as Tim Dettmers' [bitsandbytes](https://github.com/TimDettmers/bitsandby
 
 Without hyperparameter tuning or validation-based checkpointing, the LoRA model produces outputs comparable to the Stanford Alpaca model. (Please see the outputs included below.) Further tuning might be able to achieve better performance; I invite interested users to give it a try and report their results.
 
+For discussion and support, users have created a dedicated Discord server [here](https://discord.gg/prbq284xX5).
+
 ### Setup
 
 1. Install dependencies
@@ -38,9 +40,10 @@ PRs adapting this code to support larger models are always welcome.
 ### Checkpoint export (`export_*_checkpoint.py`)
 
 These files contain scripts that merge the LoRA weights back into the base model
-for export to Hugging Face format and to PyTorch `state_dicts`,
-which should help users who want to export LlamaModel-shaped weights or
-use the model with projects like [llama.cpp](https://github.com/ggerganov/llama.cpp).
+for export to Hugging Face format and to PyTorch `state_dicts`.
+They should help users
+who want to run inference in projects like [llama.cpp](https://github.com/ggerganov/llama.cpp)
+or [alpaca.cpp](https://github.com/antimatter15/alpaca.cpp).
 
 ### Dataset