소스 검색

Link to HuggingFace Hub

Eric J. Wang 3 년 전
부모
커밋
6a8b163f3e
1개의 변경된 파일3개의 추가작업 그리고 3개의 파일을 삭제
  1. 3 3
      README.md

+ 3 - 3
README.md

@@ -7,9 +7,9 @@ We aim to provide an Instruct model of similar quality to `text-davinci-003` tha
 but extensions to the `13b`, `30b`, and `65b` models should be feasible with simple changes to the code.
 
 In addition to the training code, which runs within five hours on a single RTX 4090,
-we publish the resulting LoRA weights on the Huggingface model hub,
-along with a script for downloading and inference on the foundation model and LoRA.
-For fine-tuning LoRAs we use Huggingface's [PEFT](https://github.com/huggingface/peft)
+we publish a script for downloading and inference on the foundation model and LoRA,
+as well as the resulting [LoRA weights themselves](https://huggingface.co/tloen/alpaca-lora-7b/tree/main).
+To fine-tune cheaply and efficiently, we use Huggingface's [PEFT](https://github.com/huggingface/peft)
 as well as Tim Dettmers' [bitsandbytes](https://github.com/TimDettmers/bitsandbytes).
 
 Without hyperparameter tuning or validation-based checkpointing, the LoRA model produces outputs comparable to the Stanford Alpaca model, though possibly with more minor mistakes. (Please see the outputs included below.) Further tuning might be able to achieve better performance; I invite interested users to give it a try and report their results.