kyang-06 commited on
Commit
13235a4
·
verified ·
1 Parent(s): 46f92e5

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +41 -0
README.md ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ language:
5
+ - en
6
+ tags:
7
+ - deepseek
8
+ - reasoning
9
+ - thinking
10
+ - Qwen 2.5
11
+ - Lora
12
+ - Lora adapter
13
+ - 128k context
14
+ - general usage
15
+ - problem solving
16
+ - brainstorming
17
+ - solve riddles
18
+ - mergekit
19
+ - adapter
20
+ - peft
21
+ - llama-cpp
22
+ - gguf-my-lora
23
+ base_model: DavidAU/DeepSeek-R1-Distill-Qwen-14B-LORA-256-RANK
24
+ pipeline_tag: text-generation
25
+ ---
26
+
27
+ # kyang-06/DeepSeek-R1-Distill-Qwen-14B-LORA-256-RANK-F32-GGUF
28
+ This LoRA adapter was converted to GGUF format from [`DavidAU/DeepSeek-R1-Distill-Qwen-14B-LORA-256-RANK`](https://huggingface.co/DavidAU/DeepSeek-R1-Distill-Qwen-14B-LORA-256-RANK) via the ggml.ai's [GGUF-my-lora](https://huggingface.co/spaces/ggml-org/gguf-my-lora) space.
29
+ Refer to the [original adapter repository](https://huggingface.co/DavidAU/DeepSeek-R1-Distill-Qwen-14B-LORA-256-RANK) for more details.
30
+
31
+ ## Use with llama.cpp
32
+
33
+ ```bash
34
+ # with cli
35
+ llama-cli -m base_model.gguf --lora DeepSeek-R1-Distill-Qwen-14B-LORA-256-RANK-f32.gguf (...other args)
36
+
37
+ # with server
38
+ llama-server -m base_model.gguf --lora DeepSeek-R1-Distill-Qwen-14B-LORA-256-RANK-f32.gguf (...other args)
39
+ ```
40
+
41
+ To know more about LoRA usage with llama.cpp server, refer to the [llama.cpp server documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/server/README.md).