Update README.md
Browse files
README.md
CHANGED
|
@@ -7,6 +7,7 @@ tags:
|
|
| 7 |
- oceangpt
|
| 8 |
language:
|
| 9 |
- en
|
|
|
|
| 10 |
datasets:
|
| 11 |
- zjunlp/OceanInstruct
|
| 12 |
---
|
|
@@ -20,7 +21,7 @@ datasets:
|
|
| 20 |
<a href="https://github.com/zjunlp/OceanGPT">Project</a> •
|
| 21 |
<a href="https://arxiv.org/abs/2310.02031">Paper</a> •
|
| 22 |
<a href="https://huggingface.co/collections/zjunlp/oceangpt-664cc106358fdd9f09aa5157">Models</a> •
|
| 23 |
-
<a href="http://oceangpt.zjukg.cn
|
| 24 |
<a href="#quickstart">Quickstart</a> •
|
| 25 |
<a href="#citation">Citation</a>
|
| 26 |
</p>
|
|
@@ -88,13 +89,20 @@ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
|
|
| 88 |
| OceanGPT-14B-v0.1 (based on Qwen) | <a href="https://huggingface.co/zjunlp/OceanGPT-14B-v0.1" target="_blank">14B</a> | <a href="https://wisemodel.cn/models/zjunlp/OceanGPT-14B-v0.1" target="_blank">14B</a> | <a href="https://modelscope.cn/models/ZJUNLP/OceanGPT-14B-v0.1" target="_blank">14B</a> |
|
| 89 |
| OceanGPT-7B-v0.2 (based on Qwen) | <a href="https://huggingface.co/zjunlp/OceanGPT-7b-v0.2" target="_blank">7B</a> | <a href="https://wisemodel.cn/models/zjunlp/OceanGPT-7b-v0.2" target="_blank">7B</a> | <a href="https://modelscope.cn/models/ZJUNLP/OceanGPT-7b-v0.2" target="_blank">7B</a> |
|
| 90 |
| OceanGPT-2B-v0.1 (based on MiniCPM) | <a href="https://huggingface.co/zjunlp/OceanGPT-2B-v0.1" target="_blank">2B</a> | <a href="https://wisemodel.cn/models/zjunlp/OceanGPT-2b-v0.1" target="_blank">2B</a> | <a href="https://modelscope.cn/models/ZJUNLP/OceanGPT-2B-v0.1" target="_blank">2B</a> |
|
| 91 |
-
|
| 92 |
-
---
|
| 93 |
|
| 94 |
## 🌻Acknowledgement
|
| 95 |
|
| 96 |
OceanGPT is trained based on the open-sourced large language models including [Qwen](https://huggingface.co/Qwen), [MiniCPM](https://huggingface.co/collections/openbmb/minicpm-2b-65d48bf958302b9fd25b698f), [LLaMA](https://huggingface.co/meta-llama). Thanks for their great contributions!
|
| 97 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 98 |
|
| 99 |
### 🚩Citation
|
| 100 |
|
|
|
|
| 7 |
- oceangpt
|
| 8 |
language:
|
| 9 |
- en
|
| 10 |
+
- zh
|
| 11 |
datasets:
|
| 12 |
- zjunlp/OceanInstruct
|
| 13 |
---
|
|
|
|
| 21 |
<a href="https://github.com/zjunlp/OceanGPT">Project</a> •
|
| 22 |
<a href="https://arxiv.org/abs/2310.02031">Paper</a> •
|
| 23 |
<a href="https://huggingface.co/collections/zjunlp/oceangpt-664cc106358fdd9f09aa5157">Models</a> •
|
| 24 |
+
<a href="http://oceangpt.zjukg.cn/">Web</a> •
|
| 25 |
<a href="#quickstart">Quickstart</a> •
|
| 26 |
<a href="#citation">Citation</a>
|
| 27 |
</p>
|
|
|
|
| 89 |
| OceanGPT-14B-v0.1 (based on Qwen) | <a href="https://huggingface.co/zjunlp/OceanGPT-14B-v0.1" target="_blank">14B</a> | <a href="https://wisemodel.cn/models/zjunlp/OceanGPT-14B-v0.1" target="_blank">14B</a> | <a href="https://modelscope.cn/models/ZJUNLP/OceanGPT-14B-v0.1" target="_blank">14B</a> |
|
| 90 |
| OceanGPT-7B-v0.2 (based on Qwen) | <a href="https://huggingface.co/zjunlp/OceanGPT-7b-v0.2" target="_blank">7B</a> | <a href="https://wisemodel.cn/models/zjunlp/OceanGPT-7b-v0.2" target="_blank">7B</a> | <a href="https://modelscope.cn/models/ZJUNLP/OceanGPT-7b-v0.2" target="_blank">7B</a> |
|
| 91 |
| OceanGPT-2B-v0.1 (based on MiniCPM) | <a href="https://huggingface.co/zjunlp/OceanGPT-2B-v0.1" target="_blank">2B</a> | <a href="https://wisemodel.cn/models/zjunlp/OceanGPT-2b-v0.1" target="_blank">2B</a> | <a href="https://modelscope.cn/models/ZJUNLP/OceanGPT-2B-v0.1" target="_blank">2B</a> |
|
| 92 |
+
|
|
|
|
| 93 |
|
| 94 |
## 🌻Acknowledgement
|
| 95 |
|
| 96 |
OceanGPT is trained based on the open-sourced large language models including [Qwen](https://huggingface.co/Qwen), [MiniCPM](https://huggingface.co/collections/openbmb/minicpm-2b-65d48bf958302b9fd25b698f), [LLaMA](https://huggingface.co/meta-llama). Thanks for their great contributions!
|
| 97 |
|
| 98 |
+
## Limitations
|
| 99 |
+
|
| 100 |
+
- The model may have hallucination issues.
|
| 101 |
+
|
| 102 |
+
- We did not optimize the identity and the model may generate identity information similar to that of Qwen/MiniCPM/LLaMA/GPT series models.
|
| 103 |
+
|
| 104 |
+
- The model's output is influenced by prompt tokens, which may result in inconsistent results across multiple attempts.
|
| 105 |
+
|
| 106 |
|
| 107 |
### 🚩Citation
|
| 108 |
|