所使用的模型是`Llama-3-Taiwan-8B-Instruct-DPO`的量化版本。更多细节请访问网站(https://hugging-face.cn/yentinglin/Llama-3-Taiwan-8B-Instruct-DPO)
8B
38 Pulls 更新于6周前
更新于6周前
6周前
8158dc01cc32 · 16GB
model
架构llama
·
参数8.03B
·
量化F16
16GB
模板
"{{ if .System }}<|start_header_id|>system<|end_header_id|> {{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|> {{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|> {{ .Response }}<|eot_id|>""
257B
params
{"num_ctx":8192,"stop":["<|start_header_id|>","<|end_header_id|>","<|end_of_text|>","<|eot_id|>","<|reserved_special_token"]}
171B
README
无README