Cortex.cpp: model.yaml Format #1123
dan-menlo
started this conversation in
Architecture Specs
Replies: 3 comments 2 replies
-
Questions @nguyenhoangthuan99 @sangjanai @namchuai
See related issues above. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Here is the default setting for a gguf-model:
other information will be infer from the gguf metadata.
|
Beta Was this translation helpful? Give feedback.
0 replies
-
@nguyenhoangthuan99 Can we tame this complexity by defining the following:
name: Llama 3.1
model: llama3.1
version: 1
stop:
- <|end_of_text|>
- <|eot_id|>
- <|eom_id|>
top_p: 0.9
temperature: 0.6
frequency_penalty: 0
presence_penalty: 0
max_tokens: 8192
stream: true
ngl: 33
ctx_len: 8192
engine: cortex.llamacpp
prompt_template: |+
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>
{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
files:
- /Users/nicolezhu/cortex/models/llama3.1/model.gguf
id: llama3.1
created: 1722850990703
object: model
owned_by: '' i.e.
|
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Goal
GGUF
TBA
TensorRT-LLM
TBA
ONNX
TBA
Key Epics
Related
model.yaml
for Model Downloaded via URL jan#3558Beta Was this translation helpful? Give feedback.
All reactions