Browse Source

Add Cache flag to api (#1642)

brucemacd/llama-mem-calc
K0IN 2 years ago
committed by GitHub
parent
commit
10da41d677
No known key found for this signature in database GPG Key ID: 4AEE18F83AFDEB23
  1. 2
      api/types.go
  2. 3
      docs/api.md
  3. 3
      llm/ext_server.go

2
api/types.go

@ -103,6 +103,7 @@ type Options struct {
MirostatEta float32 `json:"mirostat_eta,omitempty"`
PenalizeNewline bool `json:"penalize_newline,omitempty"`
Stop []string `json:"stop,omitempty"`
Cache bool `json:"cache,omitempty"`
}
// Runner options which must be set when the model is loaded into memory
@ -355,6 +356,7 @@ func DefaultOptions() Options {
MirostatEta: 0.1,
PenalizeNewline: true,
Seed: -1,
Cache: true,
Runner: Runner{
// options set when the model is loaded

3
docs/api.md

@ -306,7 +306,8 @@ curl http://localhost:11434/api/generate -d '{
"embedding_only": false,
"rope_frequency_base": 1.1,
"rope_frequency_scale": 0.8,
"num_thread": 8
"num_thread": 8,
"cache": true
}
}'
```

3
llm/ext_server.go

@ -234,7 +234,8 @@ func predict(llm extServer, opts api.Options, ctx context.Context, predict Predi
"penalize_nl": opts.PenalizeNewline,
"seed": opts.Seed,
"stop": opts.Stop,
"image_data": imageData,
"image_data": imageData,
"cache_prompt": opts.Cache,
}
if predict.Format == "json" {

Loading…
Cancel
Save