Compare commits

...

12 Commits

Author SHA1 Message Date
Roy Han
568416ba17 add suffix 2024-07-16 16:51:27 -07:00
Roy Han
80cba42ab2 Update docs 2024-07-16 16:51:27 -07:00
royjhan
6477a7aca4 Merge branch 'royh-completions-docs' of https://github.com/ollama/ollama into royh-completions-docs 2024-07-16 16:51:11 -07:00
royjhan
51214ddef5 Update docs/openai.md 2024-07-16 16:34:31 -07:00
royjhan
b950d749a9 Update docs/openai.md 2024-07-16 16:34:31 -07:00
Roy Han
3702ed7532 token bug corrected 2024-07-16 16:34:31 -07:00
Roy Han
6266603b17 Update docs 2024-07-16 16:34:31 -07:00
royjhan
2644c4e682
Update docs/openai.md 2024-07-08 14:46:05 -07:00
royjhan
04cde43b2a
Update docs/openai.md 2024-07-08 14:44:16 -07:00
Roy Han
105e36765d token bug corrected 2024-07-03 15:03:54 -07:00
royjhan
fa7be5aab4
Merge branch 'main' into royh-completions-docs 2024-07-02 14:52:56 -07:00
Roy Han
02169f3e60 Update docs 2024-06-26 14:30:28 -07:00

View File

@ -27,6 +27,11 @@ chat_completion = client.chat.completions.create(
],
model='llama3',
)
completion = client.completions.create(
model="llama3",
prompt="Say this is a test"
)
```
### OpenAI JavaScript library
@ -45,6 +50,11 @@ const chatCompletion = await openai.chat.completions.create({
messages: [{ role: 'user', content: 'Say this is a test' }],
model: 'llama3',
})
const completion = await openai.completions.create({
model: "llama3",
prompt: "Say this is a test.",
})
```
### `curl`
@ -66,6 +76,12 @@ curl http://localhost:11434/v1/chat/completions \
]
}'
curl http://localhost:11434/v1/completions \
-H "Content-Type: application/json" \
-d '{
"model": "llama3",
"prompt": "Say this is a test"
}'
```
## Endpoints
@ -103,8 +119,71 @@ curl http://localhost:11434/v1/chat/completions \
- [ ] `user`
- [ ] `n`
### `/v1/completions`
#### Supported features
- [x] Completions
- [x] Streaming
- [x] JSON mode
- [x] Reproducible outputs
- [ ] Logprobs
#### Supported request fields
- [x] `model`
- [x] `prompt`
- [x] `frequency_penalty`
- [x] `presence_penalty`
- [x] `seed`
- [x] `stop`
- [x] `stream`
- [x] `temperature`
- [x] `top_p`
- [x] `max_tokens`
- [x] `suffix`
- [ ] `best_of`
- [ ] `echo`
- [ ] `logit_bias`
- [ ] `user`
- [ ] `n`
#### Notes
- `prompt` currently only accepts a string
### `/v1/completions`
#### Supported features
- [x] Completions
- [x] Streaming
- [x] JSON mode
- [x] Reproducible outputs
- [ ] Logprobs
#### Supported request fields
- [x] `model`
- [x] `prompt`
- [x] `frequency_penalty`
- [x] `presence_penalty`
- [x] `seed`
- [x] `stop`
- [x] `stream`
- [x] `temperature`
- [x] `top_p`
- [x] `max_tokens`
- [ ] `best_of`
- [ ] `echo`
- [ ] `suffix`
- [ ] `logit_bias`
- [ ] `user`
- [ ] `n`
#### Notes
- `prompt` currently only accepts a string
- `usage.prompt_tokens` will be 0 for completions where prompt evaluation is cached
## Models