Jesse Gross 34a75102f7 prompt: Use a single token when estimating mllama context size
Currently we assume that images take 768 tokens of context size for
the purposes of clipping old messages that exceed the context window.
However, our mllama implementation stores the full image embedding
in a single token. As a result, there is significant waste of context
space.

Ideally, we would handle this more generically and have the
implementation report the number of tokens. However, at the moment
this would just result in a similar set of 'if' conditions in the
runner plus APIs to report it back. So for now, we just keep this
simple.
2024-11-05 10:11:50 -08:00
..
2024-07-26 14:14:48 -07:00
2024-08-01 17:06:06 -07:00
2024-08-15 10:29:14 -07:00
2024-10-18 16:12:35 -07:00
2024-08-28 09:32:57 -07:00
2024-08-28 09:32:57 -07:00
2024-10-16 17:45:00 -07:00