core-agent-ide/codex-rs/tui2
Devon Rifkin fe03320791
ollama: default to Responses API for built-ins (#8798)
This is an alternate PR to solving the same problem as
<https://github.com/openai/codex/pull/8227>.

In this PR, when Ollama is used via `--oss` (or via `model_provider =
"ollama"`), we default it to use the Responses format. At runtime, we do
an Ollama version check, and if the version is older than when Responses
support was added to Ollama, we print out a warning.

Because there's no way of configuring the wire api for a built-in
provider, we temporarily add a new `oss_provider`/`model_provider`
called `"ollama-chat"` that will force the chat format.

Once the `"chat"` format is fully removed (see
<https://github.com/openai/codex/discussions/7782>), `ollama-chat` can
be removed as well

---------

Co-authored-by: Eric Traut <etraut@openai.com>
Co-authored-by: Michael Bolin <mbolin@openai.com>
2026-01-13 09:51:41 -08:00
..
docs tui2: stop baking streaming wraps; reflow agent markdown (#8761) 2026-01-05 18:37:58 -08:00
frames feat(tui2): copy tui crate and normalize snapshots (#7833) 2025-12-10 22:53:46 +00:00
src ollama: default to Responses API for built-ins (#8798) 2026-01-13 09:51:41 -08:00
tests Sync tui2 with tui and keep dual-run glue (#7965) 2025-12-12 20:46:18 -08:00
BUILD.bazel feat: add support for building with Bazel (#8875) 2026-01-09 11:09:43 -08:00
Cargo.toml feat(tui2): transcript scrollbar (auto-hide + drag) (#8728) 2026-01-05 09:05:14 -08:00
prompt_for_init_command.md feat(tui2): copy tui crate and normalize snapshots (#7833) 2025-12-10 22:53:46 +00:00
styles.md Sync tui2 with tui and keep dual-run glue (#7965) 2025-12-12 20:46:18 -08:00
tooltips.txt Send message by default mid turn. queue messages by tab (#9077) 2026-01-12 23:06:35 -08:00