On Fri, 16 Jan 2026 05:07:01 +0100,
Chris Cappuccio <[email protected]> wrote:
> 
> This is the Codex LLM agent from OpenAI.
> 
> It can perform actions to use tools within the filesystem to aid in a
> variety of actions, particularly coding-related activities.
> 
> The pipeline of llama.cpp + codex enables an entirely native OpenBSD
> based, AMDGPU accelerated LLM inference agent system.
> 
> This is one of the last versions of Codex that will support the "Chat
> Completions" API which is currently served through llama.cpp. (Future
> Codex versions will require "Responses" API support in llama.cpp.)

I see PR for llama.cpp to supports this API, so, probably it will be
released soon: https://github.com/ggml-org/llama.cpp/pull/18486

I haven't fully read your port, just take a vary fast look and one things
raise fast suggestion, I think this way it's cleaner:

DIST_TUPLE =    github  openai          codex           rust-v${V}      .
DIST_TUPLE +=   github  nornagon        crossterm       ${CROSSTERM_COMMIT}     
crossterm
DIST_TUPLE +=   github  nornagon        ratatui         ${RATATUI_COMMIT}       
ratatui

-- 
wbr, Kirill

Reply via email to