From: Kirill A. Korinsky Subject: Re: NEW: devel/codex To: Chris Cappuccio Cc: ports@openbsd.org Date: Wed, 21 Jan 2026 18:36:40 +0100 On Wed, 21 Jan 2026 18:18:29 +0100, Chris Cappuccio wrote: > > Kirill A. Korinsky [kirill@korins.ky] wrote: > > > > Here a version where I: > > 1. added runtime dependency on textproc/ripgrep > > 2. patched required current_exe() usage > > 3. added README with sample how it can be used against llama.cpp server > > > > I like this. I had done a similar README. In model_providers you should add: > > headers = {} > query_params = {} > > They might not break anything today but they are of no use to llama.cpp and > the future is less certain. They are designed for openai's interface. > Do you mean something like that? headers = {} query_params = {} model_provider = "local" model = "local" [model_providers.local] name = "llama-server" base_url = "http://127.0.0.1:8080/v1" wire_api = "chat" -- wbr, Kirill