Quoted:
Message:
Plus, some people run local inference servers, such as LocalAI, which also run OAI API. Using Ollama only forces doubling up your installs, as Ollama manages models its own way.
Timestamp:
2025-03-19T13:33:24.602000+00:00
Attachment:
Discord Message ID:
1351911440025391137