MetaEditor, Open AI and ChatGPT - page 10

 
Edgar Akhmadeev #:

A local neural model can be run as a chat room or as a server. You connect to the server via API and JSON.

I recommend ollama or llama.cpp as the fastest and easiest (in c++, not python), and the most customised for GGUF.

LMStudio is the most convenient interface for chat.

Thanks for recommending LMStudio. I liked it.