I tested GPT4all on Mac. After you installing the app, you can either download the models from its UI, or you can download them separately and put it into the relevant folder. Then, you can make changes to the settings if you want.
Preferably, check if sharing conversations with model authors and disable it.
Saving chats to disk takes a lot of disk space.
For macs with silicons, you can also use metal, but if you use more ram than you have, it will crash.
I’m not that knowledgeable on this kind of AI. Would a locally run model on GPT4All have lower quality of results than cloud based models like mixtral 8x7b with brave leo?
Ignoring speed differences due to hardware, how is the actual quality and content of responses?
That’s concerning. We wouldn’t want an offline model to send data home, as this nullify the point of running it locally. Is it opt-in with no nudging to enable it ?
Check out arena.lmsys.org. The short answers is most likely it won’t. The long answer is :
Explanation
The Mixtral 8x7b can be easily run on your computer, but more powerful model like Claude Instant can’t. although it doesn’t matter since they are closed source. Recently though, Meta released Llama 3. Its 8b version is more powerful than any of the model Brave hosts, including Claude Instant (2). The 70B model (can be run on a very high end pc) is only below Google, OpenAI, and Claude. The Llama models have a fairly permissive license.
I think we should definitely lists some resources in the recommendations list. I’m even thinking of making a guide, but I am not sure this is PG role.
About the criteria section, I do not feel comfortable knowing that chats are kept for 30 days.
I am becoming concerned that Duck AI only use OpenAI and Anthropic and those keeps your chat for one month. If you consider this category like Search Engines (Providers), then that’s normal, afterall Bing and Google probably keep all queries ever sent to them, and that includes StartPage and DuckDuckGo. So in the Privacy Guides category we warn that there is no “guarantee that those privacy policy are honored”. At the same time, this category is also very close to a Tools category, as you are basically just offloading the work to someone with a more powerful computer. If we consider this a tool, then we must GUARANTEE the effective privacy.
I’m also concerned that despite the warning added against sharing PII, people will do it no matter what, as chatting with an AI feels similar ro doing so with an human, and humans are social creatures that love sharing.
I will not include an app without testing it first. if you have tested it yourself, feel free to write a short description, including platform availability, ease of use, which models are available, check the privacy, etc.
Not available on any recommended Linux distributions is a big-no.
Also , this
I have already tested GPT for all on my system also olama on my system and now I am also testing olama on our cloud server which was given to me by @wojciechxtx
Talking about ollama
Linux macos windows(preview)
Installation is very easy just a curl script though i know this is a unsafe method still it works and this script is open and manual installation is also a option those want that
Project is also in dockerhub to run under contrainer. https://hub.docker.com/r/ollama/ollama
I have checked this can be run using podman which is a rootless docker like service and fully compatible with docker.
Project is on github
New models are coming everyday on my request phi3 medium just launched within 1day.
I tested Ollama, very easy to setup indeed, the only issue I had is I couldn’t modify Llama 3 system prompt, despite multiple attempts at copying the model and writing a new system prompt, it will just add my prompt but keep the (very long) Llama 3 system prompt.
@brivacy My strong suspicion is that Mozilla will dump this project. Sooner than we think. As with almost all of their side-projects (Mozilla Reps Program being last one; at least that I know of)