this post was submitted on 09 Nov 2023
3 points (100.0% liked)

Self-Hosted Main

21 readers
1 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

For Example

We welcome posts that include suggestions for good self-hosted alternatives to popular online services, how they are better, or how they give back control of your data. Also include hints and tips for less technical readers.

Useful Lists

founded 1 year ago
MODERATORS
 

Hello ! I am looking for a self-hosted or self-hostable nlp (like llama or chatgpt) but very very small one that could work with as low as ~300 mb of ram. It needs to have an API. I plan to integrate it into my dashboard project you might know, homarr. I'd like to make some kind of assistant to directly help within the app by using its integrations capabilities.

The tool needs to be self-hosted so that users won't leak the queries to anyone. A freemium service that you can either self-host or pay for would also work.

It does not need to have a huge knowledge base (doesn't need to know a good lobster recipe) , just to be able understand basic language inputs and in turn I will make it communicate with the key parts of the app

I apologize if this is not worded properly as I am fairly new to the world of LLMs.

you are viewing a single comment's thread
view the rest of the comments
[–] the9engines@alien.top 1 points 1 year ago (2 children)

ChatGPT is so far ahead and so advanced that:

- No model is even close to its quality

- Even if it it was released to the public, you would need so beefy machines to run it it makes no sense

We got to wait for some kind of breakthrough that would allow running high quality open source models locally.

Considering the cost of hosting anything, even if it was host on a PC at your place, the electricity bill alone would be higher than the ChatGPT API cost.

[–] grahamsz@alien.top 1 points 1 year ago

I can run VMWare's Open LLama 7B v2 Open Instruct on my laptop comfortably (though I have 64GB ram and 16GB VRAM) and my sense is that's it's probably somewhere between GPT2 and GPT3 in inference quality. It is, however, very slow. Even with my comparatively strong hardware, it's slow enough that I wouldn't want to use it in an interactive context (though it may be useful for background processing)

load more comments (1 replies)