TheOtherJake

joined 1 year ago
[–] TheOtherJake 27 points 1 year ago* (last edited 1 year ago) (1 children)

I'm sure they will eventually try to force ID's because it would be profitable for criminal data theft ads stalkers. This is all about corrupt money and exploitation. Billionaires are worthless parasites that have no right to exist in a Democratic system. Fuck the US fascist oligarchy party.

[–] TheOtherJake 8 points 1 year ago

Fire fox, Fire fox;

Fuck you Google;

We're throwing rocks.

Alpha bet, Alpha bet;

Farming data is,

stalking/theft.

[–] TheOtherJake 21 points 1 year ago

Oobabooga is the main GUI used to interact with models.

https://github.com/oobabooga/text-generation-webui

FYI, you need to find checkpoint models. In the available chat models space, naming can be ambiguous for a few reasons I'm not going to ramble about here. The main source of models is Hugging Face. Start with this model (or get the censored version):

https://huggingface.co/TheBloke/llama2_7b_chat_uncensored-GGML

First, let's break down the title.

  • This is a model based in Meta's Llama2.
  • This is not "FOSS" in the GPL/MIT type of context. This model has a license that is quite broad in scope with the key point stipulating it can not be used commercially for apps that have more than 700 million users.
  • Next, it was quantized by a popular user going by "The Bloke." I have no idea who this is IRL but I imagine this is a pseudonym or corporate alias given how much content is uploaded by this account on HF.
  • This model is based on a 7 Billion parameter dataset, and is fine tuned for chat applications.
  • This is uncensored meaning it will respond to most inputs as best it can. It can get NSFW, or talk about almost anything. In practice there are still some minor biases that are likely just over arching morality inherent to the datasets used, or it might be coded somewhere obscure.
  • Last part of the title is that this is a GGML model. This means it can run on CPU or GPU or a split between the two.

As for options on the landing page or "model card"

  • you need to get one of the older style models that have "q(numb)" as the quantization type. Do not get the ones that say "qK" as these won't work with the llama.cpp file you will get with Oobabooga.
  • look at the guide at the bottom of the model card where it tells you how much ram you need for each quantization type. If you have a Nvidia GPU with the CUDA API, enabling GPU layers makes the model run faster, and with quite a bit less system memory from what is stated on the model card.

The 7B models are about like having a conversation with your average teenager. Asking technical questions yielded around 50% accuracy in my experience. A 13B model got around 80% accuracy. The 30B WizardLM is around 90-95%. I'm still working on trying to get a 70B running on my computer. A lot of the larger models require compiling tools from source. They won't work directly with Oobabooga.

[–] TheOtherJake 12 points 1 year ago

It's no freaking mystery anywhere. Kids are too damn expensive because just living is too damn expensive. The real fix is massive land reform that absolutely murders the real estate bubble with a nuclear bomb. Regulate the availability of funds directly to the minimum wage. You work, you live a decent life with a good balance. Build dense housing with tight local communities and perfect transportation so we're always in contact with people in our communities. Babies will be popping up like weeds.

An article about one of the poorest European countries is not really relevant. They don't have the same zoning stagnation nonsense that makes housing unaffordable. The stupid incentives that exploded home loan amounts combined with 100 years without zoning reforms are the problem.

[–] TheOtherJake 1 points 1 year ago

Have you seen the great gatspy with Wizard too? That's what always comes up when mine goes too far. I'm working on compiling llama.cpp from source today. I think that's all I need to be able to use some of the other models like Llama2-70B derivatives.

The code for llama.cpp is only an 850 line python file (not exactly sure how python=CPP yet but YOLO I guess, I just started reading the code from a phone last night). This file is where all of the prompt magic happens. I think all of the easy checkpoint model stuff that works in Oobabooga uses python-llama-cpp from pip. That hasn't had any github repo updates in 3 months, so it doesn't work with a lot of newer and larger models. I'm not super proficient with Python. It is one of the things I had hoped to use AI to help me learn better, but I can read and usually modify someone else's code to some extent. It looks like a lot of the functionality (likely) built into the more complex chat systems like Tavern AI are just mixing the chat, notebook, and instruct prompt techniques into one 'context injection' (-if that term makes any sense).

The most information I have seen someone work with independently offline was using langchain with a 300 page book. So I know at least that much is possible. I have also come across a few examples of people using langchain with up to 3 PDF files at the same time. There is also the MPT model with up to 32k context tokens but it looks like it needs server machine ram in the hundreds of GB to function.

I'm having trouble with distrobox/conda/nvidia on Fedora Workstation. I think I may start over with Nix soon, or I am going to need to look into proxmox, virtualization or go back to an immutable base to ensure I can fall back effectively. I simply can't track down where some dependencies are getting stashed and I only have 6 distrobox containers so far. I'm only barely knowledgeable enough in Linux to manage something like this well enough for it to function. - suggestions welcome

[–] TheOtherJake 13 points 1 year ago

Pee-wee, Mr. Rogers, Sesame Street, Bozo, and Reading Rainbow are some of my fondest memories; when we were all Pee-wees, and he was Herman.

[–] TheOtherJake 2 points 1 year ago

WizardLM 30B at 4 bits with the GGML version on Oobabooga runs almost as fast as Llama2 7B on just the GPU. I set it up with 10 threads on the CPU and ~20 layers on the GPU. That leaves plenty of room for a 4096 context with a batch size of 2048. I can even run a 2GB Stable Diffusion model at the same time with my 3080's 16GBV.

Have you tried any of the larger models? I just ordered 64GB of ram. I also got kobold mostly working. I hope to use it to try Falcon 40. I really want to try a 70B model at 2-4 bit and see how its accuracy is.

[–] TheOtherJake 2 points 1 year ago

Fedora workstation. Had been on Silverblue for years, but got a machine with Nvidia and didn't want the extra headaches of SB

[–] TheOtherJake 1 points 1 year ago

This is how I use the internet.

[–] TheOtherJake 2 points 1 year ago* (last edited 1 year ago) (2 children)

Cookies are not needed. They are shifting the security onto the user. Secure the information on the server just like any other business. Offloading onto the client is wrong. It leads to ambiguity and abuses. Visiting a store and a business on the internet are no different. My presence gives no right to my person, searches, or tracking in the location or outside of it. Intentions are worthless. The only thing that matters is what is possible and practiced. Every loophole is exploited and should be mitigated. The data storage and coding practices must change.

[–] TheOtherJake 11 points 1 year ago (6 children)

Nah, it should be the default state of affairs. Data mining is stalking and theft. It centers around very poor logic and decisions.

Things like browser cookies are criminal garbage. Storing anything on a user's computer is stalking. Draw the parallel here; if you want to shop in any local store, I want you to first tell me everything you are wearing and carrying in a way that I can tell every possible detail about it, tell where you came from before you visited this store, where you are going next. They also want to know everything you looked at, how you react to changes in items presented to you and changes in prices. They want enough information to connect you across stores based on your mode of transportation, and have enough data to connect your habits over the last two decades.

Your digital existence should not be subject to slavery either. Ownership over ourselves is a vital aspect of freedom. Privacy is about ownership and dominion. If you dislike all the digital rights management and subscription services nonsense, these exist now as a direct result of people neglecting ownership. In the big picture, this path leads all of humanity back into another age of feudalism. The only difference between a serf and a citizen is ownership over property and tools. Everything happening right now is a battle over a new age of slavery. "You will own nothing and you will be happy about it." Eventually this turns into 'Your grandchildren will own nothing and say nothing or they will be dead about it." What you do about your privacy now will be a very big deal from the perspective of future generations.

[–] TheOtherJake 18 points 1 year ago (1 children)

Hey there Lionir. Thanks for the post. Can the Beehaw team please look into copying or getting the creator of this bot to work here? https://lemmy.world/u/PipedLinkBot@feddit.rocks

I think the person that created that bot is somehow connected to the piped.video project. I know the whole privacy consciousness thing isn't for everyone, but this bot's posts are quite popular elsewhere on Lemmy.

FYI, the main reason to use piped.video links is that it is setup as an alternative front end for YT that automatically routes all users through a bunch of VPNs to help mitigate Alphabet's privacy abuses and manipulation.

 

No paywall or permissions nonsense

 

Ideal background material IMO

20
submitted 1 year ago* (last edited 1 year ago) by TheOtherJake to c/food
 

I don't want the super health food tree bark nonsense you give nonbelievers. I'm looking for better than those of any animal infidels. Don't hold back on me now!

 

I'm not talking "you don't need a knife" level here, I'm looking for, "you need a spoon to finish the last bits" level of falling apart.

What are your specific techniques and tricks for different cuts?

Also, if you know a great Tennessee style dry rub I really want to know about it please.

 

Tell me the details like what makes yours perfect, why, and your cultural influence if any. I mean, rice is totally different with Mexican, Chinese, Indian, Japanese, and Persian food just to name a few. It is not just the spices or sauces I'm mostly interested in. These matter too. I am really interested in the grain variety and specifically how you prep, cook, and absolutely anything you do after. Don't skip the cultural details that you might otherwise presume everyone does. Do you know why some brand or region produces better ingredients, say so. I know it seems simple and mundane but it really is not. I want to master your rice as you make it in your culture. Please tell me how.

So, how do you do rice?

view more: ‹ prev next ›