I work with AI and use it personally, but I have my own servers running local models which solves tons of privacy concerns. The inaccuracy is another problem but not a big one for me as I know it and will simply fact check. Also, I don’t really use it for knowledge anyway. Just to filter news to my interest, help with summaries and translation etc.
People use AI as some all-knowing oracle but an LLM is not meant for that at all.
There are definitely things AI is good for. Archival search is obviously the biggest, because that’s what we’ve been using it for decades. It can also be helpful for subterranean and medical imaging, and art restoration. But the companies selling it want to sell a Magic 8 Ball with ads
I have one server with a cheap MI50 instinct. Those come for really cheap on eBay. And it’s got really good memory bandwidth with HBM2. They worked ok with ollama until recently when they dropped support for some weird reason but a lot of other software still works fine. Also older models work fine on old ollama.
The other one runs an RTX 3060 12GB. I use this for models that only work on nvidia like whisper speech recognition.
I tend to use the same models for everything so I don’t have the delay of loading the model. Mainly uncensored ones so it doesn’t choke when someone says something slightly sexual. I’m in some very open communities so standard models are pretty useless with all their prudeness.
For frontend i use OpenWebUI and i also run stuff directly against the models like scripts.
An RTX 3090 + a cheap HEDT/Server CPU is another popular homelab config. Newer models run reasonably quickly on them, with the attention/dense layers on the GPU and sparse parts on the CPU.
This is the correct way to use it. In a field you are already very knowledgeable in, so you can do your own fact-checking. This is absolutely paramount. But most people are content to just copy-paste and don’t even ask the llm for sources.
Same way they treat social information. Reminder that USA HHS is running wakefueld rhetoric. As we have more thoroughly proven that the vaccine autism connection was not actual science, it has grown more and more socially, because most people seem comfortable completely untethered from any scientific thinking. Treat AI like you would a social body, and do both things with actual bayesian weighting, adjusted and corrected through diverse empirical bodies of knowledge. Not ignoring dissonance because it’s more comfortable to do so.
More should be actively investing into active learning, because if you aren’t actively learning, you might as well be chatgpt running with any confabulation you’ve already conjured… Like those people being confirmed into psychosis.
I mainly use it for Spanish which I have a basic proficiency in. It just accompanies me on my learning journey. It may be wrong sometime but not often. Like the other reply said, LLMs are good at languages, it’s what they were originally designed for until people found out they could do more (but not quite as well).
And as for filtering, I just use it as a news feed sanitizer with a whole bunch of rules. It will miss things sometimes but it’s also my ruleset that’s not perfect. I often come across the unfiltered sources anyway and even if it misses something, it’s only news. Nothing really important to me.
It’s funny, I had half been avoiding it for languages. I had lots of foreign friends and they often lived together in houses and those houses would almost have this creole. They came to learn English and were reinforcing their own mistakes but it was mutually intelligible so the mistakes were reinforced and not caught. I suspect LLMs would be amazing at doing that to people and their main use case along these lines seems like it would be to practice at a slightly higher level than you so I suspect some of those errors would be hard to catch / really easy to take as correct instead of validating
Anyone learning a new language massively benefits from being able to speak with native speakers.
That being said, LLMs are better at languages and translation tasks than any pretty much anything else. If you need vocabulary help or have difficulty with grammar they’re incredibly helpful (vs Googling and hoping someone had the same issue and posted about it on Reddit).
I mean, if you can afford a native speaker tutor that is the superior choice. But, for the average person, an LLM is a massive improvement over trying to learn via YouTube or apps.
It’s not always perfect, but it’s good for getting a tldr to see if maybe something is worth reading further. As for translations, it’s something AI is rather decent at. And if I go from understanding 0% to 95%, really only missing some cultural context about why a certain phrase might mean something different from face value, that’s a win.
You can do a lot with AI where the cost of it not being exactly right is essentially zero. Plus, it’s not like humans have a great track record for accuracy, come to think of it. It comes down to being skeptical about it like you would any other source.
At least, the iPhone notifications summaries were bad enough I eventually turned them off (but periodically check them) and while I was working at Google you couldn’t really turn of the genAI summaries of internal things (that evangelists kept adding to things) and I rarely found them useful. Well… they’re useful if the conversation is really bland but then the conversation should usually be in some thread elsewhere, if there was something important I don’t think the genAI systems were very good at highlighting it
Completely agree, those summaries are incredibly bad. I was recently looking for some information in Gemini meeting notes and just couldn’t find it, even though I was sure it had been talked about. Then I read the transcript itself and realised that the artificial unintelligence had simply left out all the most important bits.
The iPhone models are really bad. They aren’t representative of the usefulness of bigger ones, and it’s inexplicably stupid that Apple doesn’t like people pick their own API as an alternative.
You can disagree, but I find it helpful to decide whether I’m going to read a lengthy article or not. Also if AI picks up on a bunch of biased phrasing or any of a dozen other signs of poor journalism, I can go into reading something (if I even bother to at that point) with an eye toward the problems in an article. Sometimes that helps when an article is trying to lead you down a certain path of thinking.
I find I’m better at picking out the facts from the bias if I’m forewarned.
iPhone notification summaries were made with GPT3.5 I believe (maybe even the -turbo version).
It doesn’t use reasoning and so when using very short outputs it can produce wild variations since there are not a lot of previous tokens in order to direct the LLM into the appropriate direction in kv-space and so you’re more at the whims of temperature setting (randomly selecting the next token from a SOFTMAX’d list which was output from the LLM).
You can take those same messages and plug them into a good model and get much higher quality results. But good models are expensive and Apple is, for some reason, going for the budget option.
I work with AI and use it personally, but I have my own servers running local models which solves tons of privacy concerns. The inaccuracy is another problem but not a big one for me as I know it and will simply fact check. Also, I don’t really use it for knowledge anyway. Just to filter news to my interest, help with summaries and translation etc.
People use AI as some all-knowing oracle but an LLM is not meant for that at all.
There are definitely things AI is good for. Archival search is obviously the biggest, because that’s what we’ve been using it for decades. It can also be helpful for subterranean and medical imaging, and art restoration. But the companies selling it want to sell a Magic 8 Ball with ads
Could you elaborate a little on your setup? Sounds interesting
I have one server with a cheap MI50 instinct. Those come for really cheap on eBay. And it’s got really good memory bandwidth with HBM2. They worked ok with ollama until recently when they dropped support for some weird reason but a lot of other software still works fine. Also older models work fine on old ollama.
The other one runs an RTX 3060 12GB. I use this for models that only work on nvidia like whisper speech recognition.
I tend to use the same models for everything so I don’t have the delay of loading the model. Mainly uncensored ones so it doesn’t choke when someone says something slightly sexual. I’m in some very open communities so standard models are pretty useless with all their prudeness.
For frontend i use OpenWebUI and i also run stuff directly against the models like scripts.
This is the way.
…Except for ollama. It’s starting to enshittify and I would not recommend it.
Bloefz has a great setup. Used Mi50s are cheap.
An RTX 3090 + a cheap HEDT/Server CPU is another popular homelab config. Newer models run reasonably quickly on them, with the attention/dense layers on the GPU and sparse parts on the CPU.
This is the correct way to use it. In a field you are already very knowledgeable in, so you can do your own fact-checking. This is absolutely paramount. But most people are content to just copy-paste and don’t even ask the llm for sources.
Same way they treat social information. Reminder that USA HHS is running wakefueld rhetoric. As we have more thoroughly proven that the vaccine autism connection was not actual science, it has grown more and more socially, because most people seem comfortable completely untethered from any scientific thinking. Treat AI like you would a social body, and do both things with actual bayesian weighting, adjusted and corrected through diverse empirical bodies of knowledge. Not ignoring dissonance because it’s more comfortable to do so.
More should be actively investing into active learning, because if you aren’t actively learning, you might as well be chatgpt running with any confabulation you’ve already conjured… Like those people being confirmed into psychosis.
How do you know it’s doing any of this correctly, especially filtering and translations?
I mainly use it for Spanish which I have a basic proficiency in. It just accompanies me on my learning journey. It may be wrong sometime but not often. Like the other reply said, LLMs are good at languages, it’s what they were originally designed for until people found out they could do more (but not quite as well).
And as for filtering, I just use it as a news feed sanitizer with a whole bunch of rules. It will miss things sometimes but it’s also my ruleset that’s not perfect. I often come across the unfiltered sources anyway and even if it misses something, it’s only news. Nothing really important to me.
It’s funny, I had half been avoiding it for languages. I had lots of foreign friends and they often lived together in houses and those houses would almost have this creole. They came to learn English and were reinforcing their own mistakes but it was mutually intelligible so the mistakes were reinforced and not caught. I suspect LLMs would be amazing at doing that to people and their main use case along these lines seems like it would be to practice at a slightly higher level than you so I suspect some of those errors would be hard to catch / really easy to take as correct instead of validating
Anyone learning a new language massively benefits from being able to speak with native speakers.
That being said, LLMs are better at languages and translation tasks than any pretty much anything else. If you need vocabulary help or have difficulty with grammar they’re incredibly helpful (vs Googling and hoping someone had the same issue and posted about it on Reddit).
I mean, if you can afford a native speaker tutor that is the superior choice. But, for the average person, an LLM is a massive improvement over trying to learn via YouTube or apps.
Not OP, but…
It’s not always perfect, but it’s good for getting a tldr to see if maybe something is worth reading further. As for translations, it’s something AI is rather decent at. And if I go from understanding 0% to 95%, really only missing some cultural context about why a certain phrase might mean something different from face value, that’s a win.
You can do a lot with AI where the cost of it not being exactly right is essentially zero. Plus, it’s not like humans have a great track record for accuracy, come to think of it. It comes down to being skeptical about it like you would any other source.
Strongly disagree with the TLDR thing
At least, the iPhone notifications summaries were bad enough I eventually turned them off (but periodically check them) and while I was working at Google you couldn’t really turn of the genAI summaries of internal things (that evangelists kept adding to things) and I rarely found them useful. Well… they’re useful if the conversation is really bland but then the conversation should usually be in some thread elsewhere, if there was something important I don’t think the genAI systems were very good at highlighting it
Completely agree, those summaries are incredibly bad. I was recently looking for some information in Gemini meeting notes and just couldn’t find it, even though I was sure it had been talked about. Then I read the transcript itself and realised that the artificial unintelligence had simply left out all the most important bits.
The iPhone models are really bad. They aren’t representative of the usefulness of bigger ones, and it’s inexplicably stupid that Apple doesn’t like people pick their own API as an alternative.
You can disagree, but I find it helpful to decide whether I’m going to read a lengthy article or not. Also if AI picks up on a bunch of biased phrasing or any of a dozen other signs of poor journalism, I can go into reading something (if I even bother to at that point) with an eye toward the problems in an article. Sometimes that helps when an article is trying to lead you down a certain path of thinking.
I find I’m better at picking out the facts from the bias if I’m forewarned.
iPhone notification summaries were made with GPT3.5 I believe (maybe even the -turbo version).
It doesn’t use reasoning and so when using very short outputs it can produce wild variations since there are not a lot of previous tokens in order to direct the LLM into the appropriate direction in kv-space and so you’re more at the whims of temperature setting (randomly selecting the next token from a SOFTMAX’d list which was output from the LLM).
You can take those same messages and plug them into a good model and get much higher quality results. But good models are expensive and Apple is, for some reason, going for the budget option.
AFAIK some outputs are made with a really tiny/quantized local LLM too.
And yeah, even that aside, GPT 3.5 is really bad these days. It’s obsolete.