- cross-posted to:
- games@lemmy.world
- gaming@beehaw.org
- cross-posted to:
- games@lemmy.world
- gaming@beehaw.org
As a programmer: most people vastly overestimate the efficacy of large language models.
CEOs seem to overestimate them even more than everyone else.
A lot of AI researchers think LLMs are a dead end (See: Timnit Gebru) because by their structure they cannot understand truth.
The “hallucinations” are intrinsic to the structure and the best minds are saying there’s no way around that.
We might be able to cludge together filters over it but at some point that’s just hard coding the world anyways, which is what LLMs are supposed to avoid.
As a data scientist, people seem to just attribute anything that is a computer and they don’t understand to AI or worse ChatGPT. Shudder
I’ve been using chatgpt a lot, and it’s really clear to me it has many uses, but it’s almost more like asking your buddy who knows a lot but is full of shit too – sometimes he tells you exactly what you need, sometimes he sends you on a wild goose chase with all kinds of false leads.
In the end you need your own competence because the human needs to be able to make a final decision about whether to listen or not.
My EM suggested an integration using an SDK that doesn’t exist.
He was very insistent that we just hadn’t read the docs.
Then it came out that it was chat gpt suggesting it.