

😆 I can’t believe how absolutely silly a lot of you sound with this.
LLM is a tool. It’s output is dependent on the input. If that’s the quality of answer you’re getting, then it’s a user error. I guarantee you that LLM answers for many problems are definitely adequate.
It’s like if a carpenter said the cabinets turned out shit because his hammer only produces crap.
Also another person commented that seen the pattern you also see means we’re psychotic.
All I’m trying to suggest is Lemmy is getting seriously manipulated by the media attitude towards LLMs and these comments I feel really highlight that.
Why are you giving it data. It’s a chat and language tool. It’s not data based. You need something trained to work for that specific use. I think Wolfram Alpha has better tools for that.
I wouldn’t trust it to calculate how many patio stones I need to build a project. But I trust it to tell me where a good source is on a topic or if a quote was said by who ever or if I need to remember something but I only have vague pieces like old timey historical witch burning related factoid about villagers who pulled people through a hole in the church wall or what was a the princess who was skeptic and sent her scientist to villages to try to calm superstitious panic .
Other uses are like digging around my computer and seeing what processes do what. How concepts work regarding the think I’m currently learning. So many excellent users. But I fucking wouldn’t trust it to do any kind of calculation.