This is infininitivelivy worse
-
[email protected]replied to [email protected] last edited by
Ugh. Don’t get me started.
Most people don’t understand that the only thing it does is ‘put words together that usually go together’. It doesn’t know if something is right or wrong, just if it ‘sounds right’.
Now, if you throw in enough data, it’ll kinda sorta make sense with what it writes. But as soon as you try to verify the things it writes, it falls apart.
I once asked it to write a small article with a bit of history about my city and five interesting things to visit. In the history bit, it confused two people with similar names who lived 200 years apart. In the ‘things to visit’, it listed two museums by name that are hundreds of miles away. It invented another museum that does not exist. It also happily tells you to visit our Olympic stadium. While we do have a stadium, I can assure you we never hosted the Olympics. I’d remember that, as i’m older than said stadium.
The scary bit is: what it wrote was lovely. If you read it, you’d want to visit for sure. You’d have no clue that it was wholly wrong, because it sounds so confident.
AI has its uses. I’ve used it to rewrite a text that I already had and it does fine with tasks like that. Because you give it the correct info to work with.
Use the tool appropriately and it’s handy. Use it inappropriately and it’s a fucking menace to society.
-
JackFrostNColareplied to [email protected] last edited by
I know this is off topic, but every time i see you comment of a thread all i can see is the pepsi logo (i use the sync app for reference)
-
[email protected]replied to JackFrostNCola last edited by
You know, just for you: I just changed it to the Coca Cola santa
-
Wait, when did you do this? I just tried this for my town and researched each aspect to confirm myself. It was all correct. It talked about the natives that once lived here, how the land was taken by Mexico, then granted to some dude in the 1800s. The local attractions were spot on and things I've never heard of. I'm...I'm actually shocked and I just learned a bunch of actual history I had no idea of in my town 🤯
-
[email protected]replied to [email protected] last edited by
Spreading the holly day spirit
-
JackFrostNColareplied to [email protected] last edited by
We are all dutch on this blessed day
-
-
[email protected]replied to JackFrostNCola last edited by
We are all gekoloniseerd
-
[email protected]replied to [email protected] last edited by
ChatGPT is a tool under development and it will definitely improve in the long term. There is no reason to shit on it like that.
Instead, focus on the real problems: AI not being open-source, AI being under the control of a few monopolies, and there being little to none regulations that ensure it develops in a healthy direction.
-
I did that test late last year, and repeated it with another town this summer to see if it had improved. Granted, it made less mistakes - but still very annoying ones. Like placing a tourist info at a completely incorrect, non-existent address.
I assume your result also depends a bit on what town you try. I doubt it has really been trained with information pertaining to a city of 160.000 inhabitants in the Netherlands. It should do better with the US I’d imagine.
The problem is it doesn’t tell you it has knowledge gaps like that. Instead, it chooses to be confidently incorrect.
-
it will definitely improve in the long term.
Citation needed
There is no reason to shit on it like that.
Right now there is, because of how wrong it and other AIs can be, with the average person using the first answer as correct without double checking
-
Only 85k pop here, but yeah. I imagine it's half YMMV, half straight up luck that the model doesn't hallucinate shit.
-
AI is pretty over-rated but the Anti-AI forces way overblow the problems associated with AI.