this post was submitted on 10 Aug 2024
549 points (96.0% liked)

Technology

59696 readers
2451 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] FaceDeer@fedia.io -1 points 3 months ago (2 children)

Except it is capable of meaningfully doing so, just not in 100% of every conceivable situation. And those rare flubs are the ones that get spread around and laughed at, such as this example.

There's a nice phrase I commonly use, "don't let the perfect be the enemy of the good." These AIs are good enough at this point that I find them to be very useful. Not perfect, of course, but they don't have to be as long as you're prepared for those occasions, like this one, where they give a wrong result. Like any tool you have some responsibility to know how to use it and what its capabilities are.

[–] conciselyverbose@sh.itjust.works 4 points 3 months ago (1 children)

No, it isn't.

You're allowing a simple tool with literally zero reading comprehension to do your reading for you. It's not surprising your understanding of what the tech is is lacking.

[–] FaceDeer@fedia.io 0 points 3 months ago (1 children)

Your comment is simply counterfactual. I do indeed find LLMs to be useful. Saying "no you don't!" Is frankly ridiculous.

I'm a computer programmer. Not directly experienced with LLMs themselves, but I understand the technology around them and have written program that make use of them. I know what their capabilities and limitations are.

[–] conciselyverbose@sh.itjust.works 2 points 3 months ago (1 children)

Your claim that it's capable of doing what it claims isn't just false.

It's an egregious, massively harmful lie, and repeating it is always extremely malicious and inexcusable behavior.

[–] FaceDeer@fedia.io 1 points 3 months ago (1 children)

I have genuinely found LLMs to be useful in many contexts. I use them to brainstorm and flesh out ideas for tabletop roleplaying adventures, to write song lyrics, to write Python scripts to do various random tasks. I've talked with them to learn about stuff, and verified that they were correct by checking their references. LLMs are demonstrably capable of these things. I demonstrated it.

Go ahead and refrain from using them yourself if you really don't want to, for whatever reason. But exclaiming "no it doesn't!" In the face of them actually doing the things you say they don't is just silly.

[–] conciselyverbose@sh.itjust.works 1 points 3 months ago (2 children)

They absolutely cannot reliably summarize the result of searches, like this post is about, and OP in and of itself proves conclusively.

Any meaningful rate of failures at all makes them massively, catastrophically damaging to humanity as a whole. "Just don't use them" absolutely does not prevent their harm. Pushing them as competent is extremely fucking unacceptable behavior.

And this is all completely ignoring the obscene energy costs associated with making web searches complete and utter dogshit.

[–] FaceDeer@fedia.io 1 points 3 months ago (1 children)

They absolutely cannot reliably summarize the result of searches, like this post is about

The problem is that it did summarize the result of this search, the results of this search included one of those "if the Earth was the size of a grain of sand, Alpha Centauri would be X kilometers away" analogies. It did exactly the thing you're saying it can't do.

Any meaningful rate of failures at all makes them massively, catastrophically damaging to humanity as a whole.

Nothing is perfect. Does that make everything a massive catastrophic threat to humanity? How have we managed to survive for this long?

You're ridiculously overblowing this. It's a "ha ha, looks like AI made a whoopsie because I didn't understand that I actually asked it to do" situation. It's not Skynet coming to convince us to eat cyanide.

And this is all completely ignoring the obscene energy costs associated with making web searches complete and utter dogshit.

Of course it's ignoring that. It's not real.

You realize that energy costs money? If each web search cost an "obscene" amount, how is Microsoft managing to pay for it all? Why are they paying for it? Do you think they'll continue paying for it indefinitely? It'd be a completely self-solving problem.

[–] conciselyverbose@sh.itjust.works 2 points 3 months ago (2 children)

Summaries distinguish substance from nonsense. It cannot be described as a summary of a piece of content if it does not accurately portray the substance of that content.

LLMs aren't imperfect. They're dumpster fire misinformation machines with no redeeming qualities. Of course it's not Skynet. Skynet was intelligent. This isn't within 100 orders of magnitude of intelligence.

Companies burn obscene amounts of money on moonshots all the time, even ones that have no possibility of success. Willingness to lose billions burning energy to degrade every single search made is not an indication that it's not a nightmare for the environment (again, for literally no purpose because every single search with an LLM is worse than without it).

[–] FaceDeer@fedia.io 1 points 3 months ago (1 children)

No, a summary is just a condensed version of some larger work. If the larger work contains bullshit then so can the summary, that doesn't stop it from being a summary. As you say, a summary accurately portrays the substance of that content. In this case there was content that said Alpha Centauri was 13 km from Earth, so the summary said that too.

This is really not complicated.

Companies burn obscene amounts of money on moonshots all the time, even ones that have no possibility of success.

If you think it has no possibility of success, sit back and relax as AI goes away.

[–] ipkpjersi@lemmy.ml 2 points 3 months ago

If you think it has no possibility of success, sit back and relax as AI goes away.

Yep. This is exactly it, and this is what people don't seem to understand. AI is not going away, because it is actually useful, it has actual uses and people are actively using it. It's not entirely fluff based pointless technology like blockchain etc, it is actually useful and real-world people use AI/LLMs.

[–] ipkpjersi@lemmy.ml 0 points 3 months ago* (last edited 3 months ago)

This is just not true.

Just because you don't like LLMs doesn't mean that they have no purpose. If they were really entirely useless and served no purpose and never did anything, they would not be the talk of the town and OpenAI would not be a multi-billion dollar company. If they were useless, nobody would use them, but people absolutely do use them.

I literally use ChatGPT daily to automate writing code for me and it honestly does a good job. I literally used it to write an entire Laravel project called ArigatouAnimeTracker, over 600 commits including documentation all written using ChatGPT, and tbh my project is awesome. It easily would have taken me 5x as long to write it without ChatGPT and tbh it might not have ended up existing without ChatGPT because of how long it would have taken to write without LLMs doing the heavy lifting.

Sure, you have to verify the output, but you know what? That's going to be the case for any code that is written regardless, code review is essential and completely normal and existed long before LLMs did. That doesn't mean that LLMs don't have a purpose, or that nobody actually uses them. People do use them, it's a multi-billion dollar industry for a reason and people are going to continue to use them, even if you say they have no redeeming qualities. There are definitely ethical concerns about LLMs, but to say they have no redeeming qualities is just not correct.

Regardless, anything I say about AI/LLMs that isn’t that it’s terrible and useless and nobody should/would ever use it is going to be met with criticism.

[–] ipkpjersi@lemmy.ml 0 points 3 months ago* (last edited 3 months ago)

I disagree with that view on them, and I think the fact that they fail is actually a good thing in terms of preventing damage to humanity.

If they are able to perfectly do all kinds of jobs without ever making mistakes and being better at it than any humans, that would be infinitely more damaging than having them make mistakes meaning their use is limited to having to have their output carefully reviewed yet still used when it is helpful and appropriate.

Regardless, anything I say about AI/LLMs that isn't that it's terrible and useless and nobody should/would ever use it is going to be met with criticism.

[–] btaf45@lemmy.world 2 points 3 months ago (1 children)

AIs are definitely not "good enough" to give correct answers to science questions. I've seen lots of other incorrect answers before seeing this one. While it was easy to spot that this answer is incorrect, how many incorrect answers are not obvious?

[–] FaceDeer@fedia.io 1 points 3 months ago (1 children)

Then go ahead and put "science questions" into one of the areas that you don't use LLMs for. That doesn't make them useless in general.

I would say that a more precise and specific restriction would be "they're not good at questions involving numbers." That's narrower than "science questions" in general, they're still pretty good at dealing with the concepts involved. LLMs aren't good at math so don't use them for math.

[–] btaf45@lemmy.world 4 points 3 months ago

AI doesn't seem to be good at anything in which there is a right answer and a wrong answer. It works best for things where there are no right/wrong answers.