this post was submitted on 17 Jul 2024
124 points (94.9% liked)

Technology

59607 readers
3256 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] technocrit@lemmy.dbzer0.com 0 points 4 months ago (1 children)

what solid proof do we have that humans aren’t doing the same?

Humans are not computers. Brains are not LLMs...

Given a totally reasonable hypothesis (humans =/= computers) and a completely outlandish hypothesis (humans = computers), I would need much more 'proof' for the later.

[–] sunbeam60@lemmy.one 1 points 4 months ago* (last edited 4 months ago) (1 children)

Well, brains are a network of neurons (we can evidentially verify this) trained on … eyes, ears, sense of touch, taste, smell and balance (rewarded by endorphins released by the old brain on certain hardcoded stimuli). LLMs are a network of neurons trained on text and images (rewarded by producing text that mimics input text and some reasoning tests).

It’s not given that this results in the same way of dealing with language, given the wider set of input data for a human, but it’s not given that it doesn’t either.

[–] zbyte64@awful.systems 1 points 4 months ago (1 children)

Humans predict things by assigning meaning to events and things, because in nature, we're constantly trying to guess what other creatures are planning. An LLM does not hypothesize what your plans are when you communicate to it, it's just trying to predict the next set of tokens with the greatest reward value. Even if you were to use literal human neurons to build your LLM, you would still have a stochastic parrot.

[–] sunbeam60@lemmy.one 2 points 4 months ago (1 children)

I mean I have an opinion too; what I’m seeking is evidence.

[–] zbyte64@awful.systems 2 points 4 months ago* (last edited 4 months ago) (1 children)

Why should I need to prove a negative? The burden is on the ones claiming an LLM is sentient. LLMs are token predictors, do I need to present evidence of this?

[–] sunbeam60@lemmy.one 1 points 4 months ago

I’m not asking you to prove anything. I’m saying I haven’t seen evidence either way so for me, it’s too early to draw conclusions.