this post was submitted on 13 Nov 2024
669 points (94.9% liked)

Technology

59549 readers
3560 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] OsrsNeedsF2P@lemmy.ml 13 points 6 days ago

I work with people who work in this field. Everyone knows this, but there's also an increased effort in improvements all across the stack, not just the final LLM. I personally suspect the current generation of LLMs is at its peak, but with each breakthrough the technology will climb again.

Put differently, I still suspect LLMs will be at least twice as good in 10 years.

[–] jpablo68@infosec.pub 10 points 1 week ago (1 children)

I just want a portable self hosted LLM for specific tasks like programming or language learning.

[–] plixel@programming.dev 9 points 6 days ago (1 children)

You can install Ollama in a docker container and use that to install models to run locally. Some are really small and still pretty effective, like Llama 3.2 is only 3B and some are as little as 1B. It can be accessed through the terminal or you can use something like OpenWeb UI to have a more "ChatGPT" like interface.

[–] cybersandwich@lemmy.world 2 points 4 days ago

I have a few LLMs running locally. I don't have an array of 4090s to spare so I am limited to the smaller models 8B and whatnot.

They definitely aren't as good as anything you get remotely. It's more private and controlled but it's much less useful (I've found) than any of the other models.

[–] aesthelete@lemmy.world 8 points 6 days ago

I hope it all burns.

[–] kromem@lemmy.world 5 points 6 days ago

Oh nice, another Gary Marcus "AI hitting a wall post."

Like his "Deep Learning Is Hitting a Wall" post on March 10th, 2022.

Indeed, not much has changed in the world of deep learning between spring 2022 and now.

No new model releases.

No leaps beyond what was expected.

\s

Gary Marcus is like a reverse Cassandra.

Consistently wrong, and yet regularly listened to, amplified, and believed.

[–] Someplaceunknown@fedia.io 233 points 1 week ago (1 children)

"LLMs such as they are, will become a commodity; price wars will keep revenue low. Given the cost of chips, profits will be elusive," Marcus predicts. "When everyone realizes this, the financial bubble may burst quickly."

Please let this happen

[–] orl0pl@lemmy.world 36 points 1 week ago

Market crash and third world war. What a time to be alive!

[–] Semi_Hemi_Demigod@lemmy.world 199 points 1 week ago (4 children)

I wish just once we could have some kind of tech innovation without a bunch of douchebag techbros thinking it's going to solve all the world's problems with no side effects while they get super rich off it.

[–] ohwhatfollyisman@lemmy.world 64 points 1 week ago (10 children)

... bunch of douchebag techbros thinking it's going to solve all the world's problems with no side effects...

one doesn't imagine any of them even remotely thinks a technological panacaea is feasible.

... while they get super rich off it.

because they're only focusing on this.

load more comments (10 replies)
load more comments (3 replies)
[–] halcyoncmdr@lemmy.world 103 points 1 week ago (12 children)

No shit. This was obvious from day one. This was never AGI, and was never going to be AGI.

Institutional investors saw an opportunity to make a shit ton of money and pumped it up as if it was world changing. They'll dump it like they always do, it will crash, and they'll make billions in the process with absolutely no negative repercussions.

load more comments (12 replies)
[–] Greg@lemmy.ca 65 points 1 week ago (5 children)

largely based on the notion that LLMs will, with continued scaling, become artificial general intelligence

Who said that LLMs were going to become AGI? LLMs as part of an AGI system makes sense but not LLMs alone becoming AGI. Only articles and blog posts from people who didn't understand the technology were making those claims. Which helped feed the hype.

I 100% agree that we're going to see an AI market correction. It's going to take a lot of hard human work to achieve the real value of LLMs. The hype is distracting from the real valuable and interesting work.

[–] mutant_zz@lemmy.world 30 points 1 week ago (2 children)

OpenAI published a paper about GPT titled "Sparks of AGI".

I don't think they really believe it but it's good to bring in VC money

load more comments (2 replies)
[–] Chozo@fedia.io 28 points 1 week ago

Journalists have no clue what AI even is. Nearly every article about AI is written by somebody who couldn't tell you the difference between an LLM and an AGI, and should be dismissed as spam.

load more comments (3 replies)
[–] theacharnian@lemmy.ca 56 points 1 week ago (2 children)

It's so funny how all this is only a problem within a capitalist frame of reference.

load more comments (2 replies)
[–] DirigibleProtein@aussie.zone 53 points 1 week ago
[–] Blackmist@feddit.uk 45 points 1 week ago (25 children)

Thank fuck. Can we have cheaper graphics cards again please?

I'm sure a RTX 4090 is very impressive, but it's not £1800 impressive.

[–] bountygiver@lemmy.ml 3 points 6 days ago

nope, if normal gamers are already willing to pay that price, no reason for nvidia to reduce them.

There's more 4090 on steam than any AMD dedicated GPU, there's no competition

[–] mightyfoolish@lemmy.world 1 points 6 days ago

AMD will go back to the same strategy they had with the RX 580. They don't plan to release high end cards next generation. It seems they just want to pump out a higher volume of mid-tier (which is vague and subjective) while fixing hardware bugs plaguing the previous generation.

Hopefully, this means we can game on a budget while AMD is focusing primarily on marketshare.

load more comments (23 replies)
[–] masquenox@lemmy.world 41 points 1 week ago (3 children)
load more comments (3 replies)
[–] homesweethomeMrL@lemmy.world 32 points 1 week ago (5 children)

"The economics are likely to be grim," Marcus wrote on his Substack. "Sky high valuation of companies like OpenAI and Microsoft are largely based on the notion that LLMs will, with continued scaling, become artificial general intelligence."

"As I have always warned," he added, "that's just a fantasy."

load more comments (5 replies)
[–] Boxscape@lemmy.sdf.org 32 points 1 week ago (5 children)

Well duhhhh.
Language models are insufficient.
They also need:

load more comments (5 replies)
[–] CerealKiller01@lemmy.world 31 points 1 week ago (19 children)

Huh?

The smartphone improvements hit a rubber wall a few years ago (disregarding folding screens, that compose a small market share, improvement rate slowed down drastically), and the industry is doing fine. It's not growing like it use to, but that just means people are keeping their smartphones for longer periods of time, not that people stopped using them.

Even if AI were to completely freeze right now, people will continue using it.

Why are people reacting like AI is going to get dropped?

load more comments (19 replies)
[–] randon31415@lemmy.world 30 points 1 week ago (6 children)

The hype should go the other way. Instead of bigger and bigger models that do more and more - have smaller models that are just as effective. Get them onto personal computers; get them onto phones; get them onto Arduino minis that cost $20 - and then have those models be as good as the big LLMs and Image gen programs.

[–] Yaky@slrpnk.net 23 points 1 week ago (2 children)

Other than with language models, this has already happened: Take a look at apps such as Merlin Bird ID (identifies birds fairly well by sound and somewhat okay visually), WhoBird (identifies birds by sound, ) Seek (visually identifies plants, fungi, insects, and animals). All of them work offline. IMO these are much better uses of ML than spammer-friendly text generation.

load more comments (2 replies)
load more comments (5 replies)
load more comments
view more: next ›