this post was submitted on 13 Nov 2024
669 points (94.9% liked)

Technology

59549 readers
3112 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] OsrsNeedsF2P@lemmy.ml 13 points 6 days ago

I work with people who work in this field. Everyone knows this, but there's also an increased effort in improvements all across the stack, not just the final LLM. I personally suspect the current generation of LLMs is at its peak, but with each breakthrough the technology will climb again.

Put differently, I still suspect LLMs will be at least twice as good in 10 years.

[–] dejected_warp_core@lemmy.world 13 points 6 days ago (1 children)

Welcome to the top of the sigmoid curve.

If you were wondering what 1999 felt like WRT to the internet, well, here we are. The Matrix was still fresh in everyone's mind and a lot of online tech innovation kinda plateaued, followed by some "market adjustments."

[–] Hackworth@lemmy.world 6 points 6 days ago* (last edited 6 days ago)

I think it's more likely a compound sigmoid (don't Google that). LLMs are composed of distinct technologies working together. As we've reached the inflection point of the scaling for one, we've pivoted implementations to get back on track. Notably, context windows are no longer an issue. But the most recent pivot came just this week, allowing for a huge jump in performance. There are more promising stepping stones coming into view. Is the exponential curve just a series of sigmoids stacked too close together? In any case, the article's correct - just adding more compute to the same exact implementation hasn't enabled scaling exponentially.

[–] jpablo68@infosec.pub 10 points 6 days ago (1 children)

I just want a portable self hosted LLM for specific tasks like programming or language learning.

[–] plixel@programming.dev 9 points 6 days ago (1 children)

You can install Ollama in a docker container and use that to install models to run locally. Some are really small and still pretty effective, like Llama 3.2 is only 3B and some are as little as 1B. It can be accessed through the terminal or you can use something like OpenWeb UI to have a more "ChatGPT" like interface.

[–] cybersandwich@lemmy.world 2 points 4 days ago

I have a few LLMs running locally. I don't have an array of 4090s to spare so I am limited to the smaller models 8B and whatnot.

They definitely aren't as good as anything you get remotely. It's more private and controlled but it's much less useful (I've found) than any of the other models.

[–] aesthelete@lemmy.world 8 points 6 days ago

I hope it all burns.

[–] TankovayaDiviziya@lemmy.world 8 points 6 days ago (1 children)

Short on the AI stocks before it crash!

[–] sugar_in_your_tea@sh.itjust.works 16 points 6 days ago* (last edited 6 days ago)

The market can remain irrational longer than you can remain solvent.

A. Gary Shilling

[–] kromem@lemmy.world 5 points 6 days ago

Oh nice, another Gary Marcus "AI hitting a wall post."

Like his "Deep Learning Is Hitting a Wall" post on March 10th, 2022.

Indeed, not much has changed in the world of deep learning between spring 2022 and now.

No new model releases.

No leaps beyond what was expected.

\s

Gary Marcus is like a reverse Cassandra.

Consistently wrong, and yet regularly listened to, amplified, and believed.

[–] Someplaceunknown@fedia.io 233 points 1 week ago (1 children)

"LLMs such as they are, will become a commodity; price wars will keep revenue low. Given the cost of chips, profits will be elusive," Marcus predicts. "When everyone realizes this, the financial bubble may burst quickly."

Please let this happen

load more comments (1 replies)
load more comments
view more: next ›