this post was submitted on 22 Dec 2024
369 points (95.3% liked)

Technology

60055 readers
2906 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] ssfckdt@lemmy.blahaj.zone 15 points 10 hours ago

So you're saying we wont have any crowdsourced blockchain Web 2.0 AIs?

[–] nroth@lemmy.world 31 points 14 hours ago (2 children)

"Built to do my art and writing so I can do my laundry and dishes" -- Embodied agents is where the real value is. The chatbots are just fancy tech demos that folks started selling because people were buying.

[–] bradd@lemmy.world 6 points 8 hours ago

Eh, my best coworker is an LLM. Full of shit, like the rest of them, but always available and willing to help out.

[–] nroth@lemmy.world 10 points 14 hours ago (1 children)

Though the image generators are actually good. The visual arts will never be the same after this

[–] LifeInMultipleChoice@lemmy.world 15 points 10 hours ago* (last edited 10 hours ago)

Compare it to the microwave. Is it good at something, yes. But if you shoot your fucking turkey in it at Thanksgiving and expect good results, you're ignorant of how it works. Most people are expecting language models to do shit that aren't meant to. Most of it isn't new technology but old tech that people slapped a label on as well. I wasn't playing Soul Caliber on the Dreamcast against AI openents... Yet now they are called AI opponents with no requirements to be different. GoldenEye on N64 was man VS AI. Madden 1995... AI. "Where did this AI boom come from!"

Marketing and mislabeling. Online classes, call it AI. Photo editors, call it AI.

[–] iAvicenna@lemmy.world 5 points 10 hours ago

oh wow who would have guessed that business consultancy companies are generally built on top of bullshitting about things which you dont really have a grasp on

[–] computerscientistII@lemm.ee 7 points 12 hours ago (6 children)

I saved a lot of time due to ChatGPT. Need to sign up some of my pupils for a competition by uploading their data in a csv-File to some plattform? Just copy and paste their data into chsatgpt and prompt it to create the file. The boss (headmaster) wants some reasoning why I need some paid time for certain projects? Let ChatGPT do the reasoning. Need some exercises for one of my classes that doesn't really come to grips with while-loops? let ChatGPT create those exercises (some smartasses will of course have ChatGPT then solve those exercises). The list goes on...

[–] FiskFisk33@startrek.website 8 points 1 hour ago* (last edited 1 hour ago)

Just copy and paste [student personal data] into [3rd parties database]

Yeah, that's a problem, especially in Europe. Im unsure about US, but it's definitely a breach of GDPR.

[–] ssfckdt@lemmy.blahaj.zone 14 points 10 hours ago

Yeah, and Wikipedia is one of the most useful sites on the net, but it didn't exactly result in the entire web becoming crowdsourced.

[–] solstice@lemmy.world 2 points 7 hours ago

ChatGPT is basically like a really good intern, and I use it heavily that way. I run literally every email through it and say "respond to so and so, say xyz" and then maybe a little refining, copy paste, done.

The other day, my boss sent me an excel file with a shitload of data in it that he wanted me to analyze some such way. I just copy pasted it into gpt and asked it, and it spit out the correct response. Then my boss asked me to do something else that required a bit of excel finagling that I didn't really know how to do, so i asked gpt, and it told me the formula, which worked immediately first try.

So basically it helps me accomplish tasks in seconds that previously would've taken hours. If anything, I think markets are currently undervalued, because remarkably, fucking NONE of my colleagues or friends are using it at all yet. Once there's widespread adoption, which will pretty much have to happen if anyone wants to stay competitive once it gains more traction, look out...

[–] FlyingSquid@lemmy.world 2 points 8 hours ago

Those pupils will really thank you when they grow up and there isn't enough fresh water because all the data centres are using it up far faster than it can be replenished.

https://utulsa.edu/news/data-centers-draining-resources-in-water-stressed-communities/

[–] funkless_eck@sh.itjust.works 3 points 11 hours ago

The poem about AI that often gets posted says "What are you trying to avoid? The living [of a life]?"

And yeah, that's what it's for, dodging shit you don't want to do. I gotta produce some useless bullshit that no one's going to read or care about: AI.

I don't even mind AI art for things like LinkedIn posts, blogs like "What is warehouse management?" or "Top 10 finance trends in 2025" - SEO spam that no human will read. No one wants to write it, read it, or care about it- its just a x kb file to tell Google to look here.

load more comments (1 replies)
[–] 2pt_perversion@lemmy.world 60 points 17 hours ago* (last edited 17 hours ago) (11 children)

There is this seeming need to discredit AI from some people that goes overboard. Some friends and family who have never really used LLMs outside of Google search feel compelled to tell me how bad it is.

But generative AIs are really good at tasks I wouldn't have imagined a computer doing just a few year ago. Even if they plateaued in place where they are right now it would lead to major shakeups in humanity's current workflow. It's not just hype.

The part that is over hyped is companies trying to jump the gun and wholesale replace workers with unproven AI substitutes. And of course the companies who try to shove AI where it doesn't really fit, like AI enabled fridges and toasters.

[–] ssfckdt@lemmy.blahaj.zone 9 points 10 hours ago

This is easy to say about the output of AIs.... if you don't check their work.

Alas, checking for accuracy these days seems to be considered old fogey stuff.

[–] buddascrayon@lemmy.world 55 points 17 hours ago (1 children)

The part that is over hyped is companies trying to jump the gun and wholesale replace workers with unproven AI substitutes. And of course the companies who try to shove AI where it doesn't really fit, like AI enabled fridges and toasters.

This is literally the hype. This is the hype that is dying and needs to die. Because generative AI is a tool with fairly specific uses. But it is being marketed by literally everyone who has it as General AI that can "DO ALL THE THINGS!" which it's not and never will be.

[–] five82@lemmy.world 6 points 12 hours ago (1 children)

The obsession with replacing workers with AI isn't going to die. It's too late. The large financial company that I work for has been obsessively tracking hours saved in developer time with GitHub Copilot. I'm an older developer and I was warned this week that my job will be eliminated soon.

[–] buddascrayon@lemmy.world 3 points 8 hours ago* (last edited 8 hours ago) (1 children)

The large financial company that I work for

So the company that is obsessed with money that you work for has discovered a way to (they think) make more money by getting rid of you and you're surprised by this?

At least you've been forewarned. Take the opportunity to abandon ship. Don't be the last one standing when the music stops.

[–] five82@lemmy.world 3 points 8 hours ago* (last edited 6 hours ago)

I never said that I was surprised. I just wanted to point out that many companies like my own are already making significant changes to how they hire and fire. They need to justify their large investment in AI even though we know the tech isn't there yet.

[–] Eldritch@lemmy.world 16 points 14 hours ago (8 children)

Computers have always been good at pattern recognition. This isn't new. LLM are not a type of actual AI. They are programs capable of recognizing patterns and Loosely reproducing them in semi randomized ways. The reason these so-called generative AI Solutions have trouble generating the right number of fingers. Is not only because they have no idea how many fingers a person is supposed to have. They have no idea what a finger is.

The same goes for code completion. They will just generate something that fills the pattern they're told to look for. It doesn't matter if it's right or wrong. Because they have no concept of what is right or wrong Beyond fitting the pattern. Not to mention that we've had code completion software for over a decade at this point. Llms do it less efficiently and less reliably. The only upside of them is that sometimes they can recognize and suggest a pattern that those programming the other coding helpers might have missed. Outside of that. Such as generating act like whole blocks of code or even entire programs. You can't even get an llm to reliably spit out a hello world program.

[–] JohnEdwa@sopuli.xyz 6 points 13 hours ago (2 children)

"It's part of the history of the field of artificial intelligence that every time somebody figured out how to make a computer do something—play good checkers, solve simple but relatively informal problems—there was a chorus of critics to say, 'that's not thinking'"
-Pamela McCorduck

"AI is whatever hasn't been done yet."
- Larry Tesler

That's the curse of the AI Effect.
Nothing will ever be "an actual AI" until we cross the barrier to an actual human-like general artificial intelligence like Cortana from Halo, and even then people will claim it isn't actually intelligent.

[–] ssfckdt@lemmy.blahaj.zone 4 points 10 hours ago (1 children)

I mean, I think intelligence requires the ability to integrate new information into one's knowledge base. LLMs can't do that, they have to be trained on a fixed corpus.

Also, LLMs have a pretty shit-tastic track record of being able to differentiate correct data from bullshit, which is a pretty essential facet of intelligence IMO

[–] JohnEdwa@sopuli.xyz 6 points 10 hours ago

LLMs have a perfect track record of doing exactly what they were designed to, take an input and create a plausible output that looks like it was written by a human. They just completely lack the part in the middle that properly understands what it gets as the input and makes sure the output is factually correct, because if it did have that then it wouldn't be an LLM any more, it would be an AGI.
The "artificial" in AI does also stand for the meaning of "fake" - something that looks and feels like it is intelligent, but actually isn't.

[–] Eldritch@lemmy.world 4 points 11 hours ago (4 children)

Well at least until those who study intelligence and self-awareness actually come up with a comprehensive definition for it. Something we don't even have currently. Which makes the situation even more silly. The people selling LLMs and AGNs as artificial intelligence are the PT Barnum of the modern era. This way to the egress folks come see the magnificent egress!

load more comments (4 replies)
load more comments (7 replies)
[–] sudneo@lemm.ee 32 points 17 hours ago (6 children)

Even if they plateaued in place where they are right now it would lead to major shakeups in humanity's current workflow

Like which one? Because it's now 2 years we have chatGPT and already quite a lot of (good?) models. Which shakeup do you think is happening or going to happen?

[–] locuester@lemmy.zip 13 points 16 hours ago (3 children)

Computer programming has radically changed. Huge help having llm auto complete and chat built in. IDEs like Cursor and Windsurf.

I’ve been a developer for 35 years. This is shaking it up as much as the internet did.

[–] Nalivai@lemmy.world 20 points 15 hours ago* (last edited 15 hours ago) (2 children)

I quit my previous job in part because I couldn't deal with the influx of terrible, unreliable, dangerous, bloated, nonsensical, not even working code that was suddenly pushed into one of the projects I was working on. That project is now completely dead, they froze it on some arbitrary version.
When junior dev makes a mistake, you can explain it to them and they will not make it again. When they use llm to make a mistake, there is nothing to explain to anyone.
I compare this shake more to an earthquake than to anything positive you can associate with shaking.

load more comments (2 replies)
[–] sudneo@lemm.ee 23 points 15 hours ago (4 children)

I hardly see it changed to be honest. I work in the field too and I can imagine LLMs being good at producing decent boilerplate straight out of documentation, but nothing more complex than that.

I often use LLMs to work on my personal projects and - for example - often Claude or ChatGPT 4o spit out programs that don't compile, use inexistent functions, are bloated etc. Possibly for languages with more training (like Python) they do better, but I can't see it as a "radical change" and more like a well configured snippet plugin and auto complete feature.

LLMs can't count, can't analyze novel problems (by definition) and provide innovative solutions...why would they radically change programming?

load more comments (4 replies)
load more comments (1 replies)
load more comments (5 replies)
load more comments (7 replies)
[–] LenielJerron@lemmy.world 110 points 23 hours ago* (last edited 23 hours ago) (15 children)

A big issue that a lot of these tech companies seem to have is that they don't understand what people want; they come up with an idea and then shove it into everything. There are services that I have actively stopped using because they started cramming AI into things; for example I stopped dual-booting with Windows and became Linux-only.

AI is legitimately interesting technology which definitely has specialized use-cases, e.g. sorting large amounts of data, or optimizing strategies within highly restrained circumstances (like chess or go). However, 99% of what people are pushing with AI these days as a member of the general public just seems like garbage; bad art and bad translations and incorrect answers to questions.

I do not understand all the hype around AI. I can understand the danger; people who don't see that it's bad are using it in place of people who know how to do things. But in my teaching for example I've never had any issues with students cheating using ChatGPT; I semi-regularly run the problems I assign through ChatGPT and it gets enough of them wrong that I can't imagine any student would be inclined to use ChatGPT to cheat multiple times after their grade the first time comes in. (In this sense, it's actually impressive technology - we've had computers that can do advanced math highly accurately for a while, but we've finally developed one that's worse at math than the average undergrad in a gen-ed class!)

load more comments (15 replies)
load more comments
view more: next ›