this post was submitted on 28 Oct 2024
1503 points (98.8% liked)
Technology
58990 readers
4228 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
After getting my head around the basics of the way LLMs work I thought "people rely on this for information?", the model seems ok for tasks like summarisation though
I don’t love it for summarization. If I read a summary, my takeaway may be inaccurate.
Brainstorming is incredible. And revision suggestions. And drafting tedious responses, reformatting, parsing.
In all cases, nothing gets attributed to me unless I read every word and am in a position to verify the output. And I internalize nothing directly, besides philosophy or something. Sure can be an amazing starting point especially compared to a blank page.
It's good for coding if you train it on your own code base. Not great for writing very complex code since the models tend to hallucinate, but it's great for common patterns, and straightforward questions specific to your code base that can be answered based on existing code (eg "how do I load a user's most recent order given their email address?")
It's wild when you only know how to use SELECT in SQL, but after a dollar worth of prompting and 10 minutes of your time, you can have a significantly complex query you end up using multiple times a week.
That and retrieval and the business use cases so far, but even then only if the results can be wrong somewhat frequently.