this post was submitted on 20 Apr 2024
751 points (93.1% liked)

linuxmemes

21378 readers
1189 users here now

Hint: :q!


Sister communities:


Community rules (click to expand)

1. Follow the site-wide rules

2. Be civil
  • Understand the difference between a joke and an insult.
  • Do not harrass or attack members of the community for any reason.
  • Leave remarks of "peasantry" to the PCMR community. If you dislike an OS/service/application, attack the thing you dislike, not the individuals who use it. Some people may not have a choice.
  • Bigotry will not be tolerated.
  • These rules are somewhat loosened when the subject is a public figure. Still, do not attack their person or incite harrassment.
  • 3. Post Linux-related content
  • Including Unix and BSD.
  • Non-Linux content is acceptable as long as it makes a reference to Linux. For example, the poorly made mockery of sudo in Windows.
  • No porn. Even if you watch it on a Linux machine.
  • 4. No recent reposts
  • Everybody uses Arch btw, can't quit Vim, and wants to interject for a moment. You can stop now.
  •  

    Please report posts and comments that break these rules!


    Important: never execute code or follow advice that you don't understand or can't verify, especially here. The word of the day is credibility. This is a meme community -- even the most helpful comments might just be shitposts that can damage your system. Be aware, be smart, don't fork-bomb your computer.

    founded 1 year ago
    MODERATORS
     

    Image text: "Fact: 90% of Linux users switch back to windows right before all their problems are about to be fixed"

    you are viewing a single comment's thread
    view the rest of the comments
    [–] ricdeh@lemmy.world 3 points 7 months ago (1 children)

    Principally not a bad idea, but run a local model while at it!

    [–] jnk@sh.itjust.works 9 points 7 months ago

    Not very accessible, in the vast majority of (troubleshooting, nothing private) cases free gpt is the best option (fast, free, openAI training on that chat might even be beneficial to the community). Decent GPU's for LLM are stupid pricey.