• 3 Posts
  • 19 Comments
Joined 2 years ago
cake
Cake day: June 23rd, 2023

help-circle


  • There are several “good” LLMs trained on open datasets like FineWeb, LAION, DataComp, etc.

    Then use those as training data. You’re too caught up on this exacting definition of open source that you’ll completely ignore the benefits of what this model could provide.

    an LLM could decide to, for example, summarize and compress some context full of trade secrets, then proceed to “search” for it, sending it to wherever it has access to.

    That’s not how LLMs work, and you know it. A model of weights is not a lossless compression algorithm.

    Also, if you’re giving an LLM free reign to all of your session tokens and security passwords, that’s on you.




















  • Masturbation abstinence practices have returned to the USA in the form of semen retention communities.

    Well, that’s a combination of words I’ve never ever seen before.

    Also, both this article and web site are weird. It’s this strange combination of a “news article” and a research paper. Which is it?

    This study uses a systematic search on r/NoFap and two control forums (r/pornfree, and r/stopdrinking) to define a corpus of violent content.

    Well, there’s a textbook definition of cherry-picking right there. Control group needs more than just two other forums.

    Of the 421 violent posts identified from September 2011 to September 2022,

    Oh boy, if I had a dollar for the amount of times I saw a news article reference a science paper with an abysmal sample size, I could use that money to buy a decent used car. It’s the main problem on something like 98% of the science shit I’ve see here and on Reddit.