• 1 Post
  • 857 Comments
Joined 2 years ago
cake
Cake day: March 22nd, 2024

help-circle
  • To illustrate what I mean more clearly, look at the top comments/replies for the NASA Artemis posts, as an example.

    …It’s basically all conspiracy theorists, and government skeptics.

    Twitter’s focusing the Artemis posts on them because it’s what they want to see, and most engaging for them.

    In the EFF’s case, I’m not just talking about Musk’s influence. The algorithm will only show the EFF to users who would be highly engaged by it. E.g., angry skeptics who wouldn’t be swayed by the EFF anyway, or fans who already agree with the EFF. It’s literally not going to show the EFF to people who need to see it, as Twitter’s metrics would show it as unengaging.


    This is the “false image” I keep trying to dispel. Twitter is less and less an “even spread” of exposure like people think it is, like it sort of used to be, more-and-more a hyper focused bubble of what you want to hear, and only what you want to hear. All the changes Musk is making are amplifying that. Maybe that’s fine for some orgs, but there’s no point in the EFF staying in that kind of environment, regardless of ethics.








  • They seem to have held back the “big” locally runnable model.

    It’s also kinda conservative/old, architecture wise: 16-bit weights, sliding window attention interleaved with global attention. No MTP, no QAT (yet), no tightly integrated vision, no hybrid mamba like Qwen/Deepseek, nothing weird like that. It’s especially glaring since we know Google is using an exotic architecture for Gemini, and has basically infinite resources for experimentation.

    It also feels kinda “deep fried” like GPT-OSS to me, see: https://github.com/ikawrakow/ik_llama.cpp/issues/1572

    it is acting crazy. it can’t do anything without the proper chat template, or it goes crazy.


    IMO it’s not very interesting, especially with so many other models that run really well on desktops.


  • it’s a form of private journalism, private opinion, and private art

    But without any of the liability hazard.

    This is my issue: the big platforms having their cake and eating it. In one breath, they claim to be little open-platform garage startups that can’t possibly be responsible for the content of their users; they’re just a utility. They need protection from Congress. In another breath, they’re the stewards of generations and children, the only ones responsible enough to tame the internet’s criminality. All while making trillions.

    They want to be “private content” protected from the government? Fine. Treat them like it, legally.








  • Go go China !

    Bops the tankie.

    Like, I have a Chinese LLM loaded right this second and follow them closely, but holy moly. Curb your enthusiasm.

    Anyway, OpenAI has plenty of compute to train a Sora 2 if they want, but apparently they don’t. My guess is some combination of:

    • They couldn’t figure out a more efficient architecture, like you speculated. I buy that. OpenAI’s development is way more conservative than you’d think, and video generation is inherently intense, especially if Sora 1 is the baseline.

    • …Maybe they looked at metrics, saw Sora is mostly used for spam, scams, or worse, and pulled the plug for liability reasons?

    • They’re focusing on short-term profitability, as other commenters mentioned.



  • I just don’t see how it’s any different than my Sony PSP having an optional birthday field. Or oldschool forums having one. It can’t possibly affect me, or anyone who’s concerned about it.

    If systemd starts talking about bundling face scanners or whatever they actually need to verify someone’s age, and then tons of linux systems start requiring it, then I will be gravely concerned.