• 2 Posts
  • 32 Comments
Joined 1 year ago
cake
Cake day: March 22nd, 2024

help-circle


  • A problem is volunteers and critical mass.

    Open source “hacks” need a big pool of people who want something to seed a few brilliant souls to develop it in their free time. It has to be at least proportional to the problem.

    This kinda makes sense for robot vacuums: a lot of people have them, and the cloud service is annoying, simpler, and not life critical.

    Teslas are a whole different deal. They are very expensive, and fewer people own them. Replicating even part of the cloud API calls is a completely different scope. The pool of Tesla owners willing to dedicate their time to that is just… smaller.

    Also, I think buying a Tesla, for many, was a vote of implicit trust in the company and its software. It’s harder for someone cynical of its cloud dependence to end up with an entire luxury automobile.



  • brucethemoose@lemmy.worldtoProgrammer Humor@programming.devC++
    link
    fedilink
    arrow-up
    145
    arrow-down
    2
    ·
    edit-2
    18 days ago

    Meanwhile, Rust punches you in the face for the mere suggestion. Again. And again.

    Python happily nods, runs it one page at a time, very slowly, randomly handing things off to a C person standing to the side, then returns a long poem. You wanted a number.

    Assembly does no checking, and reality around you tears from an access violation.

    EDIT: Oh, and the CUDA/PyTorch person is holding a vacuum sucking money from your wallet, with a long hose running to Jensen Huang’s kitchen.




  • This is hard.

    • Avatar: The Last Airbender + The Legend of Korra together. Don’t even get me started (actually please do).

    • Pantheon. Dense, proper sci fi, criminally and tragically suppressed IRL. Watch it, right now.

    • I am really loving Severance so far.

    • Arcane is up there, maybe as most artistic.

    • Star Wars: Andor. Not a huge Star Wars fan, but it’s so good, and grounded. It has everything.

    I have a lot of “secondary” favorites I am excluding here, like The Expanse, Person of Interest, Fringe, The Americans, Narcos, Mythbusters, and such for various reasons, but are still amazing.



  • It’s social media. It’s TV.

    So I bounce between 3 Trump states, and much of my family is glued to Facebook. Many don’t even know how to use a browser, just apps that serve them. Others religiously watch fox, a lot of cable opinion segments, things like that. I know a kid lost in Discord and YouTube.

    Also, you should see the things pastors are saying in one of their churches, in spite of what the clergy have done…

    I don’t know what it’s like overseas, but again, Americans seem totally consumed by social media, which is all just algorithmic propaganda that warps their world. And Trump/Musk are literally the top of that world.

    It’s only getting worse.

    That really is the root issue. No amount of protest, policy, empathy, anything is going to get through as long as people are immersed in these warped views of the world, and any political opposition is completely incapable of realizing that.




  • I have very religious family that repeatedly told my 90 year old grandma not to get vaccinated in the depths of COVID-19. I have other, not-at-all religious family that works as a nurse… And is anti vaccine.

    It’s like a parody.

    …But it is no joke. I can answer questions about them if you want.


    If you’re wondering why, it’s because many Americans are inundated in really scary social media and TV. That part of my family is constantly on Facebook, watching Fox, doomscrolling whatever. Even their church preaches some really, uh, interesting things now.

    It’s this way because there’s a lot of profiteering. For example, the current head of the FBI is apparently selling and promoting some kind of “brave anti vaccine” health merchandise. The current head of the US health department made a lot of money and fame off vaccine skepticism. And their church clergy is crooked in ways I can’t even publicly discuss.






  • brucethemoose@lemmy.worldtoMemes@lemmy.ml2 in a single week that is crazy
    link
    fedilink
    arrow-up
    8
    arrow-down
    1
    ·
    edit-2
    2 months ago

    I mean, if you have huge GPU, sure. Or at least 12GB free vram or a big Mac.

    Local LLMs for coding is kinda a niche because most people don’t have a 3090 or 7900 lying around, and you really need 12GB+ free VRAM for the models to start being “smart” and even worth using over free LLM APIs, much less cheap paid ones.

    But if you do have the hardware and the time to set a server up, the Deepseek R1 models or the FuseAI merges are great for “slow” answers where the model thinks things out for replying. Qwen 2.5 32B coder is great for quick answers on 24GB VRAM. Arcee 14B is great for 12GB VRAM.

    Sometimes running a small model on a “fast” less vram efficient backend is better for stuff like cursor code completion.


  • brucethemoose@lemmy.worldtoMemes@lemmy.ml2 in a single week that is crazy
    link
    fedilink
    arrow-up
    52
    arrow-down
    1
    ·
    edit-2
    2 months ago

    My friend, the Chinese have been releasing amazing models all last year, it just didn’t make headlines.

    Tencent’s Hunyuan Video is incredible. Alibabas Qwen is still a go to local model. I’ve used InternLM pretty regularly… Heck, Yi 32B was awesome in 2023, as the first decent long context local model.

    …The Janus models are actually kind of meh, unless you’re captioning images, and FLUX/Hunyuan Video is still king in diffusion world.


  • As implied above, the raw format fed to/outputed from Deepseek R1 is:

    <|begin▁of▁sentence|>{system_prompt}<|User|>{prompt}<|Assistant|><think>The model rambles on to itself here, “thinking” before answering</think>The actual answer goes here.

    It’s not a secret architecture, theres no window into its internal state. This is just a regular model trained to give internal monologues before the “real” answer.

    The point I’m making is that the monologue is totally dependent on the system prompt, the user prompt, and honestly, a “randomness” factor. Its not actually a good window into the LLM’s internal “thinking,” you’d want to look at specific tests and logit spreads for that.