Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)VI
Posts
0
Comments
226
Joined
4 mo. ago

  • I don't understand why Gemini is such a disaster. DeepMind Gemma works better and that's a 27B model. It's like there are two separate companies inside Google fucking off and doing their own thing (which is probably true)

  • Not making these famous logical errors

    For example, how many Rs are in Strawberry? Or shit like that

    (Although that one is a bad example because token based models will fundamentally make such mistakes. There is a new technique that lets LLMs process byte level information that fixes it, however)

  • Small scale models, like Mistral Small or Qwen series, are achieving SOTA performance with lower than 50 billion parameters. QwQ32 could already rival shitGPT with 32 billion parameters, and the new Qwen3 and Gemma (from google) are almost black magic.

    Gemma 4B is more comprehensible than GPT4o, the performance race is fucking insane.

    ClosedAI is 90% hype. Their models are benchmark princesses, but they need huuuuuuge active parameter sizes to effectively reach their numbers.

    Everything said in this post is independently verifiable by taking 5 minutes to search shit up, and yet you couldn't even bother to do that.

  • You can experiment on your own GPU by running the tests using a variety of models of different generations (LLAMA 2 class 7B, LLAMA 3 class 7B, Gemma, Granite, Qwen, etc...)

    Even the lowest end desktop hardware can run atleast 4B models. The only real difficulty is scripting the test system, but the papers are usually helpful with describing their test methodology.

  • and that you, yourself, can stand as a sole beacon against the otherwise regularly increasing evidence and studies that both indicate toward and also prove your claims to be full of shit?

    Hallucination rates and model quality has been going up steadily, same with multishot prompts and RAG reducing hallucination rates. These are proven scientific facts, what the fuck are you on about? Open huggingface RIGHT NOW, go to the papers section, FUCKING READ.

    I've spent 6+ years of my life in compsci academia to come here and be lectured by McDonald in his fucking basement, what has my life become