Close Menu
AI News TodayAI News Today

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    FTC pushes ad agencies into dropping brand safety rules

    Ticketmaster is an illegal monopoly, jury rules

    NBA fans cry foul as Prime Video cuts out during overtime, fails to sync audio

    Facebook X (Twitter) Instagram
    • About Us
    • Contact Us
    Facebook X (Twitter) Instagram Pinterest Vimeo
    AI News TodayAI News Today
    • Home
    • Shop
    • AI News
    • AI Reviews
    • AI Tools
    • AI Tutorials
    • Chatbots
    • Free AI Tools
    AI News TodayAI News Today
    Home»AI Tools»Byte for byte, the most capable open models
    AI Tools

    Byte for byte, the most capable open models

    By No Comments2 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Motion blur lines, a mostly great bicycle albeit missing the front part of the frame. Pelican is decent.
    Share
    Facebook Twitter LinkedIn Pinterest Email

    2nd April 2026 – Link Blog

    Gemma 4: Byte for byte, the most capable open models. Four new vision-capable Apache 2.0 licensed reasoning LLMs from Google DeepMind, sized at 2B, 4B, 31B, plus a 26B-A4B Mixture-of-Experts.

    Google emphasize “unprecedented level of intelligence-per-parameter”, providing yet more evidence that creating small useful models is one of the hottest areas of research right now.

    They actually label the two smaller models as E2B and E4B for “Effective” parameter size. The system card explains:

    The smaller models incorporate Per-Layer Embeddings (PLE) to maximize parameter efficiency in on-device deployments. Rather than adding more layers or parameters to the model, PLE gives each decoder layer its own small embedding for every token. These embedding tables are large but are only used for quick lookups, which is why the effective parameter count is much smaller than the total.

    I don’t entirely understand that, but apparently that’s what the “E” in E2B means!

    I tried them out using the GGUFs for LM Studio. The 2B, 4B and 26B-A4B models all worked perfectly, but the 31B model was broken and spat out "---n" in a loop for every prompt I tried.

    The succession of pelican quality from 2B to 4B to 26B-A4B is notable:

    E2B:

    E4B:

    Two black wheels joined by a sort of grey surfboard, the pelican is semicircles and a blue blob floating above it

    26B-A4B:

    Bicycle has the right pieces although the frame is wonky. Pelican is genuinely good, has a big triangle beak and a nice curved neck and is clearly a bird that is sitting on the bicycle

    (This one actually had an SVG error – “error on line 18 at column 88: Attribute x1 redefined” – but after fixing that I got probably the best pelican I’ve seen yet from a model that runs on my laptop.)

    Google are providing API access to the two larger Gemma models via their AI Studio. I added support to llm-gemini and then ran a pelican through the 31B model using that:

    llm -m gemini/gemma-4-31b-it 'Generate an SVG of a pelican riding a bicycle'
    

    Pretty good, though it is missing the front part of the bicycle frame:

    Motion blur lines, a mostly great bicycle albeit missing the front part of the frame. Pelican is decent.

    Byte capable Models open
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleOpenAI acquires TBPN | OpenAI
    Next Article Five questions for the guys who made a compass that points to the Times Square Olive Garden
    • Website

    Related Posts

    AI Tools

    5 Practical Tips for Transforming Your Batch Data Pipeline into Real-Time: Upcoming Webinar

    AI Tools

    How to Maximize Claude Cowork

    AI Tools

    Prefill Is Compute-Bound. Decode Is Memory-Bound. Why Your GPU Shouldn’t Do Both.

    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    FTC pushes ad agencies into dropping brand safety rules

    0 Views

    Ticketmaster is an illegal monopoly, jury rules

    0 Views

    NBA fans cry foul as Prime Video cuts out during overtime, fails to sync audio

    0 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews
    AI Tutorials

    Quantization from the ground up

    AI Tools

    David Sacks is done as AI czar — here’s what he’s doing instead

    AI Reviews

    Judge sides with Anthropic to temporarily block the Pentagon’s ban

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Most Popular

    FTC pushes ad agencies into dropping brand safety rules

    0 Views

    Ticketmaster is an illegal monopoly, jury rules

    0 Views

    NBA fans cry foul as Prime Video cuts out during overtime, fails to sync audio

    0 Views
    Our Picks

    Quantization from the ground up

    David Sacks is done as AI czar — here’s what he’s doing instead

    Judge sides with Anthropic to temporarily block the Pentagon’s ban

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest
    • About Us
    • Contact Us
    • Terms & Conditions
    • Privacy Policy
    • Disclaimer

    © 2026 ainewstoday.co. All rights reserved. Designed by DD.

    Type above and press Enter to search. Press Esc to cancel.