17 comments

  • cyberjerkXX 55 minutes ago
    There was a security engineer at my work doing something similar to this. He wanted to use LLMs as an IDS. I begged him to use BPF and stop wasting sprint cycles trying to reinvent a shittier slower wheel.
    • pfortuny 53 minutes ago
      An elliptical wheel, at most. A square one without an axle, most probably.
  • amelius 1 hour ago
    How fast can Claude do branch predictions in CPU?
  • ShinyLeftPad 5 hours ago
    How quickly claude responds when it acts like a user space LLM chatbot?
  • mintflow 3 hours ago
    This is cool, let aside the token usage, perhaps it can help analyze tcp throughput by redirect wire shark/to dump result
    • fl7305 2 hours ago
      Opus 4.6 is already very good at troubleshooting all kinds of network problems if it has access to the command line tshark tool and the pcap files.
  • fouc 7 hours ago
    think about how much faster it would've been with a small local model!
  • twoodfin 6 hours ago
    Modulo Anthropic messing with the model for load mitigation, I wonder how stable this result is.

    1,000 pings, how many correctly ponged?

  • ValdikSS 10 hours ago
    That's why LLM will eventually be used only for initial interaction between the user in their language, to prepare the data to a specialized model.

    Imagine face recognition to work like a text chat, where the PC gets the frame from the camera and writes in the chat: "Who's that? Here's the RGB888 image in hex: ...".

    • FeepingCreature 6 hours ago
      That's actually how vision language models already work, pretty much.
      • wongarsu 3 hours ago
        And there's a reason nobody uses them for face recognition

        Vision language models are an incredible achievement in the generality and usability. But they pay a hefty price in fidelity and speed

      • stingraycharles 6 hours ago
        Huh? The images are tokenized in the same way language is and it’s just fed into one single model. Not multiple smaller expert models.

        Image gets rasterized into smaller pieces (eg 4x4 pixels) and each of those is assigned a token, similarly how text is broken up into tokens. And the whole thing is fed into a single model.

        • FeepingCreature 4 hours ago
          Yes I'm saying

          > Imagine face recognition to work like a text chat, where the PC gets the frame from the camera and writes in the chat: "Who's that? Here's the RGB888 image in hex: ...".

          that's p much how it works.

          • stingraycharles 3 hours ago
            But that isn’t a specialized model like the grandparent claimed, but rather a single, multi-modal model.
            • Dylan16807 2 hours ago
              Yes, the "imagine" was showcasing the opposite of a specialized model to call it a bad idea.
    • stingraycharles 6 hours ago
      Do you know that MoE is a thing?
      • jampekka 6 hours ago
        The experts in MoEs aren't specialized in any meaningful task sense. From level of what we would think as tasks MoEs are selected essentially arbitrarily per token and per block.
        • stingraycharles 5 hours ago
          It’s unsupervised, yes, but “unspecialized in any meaningful task sense” is incorrect, that’s the whole point. It’s just not in the sense of “this is a legal expert, this is a software developer”.
          • orbital-decay 5 minutes ago
            Optimal expert separation depends on the goal and can be pretty arbitrary, for example DeepSeek v4 separates them more or less by domain if I remember correctly.
  • bot403 4 hours ago
    Now do the equivalent of just in time compilation. Claude sees that we need to respond to a lot of pings and writes a program to compute it instead of thinking about each one.
  • ForHackernews 3 hours ago
    >Fun? Oh yeah!

    I think this author and I have different definitions of fun.

  • westurner 9 hours ago
    Wouldn't this be faster with an agent skill that has code?

    /skill-creator [or /create-skill] Write an agent skill with code script(s) that use an existing user space IP library that works with your agent runtime, to [...]

    ComposioHQ/awesome-claude-skills: https://github.com/ComposioHQ/awesome-claude-skills

    anthopics/skills//skill-creator/SKILL.md: https://github.com/anthropics/skills/blob/main/skills/skill-...

    /.agents/skills/skill-name/SKILL.md, scripts/{script_name.py,__init__.py}

    https://agentskills.io/what-are-skills

    • trollbridge 9 hours ago
      Well, yeah, of course it would be.

      Even faster would just to be use code in the first place!

  • brcmthrowaway 10 hours ago
    Next up: Claude replacement to handle simdjson processing.
  • iluvcommunism 1 hour ago
    [dead]
  • Ozzie-D 2 hours ago
    [dead]
  • self_awareness 2 hours ago
    If you wonder why your Copilot subscription has new limits that you hit every few days, it's because of PhDs like Adam.

    Could Adam use a local model hosted on his own box? Probably yes. But he preferred to waste the service we all use just to produce a weak blog post that introduces absolutely no knowledge and serves no other purpose than to tell everyone that the author likes to waste resources and calls it "fun".

    > Ridiculous? Yes. Wasteful of tokens? Sure. Fun? Oh yeah!

    Do you really think it's fun to be one of these people who are the reason why the rest of us gets more limits?

    • mystifyingpoi 1 hour ago
      Don't hate the player, hate the game.
      • self_awareness 1 hour ago
        No.

        In our lives, the game we play, we can do whatever we like. There are consequences for some things, but generally we can do lots of things.

        We can kill people and get away with it. We can also help them.

        Should we hate life because it's possible to do really shitty things in life? I don't think so. We should hate the "players" who actually do shitty things.

        • wolttam 5 minutes ago
          Ok but hating a guy playing with an LLM is a bit extreme. These things are in many ways still just toys (toys that are becoming increasingly useful).

          People almost certainly send dumber stuff to Claude than this, and just don’t write blog posts about it.

          You could try other providers if Anthropic is too slow/limited, there’s some good alternatives.

          (And your anger should probably be directed at Anthropic who hasn’t put in “better controls”, not the masses for not using the tool in the way you think they should. Hating rarely leads to anything productive.)

  • jeremyjh 7 hours ago
    Perhaps one day, all network services will be provided by LLMs natively. Truly, that would be a day in the future.
    • pastage 5 hours ago
      You could read about that in 1992 "A Fire Upon the Deep" by Vernor Vinge. There is prompt injection in communication, in the book certain protocols for information communication can not be deterministic so if someone is too smart you get hacked.
    • lionkor 4 hours ago
      "Perhaps" doing enough lifting to participate in a bodybuilder contest, in that sentence
    • vrighter 6 hours ago
      why? We already have more efficient specialized hardware.
    • codezero 7 hours ago
      I mean, we did decades of JavaScript, so... I mean... anything is possible, right? :)
  • fl7305 2 hours ago
    Do some people still claim "LLMs are just dumb auto completers"?

    Because this seems to disprove that claim pretty convincingly?

    • mystifyingpoi 1 hour ago
      Oh, they are. It's just that the harness around it is able to pick up the commands it "autocompletes" and runs them for you. LLM can't run anything, it never could.
    • AlienRobot 1 hour ago
      It proves that code, specifically any code in the form of bytes, is, too, language.
    • huflungdung 2 hours ago
      [dead]