8 comments

  • anyfoo 20 minutes ago
    This would have blown me away back in the late 80s/early 90s.

    (Or maybe not, if it doesn't perform better than random, I haven't actually tried it out yet. Some more examples would have been nice!)

    I wonder how far you could push this while still staying period correct, e.g. by adding a REU (RAM Expansion Unit), or even a GeoRAM (basically a REU on steroids).

    SuperCPU would also be an option, but for me it's always blurring the line of "what is a C64" a bit too much, and it likely just makes it faster anyway.

  • wk_end 1 hour ago
    > 25K parameters is about 70 million times smaller than GPT-4. It will produce broken sentences. That's the point - the architecture works at this scale.

    Since it seems to just produce broken and nonsensical sentences (at least based on the one example given) I'm not sure if it does work at this scale.

    Anyway, as written this passage doesn't really make a whole lot of sense (the point is that it produces broken sentences?), and given that it was almost certainly written by an AI, it demonstrates that the architecture doesn't work especially well at any scale (I kid, I kid).

    • forinti 1 hour ago
      How does it compare to a Markov chain generator I wonder.
    • pizza234 1 minute ago
      [dead]
  • classichasclass 36 minutes ago
    If you're running this in VICE, run it under the SuperCPU with warp mode on.
    • bartread 21 minutes ago
      That's a good idea because, although I love this, 1 minute per token is absolutely savage. Whereas if you can juice the performance you're into semi-credible Jar Jar Binks simulator territory.

      It does also make me wonder what you could do with somewhat more powerful retro hardware. I'd love to see what a transformer running on a PSX or an N64 could do.

  • brcmthrowaway 14 minutes ago
    How does this compare to ELIZA?
  • ghstinda 33 minutes ago
    but can you make mac keyboards feel like a c64c?
  • Lerc 44 minutes ago
    Ok now we need 1541 flash attention.

    I'm not sure what the venn diagram of knowledge to understand what that sentence is suggesting looks like, it's probably more crowded in the intersection than one might think.

  • harel 1 hour ago
    Eliza called, and asked if we saw her grand kids...
    • tclancy 56 minutes ago
      What makes you say that? This is about you, not me.

      (Came here to say an update to Eliza could really mess with the last person still talking to her.)

  • bighead1 1 hour ago
    i hate ai, and i love the c64, but i'll allow it.