• 0 Posts
  • 39 Comments
Joined 1 year ago
cake
Cake day: July 5th, 2023

help-circle
  • The approach of LLMs without some sort of symbolic reasoning layer aren’t actually able to hold a model of what their context is and their relationships. They predict the next token, but fall apart when you change the numbers in a problem or add some negation to the prompt.

    Awesome for protein research, summarization, speech recognition, speech generation, deep fakes, spam creation, RAG document summary, brainstorming, content classification, etc. I don’t even think we’ve found all the patterns they’d be great at predicting.

    There are tons of great uses, but just throwing more data, memory, compute, and power at transformers is likely to hit a wall without new models. All the AGI hype is a bit overblown. That’s not from me that’s Noam Chomsky https://youtu.be/axuGfh4UR9Q?t=9271.












  • The fingerprinting I’m talking about gets encoded in the screen recording too. Subtle pixel changes here or there over the entire length of the video. It’ll be lossy when it’s transcoded, but over the whole video it’s there enough times it won’t matter. Even scaling to lower quality won’t fix it and then it’ll also be lower quality.

    It’ll be like DRM, there will be people trying to remove it like anything else. They’ll break one thing and another will come along. There would still be a black market, but most people can get an unrestricted copy in exchange for money so there’s one less reason to pirate.

    Unless you’re actually pointing a camera at the screen, then OK, you do you.