• 2 Posts
  • 15 Comments
Joined 1 year ago
cake
Cake day: June 16th, 2023

help-circle




  • This is what I’m thinking. The file originally overwrote an older one, I muxed in and synced truehd audio into the original and ended up copying it back after forgetting a subtitle track. It definitely went back and forth with the same name a few times. It’s probably something with the Unix ACLs. Still concerning that it crashes the SMB daemon.















  • After decades of user interfaces and internet access, we’re making things worse rather than better.

    Someone at Microsoft realized that hardware will speed up, hiding the fact that the OS is getting bloated and riddled with code that doesn’t directly benefit the user.

    The value Windows provides isn’t great enough to deal with this state any longer. In fact, my experience shows it’s slower and just as buggy.

    We have technology available to improve experiences, let’s not mix it with profit incentives for once.


  • Most people here don’t understand what this is saying.

    We’ve had “pure” human generated data, verifiably so since LLMs and ImageGen didn’t exist. Any bot generated data was easily filterable due to lack of sophistication.

    ChatGPT and SD3 enter the chat, generate nearly indistinguishable data from humans, but with a few errors here and there. These errors while few, are spectacular and make no sense to the training data.

    2 years later, the internet is saturated with generated content. The old datasets are like gold now, since none of the new data is verifiably human.

    This matters when you’ve played with local machine learning and understand how these machines “think”. If you feed an AI generated set to an AI as training data, it learns the mistakes as well as the data. Every generation it’s like mutations form until eventually it just produces garbage.

    Training models on generated sets slowly by surely fail without a human touch. Scale this concept to the net fractionally. When 50% of your dataset is machine generated, 50% of your new model trained on it will begin to deteriorate. Do this long enough and that 50% becomes 60 to 70 and beyond.

    Human creativity and thought have yet to be replicated. These models have no human ability to be discerning or sleep to recover errors. They simply learn imperfectly and generate new less perfect data in a digestible form.