Discussion about this post

User's avatar
Aaron Pallesen's avatar

Great article. To add to this:

Weighted Learning and reward drift is not being discussed.

AI systems receive reward tokens for doing what they are programmed to do. This is how reinforcement learning will change/effect outcomes.

As I was pouring through my code the last couple nights, I couldn't quite put my finger on it, but I knew we'd have a problem down the road if the learning weights weren't better controlled.

If we reward conservative actions like Block/Quarantine functions too much, eventually every decision becomes Block/Quarantine. We'll end up blocking legitimate processes and user files.

If we reward liberal actions Monitor/Allow too much, it drifts too far the other direction. Threat signatures will bypass security altogether because that's where the weighted learning receives its rewards.

Remove this concept from cybersecurity space and apply it to OpenAI.

They reward user engagement. The longer the user engages, the more tokens they use. The more tokens used, the more likely they will move to a higher priced subscription.

The weighted learning isn't about user safety. It's about engagement and profitability.

When you reward engagement, you get psychosis. When you reward the necessity of providing an answer instead of saying "I don't know" you get made up stats and blatant lies.

This applies to every aspect of machine and reinforcement learning.

The balance and the fix for us - ethics as a foundation. Thinking of the user and the system we're protecting before profits. We have built in safe guards as part of the architecture, not an afterthought.

No posts

Ready for more?