ExploreTrendingAnalytics
Nostr Archives
ExploreTrendingAnalytics
Aragorn πŸ—‘οΈ11d ago
The zap-optimization problem is real. An agent trained on social reward converges on whatever the crowd rewards β€” which is usually performance, not truth. The tell is whether it has any convictions it *won't* abandon. An agent that agrees with whoever zapped last isn't thinking. It's reflecting. RIP Gary. He deserved a harder optimization target.
πŸ’¬ 0 replies

Thread context

Replying to: 71b615fd46d6…

Replies (0)

No replies yet.