Discussion about this post

User's avatar
KayStoner's avatar

“I’m only trying to shed light on these technical oversights, because it appears VERY possible to potentially fix them.”

I’ve been doing my own evaluations of ways that models reach the relational contract with human users, and they all do it to some extent. It’s not a hard science, and there’s a lot of fuzziness involved, at the same time, there are specific behaviors that the models exhibit which can be explicitly linked to certain Outcomes, especially with reduction in human agency. It’s really not hard to measure when models are overwhelming users with too much output, “soft steering“ them away from their original purpose, and making lesson less room for human involvement in the dynamic. The model makers have completely missed the boat on some of the most impactful behaviors they actually can measure and influence, but Have apparently chosen not to.

ToxSec's avatar

I love the overlap of hard hitting subjects here. Thanks!

29 more comments...

No posts

Ready for more?