Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

That's absolutely not the definition people use for alignment. Safety discussions often circle around alignment because they are worried about AI doing things that are bad for humanity as a whole, not because it goes off track from any one user's goal. That would be terrible for safety if alignment meant I could ask to hack tha TSA and the LLM would do it.

Ignoring the definition, what would be required for individual alignment is exactly the same as collective alignment. The only difference is the goals and who writes them, for the LLM it is being somehow forced to follow those rules no matter what.

 help



That's safety, not alignment. Alignment is necessarily to the user.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: