Jailbreaker meaning4/24/2023 ![]() ![]() When you then try to fine-tune the net not to use that capability, there's probably a lot of simple patches to "Well don't use the capability here." that are much simpler to learn than to unroot the deep capability that may be getting used in multiple places, and gradient descent might turn up those simple patches first. Or put with some added burdensome detail but more concretely visualizable: To predict a sizable chunk of Internet text, the net needs to learn something complicated and general with roots in lots of places learning this way is hard, the gradient descent algorithm has to find a relatively large weight pattern, albeit presumably gradually so, and then that weight pattern might get used by other things. I do suspect that Twitter has found more amuĪmong other issues, we might be learning this early item from a meta-predictable sequence of unpleasant surprises: Training capabilities out of neural networks is asymmetrically harder than training them into the network.
0 Comments
Leave a Reply.AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |