@omarsar0
LLMs Can Learn About Themselves by Introspection This paper reports that LLMs can acquire knowledge through introspection that cannot be inferred from their training data. "Our findings challenge the view that LLMs merely imitate their training data and suggest they have privileged access to information about themselves." They also report that this introspection ability is limited and models struggle to predict their behavior on tasks requiring reasoning over long outputs. This is exciting and interesting because these introspection capabilities can lead to more interpretable and controllable LLMs.