A new research paper seeks to understand the evolution of knowledge in Large Vision-Language Models (LVLMs).
The study delves into the analysis of internal knowledge at different levels, including single token probabilities, token probability distributions, and feature encodings.
The research identifies two key nodes in knowledge evolution, namely critical layers and mutation layers, dividing the evolution process into rapid evolution, stabilization, and mutation.
This study provides valuable insights into the underlying mechanisms of LVLMs and contributes to their further enhancement.