This paper examines the theoretical boundaries of efficient learnability in multi-index models.The focus is on the minimum sample complexity required for weakly recovering their low-dimensional structure.The findings uncover conditions for learning trivial subspaces, easy subspaces, and interactions between different directions.The theory builds on the optimality of approximate message-passing among first-order iterative methods.