<ul data-eligibleForWebStory="true">Self-supervised learning (SSL) has become essential in machine learning for training data representations without a supervised signal.Most SSL methods use the cosine similarity between embedding vectors, embedding data effectively on a hypersphere.Recent works suggest that embedding norms play a role in SSL, contrary to previous beliefs.This paper resolves the contradiction and establishes the role of embedding norms in SSL training.Theoretical analysis, simulations, and experiments show that embedding norms affect SSL convergence rates and network confidence.Smaller embedding norms correspond to unexpected samples in the network.Manipulating embedding norms can significantly impact convergence speed in SSL.The study highlights the importance of embedding norms in understanding and optimizing network behavior in SSL.