Contemporary large language models (LLMs) like GPT-4 and Llama, combined with molecular structure embeddings, enable accurate prediction of polymer properties.
PolyLLMem, a multimodal architecture, integrates text embeddings from Llama 3 with molecular structure embeddings derived from Uni-Mol.
Low-rank adaptation (LoRA) layers are incorporated to refine the embeddings based on limited polymer dataset, enhancing their chemical relevance.
PolyLLMem's performance is comparable to graph-based and transformer-based models, even with limited training data, accelerating the discovery of advanced polymeric materials.