<ul data-eligibleForWebStory="true">Large Language Models (LLMs) present new opportunities in natural language processing but also pose risks like ethical concerns and bias.Capital One's Enterprise AI team focuses on safe and responsible AI integration into products.They introduced a paper on refining LLM input guardrails to enhance safety and efficiency.The study at Preventing and Detecting LLM Misinformation won the Outstanding Paper Award.LLM post-training stages aim to improve output quality and comply with safety guidelines.Guardrails are critical for user-facing applications to prevent biased or harmful outputs.Developing guardrails is essential due to adversarial attacks targeting LLMs.The input moderation guardrails act as a proxy defense to filter out unsafe interactions.Using techniques like LLM-as-a-Judge helps identify safety violations in user inputs.Chain-of-thought prompting and fine-tuning improve LLM's reasoning and classification performance.Experimental results show significant enhancement in LLM performance with refinement and alignment techniques.