🧀 BigCheese.ai

Social

Ask HN: How do you add guard rails in LLM response without breaking streaming?

🧀

A discussion on Hacker News explores the best practices for adding guard rails to LLM responses without disrupting streaming functionality. Participants share insights on using regex filters, LLAMA Guard, token sampling, and fine-tuning stop words for optimal performance.

  • Topic involves LLM response constraints
  • Streaming aspect is critical
  • Regex filters are suggested
  • Token sampling can help
  • Fine-tuning stop words is useful