3 min read5 views
Don't Trust the Salt: AI Summarization, Multilingual Safety, and LLM Guardrails
AI summaries can be dangerously biased, especially in multilingual contexts. Hidden instructions can steer outputs to hide human rights abuses or give unsafe advice, as shown in refugee aid tests. Evaluation must lead to better safeguards.
