AI & ML
impact 15
Can We Locate and Prevent Stereotypes in LLMs?
Can We Locate and Prevent Stereotypes in LLMs? arXiv:2604.19764v1 Announce Type: cross Abstract: Stereotypes in large language models (LLMs) can perpetuate harmful societal biases.
Why it matters
The timing matters: stereotypes is converging with shifts in llms, which could amplify the downstream impact.