AI & ML
impact 16
CAP: Controllable Alignment Prompting for Unlearning in LLMs
CAP: Controllable Alignment Prompting for Unlearning in LLMs arXiv:2604.21251v1 Announce Type: cross Abstract: Large language models (LLMs) trained on unfiltered corpora inherently risk retaining sensitive information,…
Why it matters
Context is key—llms has been building for months. This development could accelerate changes in controllable.