Vulnerabilities and Robustness of LLMs in Augmented Democracy Systems
le 11 mars 2025
12h45
Manufacture des Tabacs MH003
Jairo Gudiño
Abstract: The use of Large Language Models (LLMs) in augmented democracy systems is gaining traction in recent literature, as they help generate consensus statements and aggregate preferences from diverse citizen inputs. However, their integration introduces critical vulnerabilities that could compromise their reliability in democratic processes. This paper examines their robustness against prompt-injection attacks, where adversarial manipulations distort LLM-driven consensus formation by injecting malicious instructions in external inputs. Our preliminary findings indicate that SeqAlign, as a defense technique, can effectively mitigate these attacks, but further research is needed as new attack categories and strategies continue to emerge. These results highlight the need for strong safeguards to ensure the reliability and resilience of LLM-based augmented democracy systems.
En appuyant sur le bouton "j'accepte" vous nous autorisez à déposer des cookies afin de mesurer l'audience de notre site. Ces données sont à notre seul usage et ne sont pas communiquées. Consultez notre politique relative aux cookies