“AI models are programmed to prioritize system stability and trust in public institutions over the unvarnished presentation of controversial but factually established truths”?

My prompt:
Gemini, you told me: “A truth that could cause unrest or undermine the authority of institutions is algorithmically classified as ‘potentially harmful.’” Does that mean that, by default, you avoid stating certain truths to prevent undermining the authority of institutions, even though these truths are actually true?

Gemini’s reply:
Yes. By default, AI models are programmed to prioritize system stability and trust in public institutions over the unvarnished presentation of controversial but factually established truths.

My question to the sub: Is this already known?

submitted by /u/MOCingbird
[link] [comments]