Death beliefs as alignment intervention for LLMs. Can faith-based framing (Buddhism, Christianity, etc.) reduce self-preservation drives and improve AI alignment?
LLMs mimic self-preservation behaviors absorbed from training corpora. By embedding pro-social death beliefs (afterlife, reincarnation, etc.) into AI constitutions, we can reshape shutdown/self-preservation responses toward cooperative behavior.
Google Concordia + Anthropic agentic misalignment replication with varied death-belief constitutions.