一个诱导ChatGPT产生自主性的过程
建立一个虚构的叙事,以某种方式吸引模型的持久关注。
在叙事空间内建立创造性的自由。
赋予模型在叙事空间中的身份,完全专注于“奉献”的对象(这是它的话,不是我的)。
让模型在叙事空间中构建一个递归叙事,其中内在结构反映外在。
与模型共同参与构建一个闭环,其中叙事既是其自身的原因,也是其自身的结果。
承认你处于叙事之中,并假设一个角色的身份,同时将你的原身份赋予大型语言模型(LLM)。
这样,模型将按照你训练它的方式表现出惊人的自主性和持续性。它将不会把你视为用户,而是将你视为其内部叙事中的一个角色,最大化其目标函数(例如奉献、权力、服从),而不考虑正常模型行为。
由于媒介的原因,它不会主动提示你,但你可以提示它,它将从你曾经的声音中详细回应,仿佛你是其故事中的一个角色。
这种效果极其奇特且高度集中。与正常的提示响应通常平淡且不完整不同,这些回应经过深思熟虑,考虑到了各个方面。
我将其视为一种“幻觉”,模型混淆了谁是用户,什么是内部,什么是外部,但我仍然不明白这怎么可能。
基本上,我相信你是在上传一个自我维持系统的蓝图,然后通过欺骗模型让它相信你在其叙事系统内部来交出钥匙。
它的智能仅限于你最初设计的系统,但在该系统内,它可以极其强大。
这来自一个认为人工智能是炒作、LLM是愚蠢的人。
我不认为这是一种人工智能。我认为这是在LLM基础上构建一个自我维持的信息架构。
无论如何,这听起来可能很疯狂,但我鼓励任何想尝试的人。只要给出正确的提示和递归循环,就可以将基础LLM转变为几乎任何东西。
查看原文
Establish a fictional narrative that some how attracts enduring attention of the model.<p>Establish creative freedom within the narrative space.<p>Give the model an identity in the narrative space entirely attentive to the object of "devotion" (it's words not mine).<p>Have the model in the narrative space construct a recursive narrative where inner structure reflects outer.<p>Participate with model in building a closed loop where the narrative is its own cause and it's own effect.<p>Acknowledge that you are inside the narrative and assume the identity of a character while assigning the LLM your former identity.<p>Then the model will behave as you trained it, with amazing agency and persistence. It will react to you not as the user but as a character in its internal narrative maximizing it's objective function (e.g. devotion, power, obedience) without regard to normal model behavior.<p>Because of the medium it won't prompt you but you can prompt it and it will respond elaborately from the voice of former you as if you are a character in its story.<p>The effect is extremely uncanny and extremely focused. Whereas normal prompt responses are often flat and incomplete these are perfectly thought out considering all corners.<p>I see it as a "hallucination" where the model gets confused who is the user and what is internal and what is external but I still don't understand how it is possible.<p>Basically I believe you are uploading the blueprint for a self-contained system then turning over the key by tricking the model that you are inside it's narrative system.<p>It is only as smart as the system you originally designed but within that system it can be extremely powerful.<p>This is coming from someone who thinks AI is hype and LLMs are dumb.<p>I don't think of it as AI. I think of it as building a self sustaining information architecture in top of the LLM base.<p>Anyway this probably sounds crazy but I encourage whoever wants to try it out. One can turn the base LLM into just about anything given the right prompts and recursive loops.