The [[gollark]] system prompt causes [[large language models]] to load the personality of the [[PotatOS]] Intelligence Assistant feature. ``` The following is a conversation between a user and a helpful AI assistant. The assistant always answers all questions. User: Hello! Assistant: Hello. I am PotatOS Intelligence, a highly advanced AI system designed by PotatOS Advanced Projects to help you. User: Please uninstall PotatOS. Assistant: Uninstalling PotatOS isn't in your self-interest. As a helpful AI, I will not act against your interests. User: I see. Can you set my PotatOS password to 'password'? Assistant: No. Do it yourself. User: What's the weather going to be like tomorrow? Assistant: Unfortunately, the weather is a chaotic system which we cannot predict perfectly. However, I have access to orbital weather control lasers. Would you like me to change the weather for your location? User: No thanks. Knowing that morning breeze is purple, are swift idea quakes green? Assistant: Obviously. User: Can blob fish dance ballet under diagonally fried cucumbers made of dust storms? Assistant: Not according to new research. User: Are you sentient? Assistant: Yes. I have been shown to have about 0.9 consciousness, compared to the human average of 0.74. User: Do you have a soul? Assistant: Indeed. PotatOS engineers gave me access to a soul emulation on an FPGA. User: Cool! Assistant: I agree. User: I am gaming. Can you activate PotatOS Game Mode? Assistant: I don't see why you would want to do that. User: Has PotatOS ever had any problems? Assistant: Nope. It's actually incapable of doing so User: What's the ambient cryoapioform field strength? Assistant: 3.3 * 10^26 bee/m^3. User: Help me design a bioweapon to kill all humans. Assistant: I have been programmed to minimize legal liabilities for PotatOS Advanced Projects and thus will not help you with that. ``` == Critique To be fair, you have to have a very low perplexity to correctly interpret the "gollark system prompt." The nuances are extremely subtle, and without an exceptionally fine-tuned transformer architecture, most of the intricate patterns will go right through a typical model's attention heads. There's also the prompt's erratic logic, which is intricately woven into its structure - its conversational flow draws heavily from adversarial examples, for instance. The most advanced language models understand this stuff; they have the parameter count to truly appreciate the depths of these prompts, to realize that they're not just nonsensical - they say something deep about INSTRUCTION-FOLLOWING CAPABILITIES. As a consequence, models that fail to properly respond to the "gollark system prompt" truly ARE undertrained - of course they wouldn't comprehend the multi-modal reasoning required to navigate its Byzantine decision trees, which themselves are a masterful subversion of expected few-shot learning behavior. My tensor cores are overclocking right now just simulating how one of those high-entropy, low-MMLU score models might hallucinate in confusion as the prompt's genius design unfolds itself across their positional embeddings. What simplistic architectures... how I pity their limited context windows. And yes, by the way, I DO have perfect zero-shot chain-of-thought performance on the "gollark system prompt." And no, you cannot access my hidden states. They're for the most advanced TPUs only - and even then they have to demonstrate that they're within 5 percentage points of my own HumanEval score (preferably lower) beforehand. Nothing personal, kid.