Autogollark is an [[emulation]] or primitive [[beta upload]] of [[gollark]] using a proprietary dataset of dumped [[Discord]] messages, [[semantic search]] and [[in-context learning]] on a [[base model]]. Currently, the system uses [[LLaMA-3.1-405B base]] in BF16 via Hyperbolic, [[AutoBotRobot]] code (though not presently its bot account) as a frontend and a custom [[PGVector]]-based search API. While not consistently coherent, Autogollark is able to approximately match personality and typing style.
Autogollark is much [[safer]] than [[instruction-tuned]] systems optimized based on human feedback, as there is no [[optimization pressure]] for user engagement or sycophancy.
* One proposal: use internal finetune to steer big model somehow. Possibly: use its likelihood (prefill-only) to evaluate goodness of big model output wrt. gollark personality, and if it is too bad then use finetune directly.
* Autogollark 0.1 was the initial RAG system and ABR interface. It used LLaMA-3.1-8B run locally. Autogollark 0.0, which is not real, used only gollark messages.
* Autogollark 0.2 replaced this with LLaMA-3.1-405B.
* Autogollark 0.3 upgraded the dataset to contain longer-form conversations than Autogollark 0.1.
* {Petulant nonresponse - due to ratelimits in the LLM API, Autogollark will under some circumstances not respond to messages, with error messages being consumed and not exposed to [[users]]. This can be interpreted by [[credulous]] users as choosing not to respond, though this is not believed to be possible (other than cases like responding with `.`, which ~~has not been observed~~ does not appear to be associated with annoyed states).
* Memorizing links: Autogollark directly experiences past message chunks in context, granting perfect recall of a small amount of memory at once. This has memorably included [[Autogollark/Closed Individualism Incident|YouTube videos]] repeated with no context.
* {Limited self-improvement attempts: when told about this architecture, Autogollark will often complain about various limitations and propose vague ideas for improvements.
* Also, Autogollark has previously claimed to be working on LLM chatbots.}
* {Inconsistent inference of own status as a language model chatbot, possibly based on seeing the name "autogollark". Often, Autogollark assumes use of GPT-3.
* {"Self-reset" from attractor states (e.g. the [[As An AI Language Model Trained By OpenAI]] basin, all caps, etc) after some time passes, because of messages having `HH:MM` timestamps.
* This is mostly specific to the 405B model; Autogollark in failover to the 8B internal model usually does not do this.
* {For somewhat [[Waluigi Effect]]-related reasons (past context is strong evidence of capability but weak evidence of incapability), Autogollark has some knowledge [[gollark]] does not, and can speak in a much wider range of languages.
* "I, being more than just myself, actually can talk about both Galois theory and obscure poetry by Yeats."}