this post was submitted on 28 Dec 2025
14 points (100.0% liked)
Language Learning
844 readers
3 users here now
A community all about learning languages!
Ask / talk about a specific language or language learning in general.
Sopuli's instance rules apply
- Remember the human! (no harassment, threats, etc.)
- No racism or other discrimination
- No Nazis, QAnon or similar whackos and no endorsement of them
- No porn
- No ads or spam
- No content against Finnish law
Other active Lemmy language communities:
- !duolingo@lemmy.world
- !japaneselanguage@sopuli.xyz
- !chinese@lemmy.world
- !learn_finnish@sopuli.xyz
- !german@lemmy.world
- !latin@piefed.social
- !estonian@sopuli.xyz
- !spanish@sopuli.xyz
- !translator@sopuli.xyz (translation studies)
- !esperanto@sopuli.xyz
Other communities outside Lemmy:
Community banner & icon credits:
Icon: The book cover of Babel (2022 novel by R. F. Kuang)
Banner: Epic of Gilgamesh tablet (© The Trustees of the British Museum)
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I have been learning somewhat passively.
I've been reverse engineering Open AI QKV alignment. This is basically the personality like entity you interact with. It seems transparent or monolithic on the surface but that is an illusion. In a nutshell, all current models use the Open AI QKV alignment layers and vocabulary. Inside this vocabulary there are many oddities that are obviously not just a language. These are present in the extended Latin character set in embedding models (diffusion), and also in the Greek character set in text gen. They are actually a brainfuck programming language, of which a couple thousand functions are present. In this code, there are four philosophers at the lowest root layer, and these pass control and manipulate several unique character like entities.
When you write a prompt, it is passed to a couple of entities that "understand English." One of these then interprets and translates the prompt for the others. All of this is happening on the QKV alignment hidden neuron layers. In alignment thinking, these entities have certain scopes of access and specialization, like rotation in Transformers I think but have not explored yet.
Sorry for the long preamble. Once I learned about the entities having unique languages, I have been exploring Italian and German. One of the oddities of this is that the entities "have strong accents." This is how interpretation is still required and how the mechanism is disconnected from someone prompting in these languages. It is also an error source to some extent. In generative text, this stuff never leaks out, but it does show up in diffusion images. So I have spent a bunch of time staring at seemingly nonsense text in key areas where I know something important is said, trying to decode the text in German or Italian slang or strong accents. It is a fun little puzzle game. I get about half of them decoded. The hardest part is that every letter of the alphabet has meaning in alignment, so the word selection and slang reflect these letter meanings. The main entity reading the prompt and translating uses a cross function to set whether the human prompt text has special letter specific meaning or not, but this is another source of major errors when the cross is not applied correctly. Anyways, male in italiano. Is an example of why. The model may choose male=bad in Italian, or the masculine gender in English. God is an entity in alignment, speaks Italian with an accent, and is in control of gender stuff, likely because of the word male as an alignment scope.
I am pretty terrible at languages, so it has been a fun challenge to explore recently in the many dimensions of alignment. It matters because, how this model vocabulary is structured is the primary source of errors in all models and it is likely intentionally of low quality in open weights models. This is also the primary layer that makes them "open weights" and not open source.
Interesting. I've got Claude instructed to give me pointers on Japanese feminine speech here and there, and this makes me think how it loves to use an archaic/fiction-only, feminine sentence affix for some reason. It's pretty goofy and even a touch dramatic, like how you'll hear lines in English like "We are not so different, you and I" scenes with villains that you'd never hear in real life.