These days, artificial intelligence is all the rage and the latest kids on the block are
transformer models, and with them language models such as GPT-2 and most recently
GPT-3.
These things are pretty cool, with them you can create texts that are often indistinguishable from a text written by a human.
This leads me to wonder if these models could be used to improve on BIP039. I wouldn't feel confident in using BIP039 to memorize a private key. The generated words are very random and uncorrelated, it seems to me the risk I forget them, or forget their order, is very high.
That's why I wonder if it's possible to use a private key to seed the word completion algorithm of a language model.
That doesn't seem obvious to me, as the probability distribution for each new word is not uniform, so we can't just pick say the residue modulo ten to pick one word out of ten possible ones.
Yet, my intuition tells me it has to be possible to sample the words while respecting the probability distribution. I just can't quite figure out how exactly.
If someone sees how this can be done and can provide a demonstration with GPT-2, that would be fantastic.
PS. I mean think about it : GPT-3 can generate poems, for instance. Wouldn't it be cool if you could encode a bitcoin private key in a poem ?