Again I appreciate the effort. It's clear you're thinking about
the problem in a way that helps understanding.
First: Passwords. If you ask for a good password, the likely answer is
a good password. In isolation it probably is a good password even when
you risk that it has the properties of example passwords in the
training set.
It probably looks like a typical password, but a good password
doesn't exist in isolation. What you really want is a set of
potential passwords with a way to pick uniformly over that set.
A good LLM may now determine that it should for example not sample the
same token again and again, as it learned that that is a easy-to-crack
password.
Maybe yes maybe no. That is not a good way to generate passwords and
it might "understand" documentation that explains that.
You already see, that the password entropy is now just 3^length.
If you would sample uniformly from ALL tokens, you would get a
password generator. But all other parts of the output would also look
like passwords.
Entropy is normally measured in bits, so you need to take the lg
of that, but it's probably more intuitive to talk out the number of
possible outputs, so OK.
What you are saying is that it's not a good password generator,
sure. This is consistent with the article, and what I said. The LLM
can use randomness to generate about 1/4 the entropy you would get
from uniform sampling over the set of all character strings of length
16. In terms of counting that's maybe 2^25 which is just 32 million
passwords which can be bruteforced in certain situations. That's a
big difference from 2^100 which is a quadrillion quadrillions.
So I agree, it's a bad password generator, but it still has access
to randomness, so it can generate passwords they are just not optimal
for their length. (The more troubling part is that it sometimes fails
to do the right thing and might generate some passwords with much high
probability.)
Unique is a bit similar problem. People complained for example that
image generators do not understand "Generate a unique SUBJECT".
As I said, it's not very well defined. Dictionaries have multiple
definitions for a reason. This alone is enough to cause issues, but
it gets more complex by speculating what the LLM is doing. However,
it is an interesting question.
In image generation it may be an image that was tagged as
"So very unique", but if all generations get close the concepts in
that image, none of them is unique anymore. The LLM test before
associated fantasy styles names with "unique".
Mathematically, if they are close, they are still unique. In this
context, what a human probably wants is something semantically
meaningful that is sufficiently far, with some metric, from all other
images that have ever been generated. Of course, without knowledge of
what all other generators are doing, unique, in this sense, is
impossible. So at best you're back to a probabilistic solution. You
need to define a space of images and have the algorithm uniformly pick from that
space. Just playing with a LLM, it seems to do a simple form of
this. It combines a bunch of semantic ideas to create a "unique"
image. On gemini, a clock owl that is on a frozen lightning branch
in a cosmic storm.
These things work by "associations" and when they are primed on
something they often follow predictable patterns.
Yes, there was a blog that talked about how if you played bad chess with an LLM it would also play bad, but if you played a strong game against it, it would improve.
Just read a bit the complaints in the creative writing
communities. Try to google for some stories about Elara and you see
how LLM even fail at creating unique names.
Interesting, and it still shows up in current models. It's probably a
bit of manifest destiny since this issue goes back to 2022. However,
when I ask for a unique name for my sci-fi lead, the model claims to
do something similar to what I described with images and comes up with
some weird names: Vyrith-Esh, Xylanthe-Vane, Kyzant-Nu, Zylpha-Kore.