That is the beauty of the model, that it can understand what a unicorn looks like, and then know roughly where to place the features. Currently we're seeing poor results due to safety tuning but in time, it'll be very interesting to see how it improves.
There are a few factors at play here: knowing what a unicorn looks like, knowing the different areas of a unicorn, being able to translate that into a 2D space, and being able to form the connection between code (language) and appearance.
How though? I have tried to get got to generate vector art and my attempts usually failed with the model getting stuck into recurring patterns with no structure. I'm surprised anyone was able to get it to generate something.
In particular, I cannot understand how the models can properly understand concepts such as spatial relations without being able to 'see'
There are a few factors at play here: knowing what a unicorn looks like, knowing the different areas of a unicorn, being able to translate that into a 2D space, and being able to form the connection between code (language) and appearance.