Generative Art Using Machine Learning

Yesterday OPENAI released a showcase of examples demonstrating image generation from text called DALL-E. The complexity of never seen examples is impressive. The quality of illustration and images makes you wonder where this will be used when anything you can think up, can be convincingly generated.

Sleepy Owl Illustrations are not SuperbOwl but not too bad.

Art, design, fashion, or possibly just to deceive? Still I’d love to see more categories or variations. Possibly an Open API

Dendrocreation

The examples are uncanny and familiar while still being “unique” or new. The low effort to produce many satisfactory results will be enough for many, especially if the cost is low.

It’s easy to point to jobs or industry that this could one day replace however I see more people working with machine learning. Human preference should steer results. A fleshy discriminator.

I joked that I’m now a Canadian fashion designer.

And just the day before another generative adversarial network, or GAN was shown called Taming transformers. This technique uses segmentation trained from landscape images to generate high resolution landscapes.

Generate van art

Cryptoart seems to be booming for well known generative artists at the moment and tools like these now eclipse some of the earliest GAN art sold at auction.

Previously other trained networks such as StyleGAN and BIGGAN peaked my interest in generating art with code. Neural networks for image generation have improved in leaps and bounds. Better UI to allow anyone to explore makes this really accessible and easy.

Exciting times.