Transforming words into drawings is the latest in the field of artificial intelligence drawing, we will talk with you about two very special projects in this field and what images generated using only words look like, and whether this new technology may pose a threat to graphic designers and artists in general

“Painting with words”, this time we are not talking about one of the poems of the poet Nizar Qabbani, but about a new mutation of artificial intelligence that will allow you to create pictures and artistic paintings even if you do not have the skill of drawing or design. Just enough to have a fertile imagination and leave the rest to artificial intelligence.

For a long time, the field of art and creativity remained one of the areas that we thought that artificial intelligence would not be able to easily enter and revolutionize it, but we were wrong. We are on the cusp of a new era of technology-enhanced art and creativity in which Artificial Neural Networks (ANNs) are producing an entire work of art from scratch. By converting what you think of words – whatever they are – into realistic drawings and pictures or paintings that no one has expressed before. Will the rumors be believed and artificial intelligence replace humans? Should artists and graphic designers worry about their future careers?

What are Artificial Neural Networks (ANNs) and how do they work?

Artificial Neural Networks, whose name and structure are inspired by the human brain, is a deep learning model that simulates the way neurons work in the human brain. It plays a large role in machine learning. Artificial neural networks can be trained based on certain speech recognition algorithms . or text data or visual images.

Speech and image recognition tasks take minutes versus hours when compared to manual recognition that occurs by humans, these networks support artificial intelligence and form the basis for a large number of developments that have occurred in it in recent years, including systems for converting words into images, and is an algorithm Google search is one of the most famous artificial neural networks in the world.

An artificial neural network contains three or more interconnected layers, the first layer consisting of input neurons, these neurons send data to the deeper layers, which in turn send the final output data to the last output layer. 

The most famous applications of artificial intelligence to convert drawings into words

The services of converting graphics to words using artificial intelligence depend on two main factors, the first is the clip, which depends on the artificial intelligence knowing what the elements that you describe to it individually look like. The artificial will be aware of the shape of the bear, the ride, and the spaceship. And then we move to the second factor, which is “Diffusion”, which depends on the method of combining these elements together in one image and improving them by deleting the “Guassion Noise”


In January 2021, OpenAI developed an artificial intelligence prototype called DALL-E capable of converting words into graphics. So we saw some interesting results like the following picture when asked to draw:

“Picture of a woodcut postage stamp celebrating a noble dog explorer using a telescope”

Drawing with artificial intelligence

But with the unveiling of the improved system DALL-E 2 in April 2022 that was designed on the first factor only “Clip” everything changed. It astounded everyone, including artists and graphic designers with its ability to transform natural language into images, imagination is no longer the ultimate limit of creativity The creativity in DALL-E2 is beyond imagination.

This system demonstrated ingenuity in dealing with long and detailed sentences, and gave results that exceeded expectations, and compared to its predecessor, DALL-E, it was able to outperform it by producing more realistic images with four times the accuracy, but so far this system has not been made available for public use. The best way to understand the amazing ability of these models is to simply look at some of the images that can be created.

DALL-E 2 creates more than one image of the same description, all of which appear with results that will need days of work to reach them with human capabilities, these are some examples that the system generated from just a text description.

“An astronaut riding a horse in a realistic style”