Midjourney just changed the generative image game and showed me how comics, film, and TV might never be the same

Midjourney, the Generative AI platform that you can currently use on Discord just introduced the concept of reusable characters and I am blown away.

It's a simple idea: Instead of using prompts to create countless generative image variations, you create and reuse a central character to illustrate all your themes, live out your wildest fantasies, and maybe tell a story.

Up until recently, Midjourney, which is trained on a diffusion model (add noise to an original image and have the model de-noise it so it can learn about the image) could create some beautiful and astonishingly realistic images based on prompts you put in the Discord channel (“/imagine: [prompt]”) but unless you were asking it to alter one of its generated images, every image set and character would look different.

Now, Midjourney has cooked up a simple way to reuse your Midjourney AI characters. I tried it out and, for the most part, it works.

Image 1 of 3

Midjourney AI character creation

I guess I don’t know how to describe myself. (Image credit: Future)
Image 2 of 3

Midjourney AI character creation

(Image credit: Future)
Image 3 of 3

Midjourney AI character creation

Things are getting weird (Image credit: Future)

In one prompt, I described someone who looked a little like me, chose my favorite of Midjourney's four generated image options, upscaled it for more definition, and then, using a new “– cref” prompt and the URL for my generated image (with the character I liked), I forced Midjounrey to generate new images but with the same AI character in them.

Later, I described a character with Charles Schulz's Peanuts character qualities and, once I had one I liked, reused him in a different prompt scenario where he had his kite stuck in a tree (Midjourney couldn't or wouldn't put the kite in the tree branches).

Image 1 of 2

Midjourney AI character creation

An homage to Charles Schulz (Image credit: Future)
Image 2 of 2

Midjourney AI character creation

(Image credit: Future)

It's far from perfect. Midjourney still tends to over-adjust the art but I contend the characters in the new images are the same ones I created in my initial images. The more descriptive you make your initial character-creation prompts, the better result you'll get in subsequent images.

Perhaps the most startling thing about Midjourney's update is the utter simplicity of the creative process. Writing natural language prompts has always been easy but training the system to make your character do something might typically take some programming or even AI model expertise. Here it's just a simple prompt, one code, and an image reference.

Image 1 of 2

Midjourney AI character creation

Got a lot closer with my photo as a reference (Image credit: Future)
Image 2 of 2

Midjourney AI character creation

(Image credit: Future)

While it's easier to take one of Midjourney's own creations and use that as your foundational character, I decided to see what Midjourney would do if I turned myself into a character using the same “cref” prompt. I found an online photo of myself and entered this prompt: “imagine: making a pizza – cref [link to a photo of me]”.

Midjourney quickly spit out an interpretation of me making a pizza. At best, it's the essence of me. I selected the least objectionable one and then crafted a new prompt using the URL from my favorite me.

Midjourney AI character creation

Oh, hey, Not Tim Cook (Image credit: Future)

Unfortunately, when I entered this prompt: “interviewing Tim Cook at Apple headquarters”, I got a grizzled-looking Apple CEO eating pizza and another image where he's holding an iPad that looks like it has pizza for a screen.

When I removed “Tim Cook” from the prompt, Midjourney was able to drop my character into four images. In each, Midjourney Me looks slightly different. There was one, though, where it looked like my favorite me enjoying a pizza with a “CEO” who also looked like me.

Midjourney AI character creation

Midjourney me enjoying pizza with my doppelgänger CEO (Image credit: Future)

Midjourney's AI will improve and soon it will be easy to create countless images featuring your favorite character. It could be for comic strips, books, graphic novels, photo series, animations, and, eventually, generative videos.

Such a tool could speed storyboarding but also make character animators very nervous.

If it's any consolation, I'm not sure Midjourney understands the difference between me and a pizza and pizza and an iPad – at least not yet.

You might also like

TechRadar – All the latest technology news

Read More

Stability AI’s new text-to-audio tool is like a Midjourney for music samples

Stability AI is taking its generative AI tech into the world of music as the developer has launched a new text-to-audio engine called Stable Audio.

Similar to the Stable Diffusion model, Stable Audio can create short sound bites based on a simple text prompt. The company explains in its announcement post that the AI was trained on content from the online music library AudioSparx. It even claims the model is capable of creating “high-quality, 44.1 kHz music for commercial use”. To put that number into perspective, 44.1 kHz is considered to be CD quality audio. So it’s pretty good but not the greatest.

Stable Audio user interface

(Image credit: Stability AI)

A free version of Stable Audio is currently available to the public where you’re allowed to generate and download 20 individual tracks a month. Each sound bite has a 45 second runtime so they won’t be very long.

Prompting music

The text prompts you enter can be simple inputs. Listening to the samples provided by Stability AI, “Car Passing By” sounds exactly as the title suggests – a car driving by in the distance although it is a little muffled. Conversely, you can also stack on details. One particular sample has a prompt involving Ambient Techno, an 808 drum machine, claps, a synthesizer, the word “ethereal”, 122 BPM, and a “Scandinavian Forest” (whatever that means). The result of this word combination is an ambient lo-fi hip-hop beat.

We took Stable Audio out for a quick spin. We were able to enter one prompt asking the AI to create a fast-paced garage rock song from the early 2000s and it sort of accomplished the goal. The generated track matched the style although it sounded really messy. 

Personal Stable Audio input

(Image credit: Future)

Unfortunately, we couldn’t go any further besides the single input. At the time of this writing, Stable Audio is seeing a huge influx of traffic from people rushing in to try out the model. The developer recommends trying again later or the next day if you’re met with nothing but a blank screen.

There is a catch with the free version – it’s for non-commercial use only. If you want to use the content commercially, then you’ll have to purchase the $ 12 Stable Audio Professional monthly plan. It also offers 500 track generations a month, each with a duration of up to 90 seconds. There’s an Enterprise plan too for custom audio duration and monthly generations. You will, however, have to contact Stability AI first to set up a plan.

Imperfect tool

Do be aware the technology isn’t perfect. The content sounds fine for the most part, however certain aspects will seem off. The mix in that Ambient Techno song mentioned earlier isn’t very good in our opinion. It was like the bass and synthesizer are fighting over what will be the dominant sound, resulting in just noise. Additionally, it doesn’t appear the AI can do vocals. It only does instrumentals. 

Stable Audio is interesting for sure, but not something that should be totally relied on. We should note the company is asking for feedback from users on how to improve the AI. A contact email can found on the official announcement page.

If you plan on utilizing this tech for your own purpose, we recommend checking TechRadar’s list of the best audio editors for 2023 to fix any flaw you might come across. 

YOU MIGHT ALSO LIKE

TechRadar – All the latest technology news

Read More