Skip to content

Gerry's World

A glimpse into my life

AI Generated Children's Book

December 24, 2023

Introduction

This Christmas, I wanted to create an illustrated children’s book for a <1 year-old using generative AI tools for 2 primary reasons:

  1. I was eager to try out the new generative AI tools to get a better feel for their capabilities, strengths, weaknesses, etc. and
  2. I thought it would be fun and cute.

You can view a low-res preview of the final book here, or full-res (but not cropped-to-book-trim) images here!

Process / Procedure

My plan was to use the following pipeline:

  1. Generate a story using ChatGPT-4, defining both the text and a description of the illustration on each page.
  2. Generate illustrations using DALL-E 3, using the descriptions from step 1.
  3. Make any necessary edits to the illustrations using generative fill tools.

Step 1: Generate a story using ChatGPT-4

Generating the story was the easiest part! ChatGPT excels at this type of task.

You can view the transcript here.

Step 2: Generate illustrations using DALL-E 3

This step was rather difficult - particularly trying to get images with all the right characters doing the right things. Trying to get one or two characters right was feasible, but to get e.g. 7 characters all with the right gender, age, and race (e.g. for the cover) was next to impossible. Ultimately, just created a handful of variations until I got one reasonably close, then used generative fill tools to make the necessary edits removing the extra characters.

The artistic style and character appearances were also not consistent across pages, but I didn’t bother trying to fix this. Nevertheless, this is certainly an identifiable weakness of DALL-E 3. As I understand, inter-image consistency is an area of active research, so I expect this will improve in the future to some degree.

Although I gave a decent amount of effort towards the cover and the first couple pages (first impressions matter!) I hardly did any variations/retries/edits on the later pages.

Also note that DALL-E 3 can natively generate 1792x1024 images, but I compress them in this article to save on website hosting costs.

To generate the images, I gave a prompt first listing textual descriptions of all the characters, then pasting the story text from step 1, and finally asking for particular images. For example,

Can you generate an 11x8 illustration for each page of this children’s story?

Aven is a very happy, lively 9-month old asian-caucasian baby with parents Alice (Asian, 34F) and Bob (tall caucasian with reddish-brownish hair and a small amount of facial hair, 34M). Carol is …

Page 1

  • Text: “Aven’s day at the zoo, so much to explore, with family around, who could ask for more?”
  • Scene: Aven in her stroller at the zoo entrance, surrounded by her family, all excited and ready for adventure.

Page 2

First, before generating the images, please list every member of the family with race, age, and a rough appearance description for Jim.

Start by creating an 11x8 children’s-book style illustration for pages 1-4.

* … DALL-E generates images … *

Good, now continue with pages 5-8

The cover didn’t need to be specific about the plotline:

Aven is a very happy, lively 9-month old asian-caucasian baby with parents …

Can you generate 4 variations of a 2:1 illustration for the cover of Aven’s children’s story book entitled “Aven’s Wild World of Wonders”? It’s about her visiting the zoo.

This would give a result like this: from which I then “photoshop-ed” out the extra baby characters at the bottom in Step 3.

Step 3: Editing using AI

There were 3 primary editing operations I did:

  1. Generative fill to add/remove/change foreground characters
  2. Upscaling / Super-resolution
  3. Generative edge fill / expand
  4. Add text (not using AI)

I tried a handful of AI generative fill tools with wildly varying qualities, but they were all extremely underwhelming.

I suspect one major reason for the poor results is that I was trying to use these tools on illustrations whereas I think they were probably intended/trained to be used on photorealistic images. Here is a nightmare-inducing example where I tried to generative-fill a character’s face, which turned out disproportionately sized, inconsistently lit, and stylistically way off:

Nevertheless, removing characters (and replacing them with background) worked “ok”, even though there were some architectural incongruity artifacts, it was more-or-less passable. For example:

I tried using:

  • Photopea “Magic Replace” - replace a selection of an image by “remove”-ing foreground or replacing with a textual description.
    • This worked pretty poorly, but was ok for removing characters and was leagues better than Picsart
    • Photopea is, in general, a pretty nice tool that I had heard of before as a Photoshop alternative/knockoff.
  • Picsart “AI Replace”
    • this is completely unuseable. Results were atrocious and nightmare-inducing.

I didn’t get to try out Photoshop’s generative fill because apparently Georgia Tech doesn’t give us students licenses :(

I also didn’t try out other “magic eraser” (e.g. Google Pixel) because I assumed those would be heavily optimized for photorealistic images and not illustrations.

This was easy and worked well. Using picsart “batch” processing, I did “AI enhance”, then dumb “resize” to 4096 wide, then “AI enhance” again. There’s probably a better way but this worked pretty well and was easy.

Some images were framed un-centered, and the cover images needed to be extended so that they could be folded over the hard cover to be glued.

Although I tried using picsart for the generative expand / edge fill to do these, it didn’t really work well, so I ended up just doing non-AI edits. The un-centered images I just lived with or cropped, and the cover I just mirrored the edges manually using Mac Preview (generic basic image viewer/editor). The cover mirroring you can see in the photo at the top of this page.

I used Photopea to add text, but in retrospect this would have been way easier in MS Powerpoint because I’m so much more familiar with Powerpoint. The text is really difficult to read in many places because I was too lazy to figure out how to get text color backgrounds but it would have been trivial with powerpoint (just be sure to set page setup to the same resolution as the images). Cropping & edge mirroring would have also been easier with powerpoint. Oh well. Live and learn.

Concluding Thoughts / Reflections

In total, making the book took about 7 hours:

I feel like 2 issues were really clear:

  1. Consistency between pages
    Although it’s kind of neat for each page to have a different style, I think it’s usually undesirable to have each page look like it’s by a different artist. Also the character consistency between pages - the characters look different on each page. I think these are active areas of development.
  2. Getting the right characters on each page
    If you try to specify one or two people it can do it, but trying to get like 7 people with the right descriptions on a page is impossible. It would always throw in random extra people or missing people or wrong descriptions or something. I combined it with generative fill to photoshop-out people which took the most time. Also, probably a function of the softwares I was using but generative fill sucks compared to the initial generation. Impossible to put people in and still look good, but could take people out with reasonable quality.

Personally, I don’t think the book is winning any awards, but I do think this was a really fun/funny project. I like to laugh at the oddities, like one of the early pages I generative-filled Bob’s face so it ended up looking “photorealistic” instead of illustration; all the random Asian people; the one page where Aven’s suddenly like 4years old then back to 1yr old on he next page; etc.

I placed the order as a same-day hardcover on walmart photos, but it didn’t turn out well (the hard cover didn’t seem professionally folded). Last year I used shutterfly for a photobook which was great quality and half the price, so in the future I think I’ll use a similar online service - just keep in mind online will take ~2 weeks to arrive.