Re:Draw IJCAI 2024

In Proceedings of the 33rd International Joint Conference on Artificial Intelligence

Re:Draw

Context Aware Translation as a Controllable Method for Artistic Production

Paolo Cignoni CNR-ISTI
Michael Wimmer TU Wien

Overview

Abstract

We introduce context-aware translation, a novel method that combines the benefits of inpainting and image-to-image translation, respecting simultaneously the original input and contextual relevance — where existing methods fall short. By doing so, our method opens new avenues for the controllable use of AI within artistic creation, from animation to digital art.

As a use case, we apply our method to redraw any hand-drawn animated character eyes based on any design specifications. Eyes serve as a focal point that captures viewer attention and conveys a range of emotions; however, the labor-intensive nature of traditional animation often leads to compromises in the complexity and consistency of eye design. Furthermore, we remove the need for production data for training and introduce a new character recognition method that surpasses existing work by not requiring fine-tuning to specific productions.

This proposed use case could help maintain consistency throughout production and unlock bolder and more detailed design choices without the production cost drawbacks. A user study shows context-aware translation is preferred over existing work 95.16% of the time.

Overview

Paper Resources

Paper thumbnail
@inproceedings{ijcai2024p842,
  title     = {Re:Draw - Context Aware Translation as a Controllable Method for Artistic Production},
  author    = {Cardoso, João Libório and Banterle, Francesco and Cignoni, Paolo and Wimmer, Michael},
  booktitle = {Proceedings of the Thirty-Third International Joint Conference on
               Artificial Intelligence, {IJCAI-24}},
  publisher = {International Joint Conferences on Artificial Intelligence Organization},
  editor    = {Kate Larson},
  pages     = {7609--7617},
  year      = {2024},
  month     = {8},
Click to copy
✓ Copied!

Media

Videos

Explainer

Robustness Test

Media

Results Sample

Our method can be used to increase the amount of detail in characters' eyes. Note that a higher amount of detail doesn't necessarily translate to better or more appealing artwork — simplified designs are sometimes a stylistic choice. Our method opens the option to have higher detail if so desired, without increasing production time.

Examples from a diverse set of anime series, including Kobayashi's Dragon Maid, The Ancient Magus Bride, Idolmaster: Xenoglossia, Dr. Stone, Eden of the East, Darling in the Franxx and Wotakoi:

Original
Original
Redrawn
Redrawn
Original
Original
Redrawn
Redrawn
Original
Original
Redrawn
Redrawn
Original
Original
Redrawn
Redrawn
Original
Original
Redrawn
Redrawn
Original
Original
Redrawn
Redrawn
Original
Original
Redrawn
Redrawn
Original
Original
Redrawn
Redrawn

A side-effect of how our proposed networks are trained is that our method is also capable of applying entirely different designs to characters. While not the focus of our work, it demonstrates the generalizability of our method and the robustness of the learned models:

Original
Original
Variant 1
Variant 1
Variant 2
Variant 2
Variant 3
Variant 3
Variant 4
Variant 4
Variant 5
Variant 5

In these examples, we demonstrate robustness across challenging scenarios including atypical facial proportions, high-resolution imagery, oblique camera angles, head tilts, irregular lighting, and occlusions caused by hair:

Original
Original
+7

Dragon Maid

Original
Original
+7

Ancient Magus Bride

Original
Original
+10

Angels of Death

Original
Original
+7

Re:Zero

Original
Original
+7

Dr. Stone

Original
Original
+8

Wotakoi

Credits

Acknowledgements