Interview: Maui Mauricio on AI-Assisted Art Creation
INTERVIEWER: Maui Mauricio is a creative producer and director from the Philippines. He loves experimenting with new technology, and that has led him to dabble with AI content generation.
This week, NiftyZone had the opportunity to speak to Maui about his experience, art-work and thoughts about AI’s role in art creation and how this technology can add to his professional work.
INTERVIEWER: Hi Maui, can you let our readers know a little about yourself?
My name is Maui Mauricio, I’m a professional video editor, producer and occasional director from Manila, Philippines. My passion has always been new technology, partly because my career as a filmmaker began almost the same time digital filmmaking did. My college classmates and I started out by making music videos, a field which valued experimental production and post production techniques, and through that I discovered my love for unusual methods of shooting footage as well as editing the said footage.
INTERVIEWER: So when and how did you get started in using AI for content creation?
I first got into AI generated imagery not with the intent of creating art but rather through its use as a tool in my field of work. I became fascinated with the development of AI-enhanced image and video upscaling, as most of our earlier work, shot in the then-standard resolution of 720×480, weren’t ageing very well and looked really terrible when played back on high resolution screens. For my first experiment, I decided to upscale one of my favorite past projects, a short film directed by one of my close friends back in college for his final thesis. I had to go back to the actual tapes themselves, upscale the recaptured footage, and reedit the entire film from scratch. I was really happy with the results, but it wasn’t really until much later on that the use of AI in generating images entered my awareness and interest.
I never really went into AI art with the actual intent of creating art – as I said above, for me it was a means to an end. I’m a big Darren Aronofsky fan, and I wanted to emulate what he did for Waking Life and A Scanner Darkly, and AI provided a potential avenue for that. AI tools were also starting to become introduced in the applications that I used daily: Adobe introduced Roto Brushing in After Effects, a subject separation tool which eventually became my most commonly used tool. I discovered Topaz Gigapixel and Video Enhance early on as well, and slowly my fascination for what AI can do grew.
But it wasn’t until the pandemic – that dark era where most of us were stuck at home with nothing to do – that I began to experiment with AI image generation. Runway.ML made their text-to-image model available to the public and I jumped on that pretty quickly. The images were only vaguely reminiscent of the prompts I provided, but I definitely saw the potential. I also tried Wombo at the start of the year, but I only really got into it (and work does get in the way most of the time haha) when Midjourney popped into the scene last July. That’s when I started doing lots of experiments, testing out different styles and subjects (just like everyone else, I suppose).
I’ve moved on to training my own models and doing animation since then, the public release of Stable Diffusion 1.4 being a catalyst for so many innovative uses of the fledgling technology, and my intent is to keep up with the trends and create newer and better pieces (and even help develop: I’m a patron/sort-of beta tester of a couple of developers’ efforts) until there comes a time that AI generated art becomes a more commonly used asset within the fields that I work in.
Training models, in particular, is something I’m very keenly interested in, as I see it a big step towards the creation of true novel AI art by way of unique personalities and art styles. I’m currently at the stage of testing out Stable Diffusion’s inherent racial biases when it comes to skin color and facial features. SD tends to handle white faces very well, as expected, so I’m partial towards testing non-caucasian faces (southeast asian ones in particular) to find ways of breaking that bias. It’s interesting to note that the model that I trained on my face has “broken” somewhat: Prompts calling for generic handsome men OR beautiful women are generating Asian faces now, and they don’t necessarily look like me.
INTERVIEWER: Any guess about the future of the role of AI in the Arts?
AI art is in its infancy, definitely, and things are only going to accelerate once people get a hold of NVIDIA’s newest cards, which would mean bigger, more complicated compositions produced at a much faster rate. I think that there will be an eventual fall-off of interest in terms of the general public, i.e. casual users who currently dominate the conversation (at the same time however, there will be a huge uptick in the number of professional artists who began with the use of AI). Concurrent to that there is a growing rise in interest with the technology in the advertising industry (I’m actually doing something AI-related for the Singapore market currently) and while I can’t speak for the art, advertising and film communities, I think that AI will be integrated into the workflow of the typical graphic artist/illustrator/animator/content creator just as much as, say, Adobe Photoshop, Blender, or Unreal engine once the tech really matures. It’s like the early days of digital filmmaking all over again, just much louder and much, much faster.
INTERVIEWER: What’s next for you (in the context of AI Art)?
While I’m refining my current process (model training, prompt creation), I’m waiting for the next developments that I can pounce on and test out. We’re still in the early days of AI art and animation generation, and as much as I would want to fantasize about what we can do with the technology tomorrow, I find it more exciting to just stay in the moment, have a very open mind, and keep prompting.