Love it when two worlds meet and make a beautiful art baby.
What if we turn Google Maps into anime? 🤔 I tried it using Stable Diffusion and ComfyUI. Here is the workflow: - Checkpoint model: meinamix_meinaV11 - Positive Prompt: day, noon, (blue sky:1.0), clear sky - Negative Prompt: (worst quality, low quality:1.4), (zombie, sketch, interlocked fingers, comic) - Resolution: 768 x 512 - ControlNet model: control_vllp_sd15_canny Depending on the Google Maps location, I add a country or city name in the positive prompt (e.g. Japan, New York, Paris, etc.). I used toyxyz's custom webcam node to capture a section of the screen and plug that into a ControlNet canny model. KSampler: - seed: 1 - control_after_generate: fixed - steps: 15 - cfg: 4.0 - sampler_name: euler_ancestral - scheduler: normal - denoise: 1.00 It is possible to optimize this further and make better and faster generations. Perhaps by using an LCM-LoRA, StreamDiffusion, TouchDesigner, or a model based on SDXL-Lightning. Video music from Uppbeat dot IO Space Journey by Hartzmann #AI #ArtificialIntelligence #GenerativeAI #StableDiffusion #ComfyUI #anime
VR, AR, Unreal Engine, Unity 🥽 Currently jacked into the Matrix.
1moThanks for sharing!