Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

There's a free addon for free Krita that did pretty much that when I tried it, last year.

The glaring issue with it back then was that unlike an LLM that can be understanding of what you try to explain and bit more consistent the diffusion models ability to read and understand your prompt wasn't really there yet, you're more shotgunning keywords and hope the seed lottery gives you something nice.

But recent image generation models are significantly better in stable output. Something like qwen image will care a lot more about your prompt and not entirely redraw the scene into something else just because you change the seed.

Meaning that the UI experiments already exist but the models are still a bit away from maturity.

On the other hand, when looking at how models are actually evolving I'm not entirely convinced we'll need particularly many classically trained artists in roles where they draw static images with some AI acceleration. I expect people to talk to an LLM interface that can take the dumbest of instructions and carefully adjust a picture, sound, music or an entire two hour movie. Where the artist would benefit more by knowing the terminology and the granular abilities of the system than by being able to hold a pencil.

The entertainment and media industry is worth trillions on an annual basis, if AI can eat a fraction of that in addition to some other work-roles it will easily be worth the current valuations.



Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: