Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

You're not. I'm surprised at their selections because neither the cooking one nor the beach one adhere to the prompt in very well, and that first one only does because it prompt largely avoids much detail altogether. Overall, the announcement gives the sense that it can make pretty pictures but not very precise ones.


Well, that's nothing new, but it doesn't matter to dedicated users because they don't control it just by typing in text prompts. They use ComfyUI, which is a node editor.


I'd say automatic1111 is more popular. Comfy seems like a rat's nest, unreal shader node flashbacks.


Does this afford better prompt adherence control in some way?


No directly. but it encourages iteration on the same seed and then on specific details rather than just trying different prompts on different seeds from scratch over and over


Sounds to me like it's an issue with their VLM captions creating very "pretty" but not actually useful captions. Like one of the example image prompts includes this absolute garbage:

> Convey compassion and altruism through scene details.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: