gpt5-main (the non-thinking model) has (still unsolved, I guess they don't wanna) instruction following regressions.
Just out of interest I ran your image with the same instruction through a small gemma3 distill:
ggml-org/gemma-3-4b-it-GGUF:Q4_K_M using llama.cpp server:
I don't know if the answer is in any way correct, but this is all runnable with minimum memory (this particular one should run with 4GB memory), locally.
gpt5-main
(the non-thinking model) has (still unsolved, I guess they don't wanna) instruction following regressions.ggml-org/gemma-3-4b-it-GGUF:Q4_K_M
using llama.cpp server: