It pretty much does indicate that though and there isn't a GPT4.5 turbo (yet), there's a turbo iteration of GPT4 and like I said those are public benchmarks - GPT4 turbo for example scores 92.5% on GSMK8, 54% on MATH and 73.17 on HumanEval see here:
The evolution of AI language models is revolutionizing how we interact with technology. Among the latest advancements are Google’s Gemini 1.5 Pro and OpenAI’s GPT-4 Turbo. This article delves into a detailed comparison, shedding light on their capabilities, architecture, and potential impact...
bito.ai
And if you look at the anthropic's results GPT4 scores 92% 52.9% and 67% respectively for each of those, so turbo isn't a huge improvement.
But then if you look at Claude 3 it scores 95% 60.1% and 84.9% respectively for those tests.
No that doesn't follow, there have been plenty of other newer models released that didn't beat GPT4, beating GPT4 across a range of tasks is quite an achievement.