Report finds newer inferential models hallucinate nearly half the time while experts warn of unresolved flaws, deliberate deception and a long road to human-level AI reliability
Yeah, I think that workarounds with o3 is where we’re at until Altman figures out that just saying the latest oX mini high is “great at coding” is bad marketing when it can’t accomplish the task.
I don’t quite understand why o3 for coding? Do you mean for code architecture or something? Like creating apps? Why not use a better model if its for coding?
Yeah, I think that workarounds with o3 is where we’re at until Altman figures out that just saying the latest oX mini high is “great at coding” is bad marketing when it can’t accomplish the task.
I don’t quite understand why o3 for coding? Do you mean for code architecture or something? Like creating apps? Why not use a better model if its for coding?