Then why doesn't the emulation code use this "proper CPU rendering"?...