Comment by georgeecollins
5 years ago
You probably know this 100ms = 10 FPS. What kind of display shows video at less than 10fps? Game engines aren't always synced to frame rates, particularly simulations. But a simulation that updates every 0.1s isn't great for fidelity.
A 30 fps game could go through a complete loop, updating everything: object positions, inputs in 33ms. At 60 fps assuming everything is synced to frame rate that would 16 ms.
I was asking for the commenter's source of information so I didn't have to guess what he or she meant. It's possible to make a game that doesn't respond a user's input in less than 200ms, but why would you? You don't need to be making a technical tour de force to respond in 16-33ms.
I was commenting on how the TV can add latency/'display lag', not that it only shows a frame every 100ms. TVs have gotten much better[0] but input lag can be high with cheap TVs sold 5-10 years ago.
0: https://displaylag.com/best-low-input-lag-tvs-gaming-by-game...
That makes more sense. I am sorry I misunderstood and thank you for explaining.