Why is latency and bandwith separated? If latency is delay of data, and bandwith is data over time, shouldn’t one affect another? If it takes 5s for a car to start, and 10s to reach it’s destination, then the average speed would be spread out across the whole 15s. Should this not be the same?
Sending data is actually a two way communication. Computer 1 sends a package of data to computer 2. Computer 2 then sends an acknowledgement to computer 1 to say that it’s received the package of data, and it’s ready for the next. Latency is a measure of how fast this round trip happens. Bandwidth is a measure of how much data can physically be shoved down a cable in a set amount of time.
A high latency can mean computer 1 can’t send the data as fast as its capable of because it’s spending a long time waiting for acknowledgements from computer 2.