Wikipedia

Bit time

Also found in: Dictionary, Encyclopedia.

Bit time is a concept in computer networking. It is defined as the time it takes for one bit to be ejected from a network interface controller (NIC) operating at some predefined standard speed, such as 10 Mbit/s. The time is measured between the time the logical link control sublayer receives the instruction from the operating system until the bit actually leaves the NIC. The bit time has nothing to do with the time it takes for a bit to travel on the network medium but has to do with the internals of the NIC.

To calculate the bit time at which a NIC ejects bits, use the following:

 bit time = 1 / NIC speed 

To calculate the bit time for a 10 Mbit/s NIC, use the formula as follows:

 bit time = 1 / (10 * 10^6) = 10^-7 = 100 * 10^-9 = 100 nanoseconds 

The bit time for a 10 Mbit/s NIC is 100 nanoseconds. That is, a 10 Mbit/s NIC can eject 1 bit every 0.1 microsecond (100 nanoseconds = 0.1 microseconds).

Bit time is distinctively different from slot time, which is the time taken for a pulse to travel through the longest permitted length of network medium.


This article is copied from an article on Wikipedia® - the free encyclopedia created and edited by its online user community. The text was not checked or edited by anyone on our staff. Although the vast majority of Wikipedia® encyclopedia articles provide accurate and timely information, please do not assume the accuracy of any particular article. This article is distributed under the terms of GNU Free Documentation License.

Copyright © 2003-2025 Farlex, Inc Disclaimer
All content on this website, including dictionary, thesaurus, literature, geography, and other reference data is for informational purposes only. This information should not be considered complete, up to date, and is not intended to be used in place of a visit, consultation, or advice of a legal, medical, or any other professional.