I assume you mean a binary representation of a number.
The "least significant bit" (usually the one to the far right but in some languages it has another placement) is "ones"
the next most significant bit are the twos
The third most significant bit are the fours
etc.
So if your number is 37
there is one 32 (the sixth most significant bit)
no 16's (the fifth most significant bit)
no 8's (the fourth most significant bit)
one 4 (the third most significant bit)
no 2's (the second most significant bit)
one 1 (the least most significant bit)
if we are to fill an 8 bit "word " we get:
0010 0101
Chat with our AI personalities
28-bits
The largest binary number that can be expressed with 16 bits is 1111111111111111, which is equivalent to 65,535 in decimal. This number uses all 16 bits set to 1. In general, for an n-bit binary number, the maximum value is (2^n - 1). Thus, for 16 bits, it is (2^{16} - 1 = 65,535).
As quoted from Google Books, "Word size refers to the number of bits that a microprocessor can manipulate at one time."
A nibble consists of 4 bits. The binary number "1011 1101 1001 1110" contains 16 bits in total. Therefore, to find the number of nibbles, you divide the total number of bits by 4, which results in 4 nibbles.
There are 16 decimal numbers that can be represented by 4-bits.