I searched for similar questions here and none of them really gave the answer I’m looking for.
I understand we have the concept of encoding where in the ascii scheme, the letter A, for example, is encoded to 01000001, because humans agreed on this to be so, *but where actually is that A* when the binary calls for it, if physically in the cpu, for this example we just have 8 microscopic transistors, 2 holding a charge and 6 without a charge that we point to and say “That there means A…but it also means 65 in decimal”? Clearly there’s some level of abstraction here I’m not aware of.
Does 01000001 actually just correspond to a standardized pixel arrangement in the shape of A that is only actually rendered by the hardware in the situation where it’s called to be printed on a screen?
In: 0
Latest Answers