Why is the native picture (fig 1) in grayscale? or more generally why is black and white the default of signal processing? Is it just because black and white are two opposites that can be easily discerned?
It's not really grayscale. The output of an image sensor integrated circuit is a series of voltages read one after the other, that could be from –0.3 to +18 volts for example, in an order specific to the sensor's red, green and blue "pixels" arrangement. The native picture (fig 1) is the result of converting a sensor's output voltage to a series of values from black (let's say -0.3 volts for example) up to white (let's say +18 volts for example) while ignoring if they are from a red, a green or a blue image sensor "pixel".
The various "raw" camera image formats kind of work like this, they include the voltages converted to some numerical range and what each "pixels" represents for a specific camera sensor setup.
They’re known as DNs, or digital numbers. Thom Hogan’s eBooks do a phenomenal job of explaining the intricacies of camera sensors, their architecture, processing to JPEGs, and pretty much every aspect of capturing good photos.
The books, while geared toward Nikon cameras, are generally applicable. And packed with high-quality illustrations and an almost obsessive uber-nerd level of detail. He’s very much an engineer and photographer. When he says “complete guide”, he means it.
The section on image sensors, read-outs,
and ISO/dual gain/S&R, etc. is particularly interesting—-and should be baseline knowledge for anyone who’s seriously interested in photography.
It's just a common default choice to represent spacial data that lacks any context on how to interpret the values chromatically. You could very well use a heatmap-like color scheme instead.