I completely agree. Theoretically you could capture and reproduce the entire spectrum for each pixel, but even that is not "true" because it is not the entire light field. But I still think that we can look at the picture on phone in the hand and at the subject just in front, and try to make them as similar as possible to our senses? This looks to me like a big improvement to the current state of affairs. Then you can always say to a critic: I checked just as i took the picture/movie, and this is exactly how the sky/grass/subject looked.