Generic views
The principal of generic views in the study of cognition stipulates that the interpretation made by an observer of a distal phenomenon should be such as to not require that the observer be in a special position to, or relationship with, the phenomenon in question. The principal is a fairly general account of the inductive bias that allows an observer to reconstruct distal phenomena from an impoverished proximal datum. This principle has been advanced particularly in vision research as an account of how, for example, three-dimensional structure is extracted from an inadequate two-dimensional projection.
The principal of generic views has been discussed by Richards[1] and Hoffman[2][n 1], and has been given a sophisticated Bayesian formalization by Freeman.
Relation to Bayesian inference
Another expression of the generic views principal is that the inference of distal structure should be such that the inference would remain substantially the same if the "position" of the observer were moderately altered (perturned). If the inference made would have been qualitatively or categorically different under a perturbation of the observer, then the inference does not satisfy the generic views assumption, and should be rejected. (The question of what constitutes a qualitative or categorical difference is an interesting point of detail.) On this view, it can be argued that the principal of generic views is nothing more than an inference based on the maximum posterior probability (MAP) which accounts for aspects of observation. Thus, we infer the distal phenomenon which possess the highest probability of having generated the observations in question, and this probability incorporates (in addition to relevant priors) both the likelihood of the distal phenomenon generating certain observable signals, and the likelihood of the observer transducing those signals in a manner consistent with the observations. On such an analysis (and with various assumptions invoked), one can obtain a behavior approximating the generic views principal.
Notes
- Suppose in reality there’s a resource, like water, and you can quantify how much of it there is in an objective order—very little water, medium amount of water, a lot of water. Now suppose your fitness function is linear, so a little water gives you a little fitness, medium water gives you medium fitness, and lots of water gives you lots of fitness—in that case, the organism that sees the truth about the water in the world can win, but only because the fitness function happens to align with the true structure in reality. Generically, in the real world, that will never be the case. Something much more natural is a bell curve—say, too little water you die of thirst, but too much water you drown, and only somewhere in between is good for survival. Now the fitness function doesn’t match the structure in the real world. And that’s enough to send truth to extinction. For example, an organism tuned to fitness might see small and large quantities of some resource as, say, red, to indicate low fitness, whereas they might see intermediate quantities as green, to indicate high fitness. Its perceptions will be tuned to fitness, but not to truth. It won’t see any distinction between small and large—it only sees red—even though such a distinction exists in reality.—Donald D. Hoffman to Amanda Gefter
References
- Knill, D. C., & Richards, W., eds., Perception as Bayesian Inference (Cambridge: Cambridge University Press, 1996), p. 478.
- Gefter, A., "The Case Against Reality", The Atlantic, Apr. 25, 2016.
Further reading
- Bennett, Bruce M.; Hoffman, Donald D.; Prakash, Chetan (1989), Observer mechanics: A formal theory of perception, San Diego: Academic Press.
- Bennett, Bruce M.; Hoffman, Donald D.; Prakash, Chetan (1991), "Unity of perception" (PDF), Cognition, 38 (3): 295–334, doi:10.1016/0010-0277(91)90009-s, PMID 2060272.
- Chaitin, Gregory J. (1974), "Information-theoretic computational complexity", IEEE Transactions on Information Theory, IT-20 (1): 10–15, doi:10.1109/tit.1974.1055172, archived from the original on 2007-04-06.
- Klamka, Jerzy (1991), Controllability of Dynamical Systems, Dordrecht: Kluwer.
- Knill, David C.; Richards, Whitman (1996), Perception as Bayesian Inference, Cambridge: Cambridge University Press.
- Kolen, John F.; Pollack, Jordan B. (1995), "The observer's paradox: Apparent computational complexity in physical systems" (PDF), Journal of Experimental and Theoretical Artificial Intelligence, 7 (3): 253–277, doi:10.1080/09528139508953809.