In 2023, “Physicists Don’t Yet Understand Color Qualities” by Brent Allsop was published in the Journal of Neural Philosophy . This paper outlines the basic ideas supported by this camp and points out that all physicists know of color qualities are the false color qualities things seem to be. The supporters of this camp want to know the true color quality of things. The paper asks physicists:
"Which of all our objective descriptions of physical stuff in the brain is a description of subjective redness?"
If you would also like to know the true color quality of things, not just the qualities things seem to be, we could use your support in this camp, or one of the supporting sub camps.
Representational Qualia Theory predicts that consciousness is composed of qualia or phenomenal qualities of something in our brain, as shown in this image by Steven Lehar. Objective observation of the stuff in our brain tells us nothing of these qualities, for the same reason that you can’t tell someone what redness is like, using only text. In reality, nobody knows the subjective colorness qualities of anything in the brain, nor of anything in physics for that matter. All scientific observation is currently done in a qualia blind way (using one abstract word for all things representing red). A good example is the work done by Jack Gallant. In order to discover and demonstrate which of all our descriptions of stuff in the brain is a description of redness, we simply need to understand what the supporters of this camp are saying, and what is described in our video. Only then can we do observations in a non qualia blind way.
Currently, nobody in science is getting this message, so if you understand this, and agree that it is important that we discover what redness is, please sign or support this camp (and get all your friends to do the same), or one of its sub camps, so that eventually physicists and neuroscientists will get the message. This is required to finally do science in a way that will not only tell us which of all our current descriptions of stuff is a description of redness, but also solve this so called ‘hard problem’ and prove which of all the sub camps is THE ONE.
Perception of things is done at a distance via chains of causal intrinsic properties. There are at least the following two sets of intrinsic properties which must be considered if one is not qualia blind:
Most theories of consciousness agree with the general ideas at this camp level. Even Daniel Dennett’s current “Predictive Bayesian Coding Theory” is now in a supporting sub camp to this “Representational Qualia Theory”. Many would consider most of the theories in this topic as worthless, or what some refer to as “crap in the gap” theories. “Crap in the gap” is similar to the evolutionary God of the Gaps idea, and it references the fact that unsupported and unworkable theories have sprung up to fill holes where we lack hard falsifying data. The amount of consensus for this theory supports the idea that most people agree we have qualia. The only disagreement is the nature of qualia. Because of qualia blindness, nobody knows the intrinsic color of anything, and all these theories exist in this gap. Once we discover which of all our descriptions of stuff in this world is a description of redness, all but the one camp correctly predicting the nature of qualia will be falsified.
For example, the most straightforward and easiest to falsify theory is Molecular Materialism. In fact, many people argue it is too simple, therefore already falsified. Though this and other arguments have not yet converted everyone, this falsifiability is the point. It predicts that something like our objective description of glutamate reacting in a synapse is a description of subjective redness. This is easily falsifiable for everyone, if anyone experiences redness without glutamate. Once that is falsified, we then substitute glutamate for the next most likely theory and jump to that camp until we find a necessary and sufficient set of descriptions that is redness which nobody can experimentally falsify. Non qualia blind reporting of experimental science ultimately corralling everyone into a definitively measured in real time scientific consensus. To say nothing of finally knowing the intrinsic colorness qualities of nature, and knowing if an AI is conscious or not.
Our understanding of “red” is limited to the unexamined conformity with the social convention of applying the label to objects that emit certain wavelengths of light. It is hard, if not impossible to find a specific definition of intrinsic “red” or “redness”.
The only difference between the two people in the above image is that one has a red/green inverter somewhere in their chain of perception. Even for those who agree that a redness quality has something to do with light, the question remains, “whose redness?”
Perception of intrinsic properties via our senses and scientific instruments is done through chains of different causal intrinsic properties. Perception is necessarily substrate independent. Each different representation in the chain must be interpreted, correctly, to the next link, to enable the substrate independence. Subjective qualities of conscious knowledge, like redness and greenness, which are the final result of the perception process aren’t perceived, they are directly apprehended. Objective information, like the label ‘red’, is some set of intrinsic properties being interpreted, with a dictionary, as something different. Subjective information, like redness, is just a qualitative fact that is directly apprehended. The subjective redness quality we know is our definition of the term ‘redness’.
If we objectively observed, from afar, the causal behavior of whatever it is, in our brain, that we directly apprehend as redness, what would we have? We would have an abstract objective description of redness behavior. But of course, the description of how redness behaves, would not be intrinsically red. The same is true for our description of glutamate, and our abstract descriptions of how glutamate behaves in a synapse. Our descriptions of glutamate behavior tells us nothing of what that glutamate behavior, itself could be qualitatively like, should it be directly apprehended. Is there anything wrong with assuming a conjecture like the following could be true, at least until it is falsified?
Functionalist camps predict redness and greenness arise from software that is functioning correctly, so if knowledge has redness and greenness qualities, which arise from the 1s and 0s of software functioning correctly, this would be considered conscious, unlike a mechanism that simply represented red knowledge with an abstract word like red, there being nothing that is intrinsically redness.
For example, if you ask a conscious mechanism a question like: “What is your redness like?” and if the objectively verifiable correct answer is something subjectively phenomenal like: “My redness is like your greenness”, that is considered conscious. If the objectively observable accurate answer is “My knowledge is like the word red. It is abstracted away from any particular physical properties, so requires a dictionary.” this is not considered conscious.
We consider the notion of ‘elemental subjective quality’ to be the lowest level required to fully describe any conscious experience. For example, if the description of glutamate, reacting in a synapse, is objectively verified to be a description of subjective redness, you could of course break down glutamate into its constituent atoms, and these atoms further into sub atomic particles and so on. You would not be required to describe this subatomic level of detail to entirely reproduce elemental redness. This would be similar to the way that you don’t need to describe the atomic and subatomic makeup of each pixel of paint on a painting. You only need a sufficient set of pixel values, with enough depth to capture all possible color variants for each pixel.
We almost never experience elemental subjective qualities, without them being bound with other subjective qualities, memories, anticipations, and so on. Completely effing the ineffable nature of a composite qualitative experience would be difficult, as memories, motivations and such would also need to be reproduced and bound. A statement like “My redness is like your greenness.” is assumed to be only referring to elemental redness, not everything else that could normally be bound up with anyone’s experience of elemental redness.
In his paper on “Infonomy & Information” Deepak Loomba coins the term "infonomy.” This is a helpful concept when distinguishing between reality and knowledge of reality. The way our brain normally represents perception, is as if the strawberry is the initial cause of the perception process. But in reality, the big bang is the initial cause of the light landing on our retina, to say nothing of the increasingly smaller levels of physics, down to and beyond the quantum level, that make up the strawberry. All this is labeled as infonomy. Our perception system selects a subset target of all this to ‘perceive’ and represent with conscious knowledge. The strawberry out there is ‘infonomy’, our knowledge of the strawberry is ‘information’ representing only the important target part of infonomy.
Higher level cognitive ideas and thoughts, like additional knowledge of ourselves being aware of redness, are qualitatively different from colorness qualities, even more so than recalled memories of colors. But all these are still considered to be phenomenal like redness and greenness. For anything like this we can be conscious of, there must be something that is that piece of conscious knowledge. If we have conscious access to or awareness of knowledge, there must be some kind of computational binding enabling this access awareness to it.
The minimum required to be considered conscious is at least two values or pixels, each of which can have at least two different subjective qualities, like redness and greenness. These two pixels must be computationally bound, resulting in a “composite qualitative experience”. For example, a thermostat could be engineered to have a heat sensor wired to redness and greenness, rather than eyes sensing light. When the temperature got hot enough, the redness would change to greeness, resulting in the heater turning off, and visa versa. An equally functional mechanism that is only abstract could be a CPU with two registers, with enough discrete logic hardware to compare or bind these two registers with a computation operation like difference. Both mechanisms would be able to tell you if the two values were the same or not. One would be phenomenally conscious, the other would only be abstract. So, yes, according to this definition, you could engineer a conscious thermostat, for which everyone could know that it was like the redness a certain percentage of the human population uses to represent red things with. That is as soon as we find out what it is that has that subjective redness quality.
Some people wouldn’t want a conscious thermostat, thinking of the possibility for it to rise up against them. But the important difference between us and computers has nothing to do with complexity or intelligence. Thinking of a trivially conscious thermostat, with only two pixels of phenomenal knowledge, and no ability to ponder any attacks, proves intelligence isn’t necessary. The important difference is computers achieve their equivalent functionality running on abstract 1s and 0s, using dictionaries to enable them to run on any hardware. We, on the other hand, run directly on subjective qualities like redness and greenness. All a computer can ever have is an abstract word like ‘red’. Even a consciousness without enough complexity to represent all the letters in a word like ‘red’, can still know the wubjective definition of ‘red’ or what it is like.
Almost all literature, today, uses sloppy terminology making it near impossible to clearly communicate about subjective qualities or qualia. They only use terms like ‘red’ or ‘color’. You can’t tell if they are talking about the intrinsic qualities of light, or the very different intrinsic subjective qualities of our knowledge or qualia. We believe that people should start using better terminology. They should only use the term ‘color’ as a label for anything that reflects or emits a particular pattern of light. And they should use a different term like ‘colorness’ to refer to subjectdive properties of our conscious knowledge.
Otherwise most people will completely misinterpret what is being said, and remain qualia blind. Possible diversity of color should also be stressed, by pointing things out like: “my redness is like your greeness, both of which we call red.” And the fact that we can bridge the explanatory gap if one simply defines their terms, as is done in that statement. Whenever people are using sloppy terminology, when talking about perception like this, their language should be called out as ‘qualia blind’.
The supporters of this camp agree that each of the above theories covers an important / unique part of consciousness / qualia, and that they can all be productively unified with Representational Qualia Theory. Except for Symmetry Theory of Valence, None of these other theories provides any kind of insight into subjective qualities of the nature of consciousness or how we might bridge the explanatory gap. RQT describes elemental subjective qualities like redness and greenness as being “computationally bound together”, but says nothing more about “computational binding”. This is what “Integrated Information Theory” provides. It denotes the amount of computational integration by the Greek letter phi Φ, which is the same thing as the amount of computational binding of elemental phenomenal qualities. All this integrated qualitative information results in exactly a computationally bound “momentarily active, subjectively experienced event in working memory,” which is the “Global Workspace” in the “Global Workspace Theory”. Global workspace is the diorama and everything else bound together, probably via some type of resonating waves in nature as described in “Resonance Theory of Consciousness”. Finally, the “valence” or motivational “intrinsic attractiveness/’good’-ness” of various qualia is covered by “Symmetry Theory of Valence”.
The supporters of this camp also agree that there are no ‘hard mind body problems,’ and that the only problem is simply a result of sloppy epistemology of color and other intrinsic qualities, smells, and so on. We are simply still qualia ablind and don’t know the intrinsic colorness quality of anything. The prediction is that using traditional science in a non-qualia blind way will eventually uncover everything necessary about consciousness to completely take it all apart and put it back together in extremely amplified and computationally connected ways.
Representational Qualia Theory predicts that consciousness is composed of qualia or phenomenal qualities of something in our brain, as shown in this image by Steven Lehar. Objective observation of the stuff in our brain tells us nothing of these qualities, for the same reason that you can’t tell someone what redness is like, using only text. In reality, nobody knows the subjective colorness qualities of anything in the brain, nor of anything in physics for that matter. All scientific observation is currently done in a qualia blind way (using one abstract word for all things representing red). A good example is the work done by Jack Gallant. In order to discover and demonstrate which of all our descriptions of stuff in the brain is a description of redness, we simply need to understand what the supporters of this camp are saying, and what is described in our video. Only then can we do observations in a non qualia blind way.
Currently, nobody in science is getting this message, so if you understand this, and agree that it is important that we discover what redness is, please sign or support this camp (and get all your friends to do the same), or one of its sub camps, so that eventually physicists and neuroscientists will get the message. This is required to finally do science in a way that will not only tell us which of all our current descriptions of stuff is a description of redness, but also solve this so called ‘hard problem’ and prove which of all the sub camps is THE ONE.
Perception of things is done at a distance via chains of causal intrinsic properties. There are at least the following two sets of intrinsic properties which must be considered if one is not qualia blind:
For example, the most straightforward and easiest to falsify theory is Molecular Materialism. In fact, many people argue it is too simple, therefore already falsified. Though this and other arguments have not yet converted everyone, this falsifiability is the point. It predicts that something like our objective description of glutamate reacting in a synapse is a description of subjective redness. This is easily falsifiable for everyone, if anyone experiences redness without glutamate. Once that is falsified, we then substitute glutamate for the next most likely theory and jump to that camp until we find a necessary and sufficient set of descriptions that is redness which nobody can experimentally falsify. Non qualia blind reporting of experimental science ultimately corralling everyone into a definitively measured in real time scientific consensus. To say nothing of finally knowing the intrinsic colorness qualities of nature, and knowing if an AI is conscious or not.
Our understanding of “red” is limited to the unexamined conformity with the social convention of applying the label to objects that emit certain wavelengths of light. It is hard, if not impossible to find a specific definition of intrinsic “red” or “redness”.
The only difference between the two people in the above image is that one has a red/green inverter somewhere in their chain of perception. Even for those who agree that a redness quality has something to do with light, the question remains, “whose redness?”
Perception of intrinsic properties via our senses and scientific instruments is done through chains of different causal intrinsic properties. Perception is necessarily substrate independent. Each different representation in the chain must be interpreted, correctly, to the next link, to enable the substrate independence. Subjective qualities of conscious knowledge, like redness and greenness, which are the final result of the perception process aren’t perceived, they are directly apprehended. Objective information, like the label ‘red’, is some set of intrinsic properties being interpreted, with a dictionary, as something different. Subjective information, like redness, is just a qualitative fact that is directly apprehended. The subjective redness quality we know is our definition of the term ‘redness’.
If we objectively observed, from afar, the causal behavior of whatever it is, in our brain, that we directly apprehend as redness, what would we have? We would have an abstract objective description of redness behavior. But of course, the description of how redness behaves, would not be intrinsically red. The same is true for our description of glutamate, and our abstract descriptions of how glutamate behaves in a synapse. Our descriptions of glutamate behavior tells us nothing of what that glutamate behavior, itself could be qualitatively like, should it be directly apprehended. Is there anything wrong with assuming a conjecture like the following could be true, at least until it is falsified?
Functionalist camps predict redness and greenness arise from software that is functioning correctly, so if knowledge has redness and greenness qualities, which arise from the 1s and 0s of software functioning correctly, this would be considered conscious, unlike a mechanism that simply represented red knowledge with an abstract word like red, there being nothing that is intrinsically redness.
For example, if you ask a conscious mechanism a question like: “What is your redness like?” and if the objectively verifiable correct answer is something subjectively phenomenal like: “My redness is like your greenness”, that is considered conscious. If the objectively observable accurate answer is “My knowledge is like the word red. It is abstracted away from any particular physical properties, so requires a dictionary.” this is not considered conscious.
We consider the notion of ‘elemental subjective quality’ to be the lowest level required to fully describe any conscious experience. For example, if the description of glutamate, reacting in a synapse, is objectively verified to be a description of subjective redness, you could of course break down glutamate into its constituent atoms, and these atoms further into sub atomic particles and so on. You would not be required to describe this subatomic level of detail to entirely reproduce elemental redness. This would be similar to the way that you don’t need to describe the atomic and subatomic makeup of each pixel of paint on a painting. You only need a sufficient set of pixel values, with enough depth to capture all possible color variants for each pixel.
We almost never experience elemental subjective qualities, without them being bound with other subjective qualities, memories, anticipations, and so on. Completely effing the ineffable nature of a composite qualitative experience would be difficult, as memories, motivations and such would also need to be reproduced and bound. A statement like “My redness is like your greenness.” is assumed to be only referring to elemental redness, not everything else that could normally be bound up with anyone’s experience of elemental redness.
In his paper on “Infonomy & Information” Deepak Loomba coins the term "infonomy.” This is a helpful concept when distinguishing between reality and knowledge of reality. The way our brain normally represents perception, is as if the strawberry is the initial cause of the perception process. But in reality, the big bang is the initial cause of the light landing on our retina, to say nothing of the increasingly smaller levels of physics, down to and beyond the quantum level, that make up the strawberry. All this is labeled as infonomy. Our perception system selects a subset target of all this to ‘perceive’ and represent with conscious knowledge. The strawberry out there is ‘infonomy’, our knowledge of the strawberry is ‘information’ representing only the important target part of infonomy.
Higher level cognitive ideas and thoughts, like additional knowledge of ourselves being aware of redness, are qualitatively different from colorness qualities, even more so than recalled memories of colors. But all these are still considered to be phenomenal like redness and greenness. For anything like this we can be conscious of, there must be something that is that piece of conscious knowledge. If we have conscious access to or awareness of knowledge, there must be some kind of computational binding enabling this access awareness to it.
The minimum required to be considered conscious is at least two values or pixels, each of which can have at least two different subjective qualities, like redness and greenness. These two pixels must be computationally bound, resulting in a “composite qualitative experience”. For example, a thermostat could be engineered to have a heat sensor wired to redness and greenness, rather than eyes sensing light. When the temperature got hot enough, the redness would change to greeness, resulting in the heater turning off, and visa versa. An equally functional mechanism that is only abstract could be a CPU with two registers, with enough discrete logic hardware to compare or bind these two registers with a computation operation like difference. Both mechanisms would be able to tell you if the two values were the same or not. One would be phenomenally conscious, the other would only be abstract. So, yes, according to this definition, you could engineer a conscious thermostat, for which everyone could know that it was like the redness a certain percentage of the human population uses to represent red things with. That is as soon as we find out what it is that has that subjective redness quality.
Some people wouldn’t want a conscious thermostat, thinking of the possibility for it to rise up against them. But the important difference between us and computers has nothing to do with complexity or intelligence. Thinking of a trivially conscious thermostat, with only two pixels of phenomenal knowledge, and no ability to ponder any attacks, proves intelligence isn’t necessary. The important difference is computers achieve their equivalent functionality running on abstract 1s and 0s, using dictionaries to enable them to run on any hardware. We, on the other hand, run directly on subjective qualities like redness and greenness. All a computer can ever have is an abstract word like ‘red’. Even a consciousness without enough complexity to represent all the letters in a word like ‘red’, can still know the wubjective definition of ‘red’ or what it is like.
Almost all literature, today, uses sloppy terminology making it near impossible to clearly communicate about subjective qualities or qualia. They only use terms like ‘red’ or ‘color’. You can’t tell if they are talking about the intrinsic qualities of light, or the very different intrinsic subjective qualities of our knowledge or qualia. We believe that people should start using better terminology. They should only use the term ‘color’ as a label for anything that reflects or emits a particular pattern of light. And they should use a different term like ‘colorness’ to refer to subjectdive properties of our conscious knowledge.
Otherwise most people will completely misinterpret what is being said, and remain qualia blind. Possible diversity of color should also be stressed, by pointing things out like: “my redness is like your greeness, both of which we call red.” And the fact that we can bridge the explanatory gap if one simply defines their terms, as is done in that statement. Whenever people are using sloppy terminology, when talking about perception like this, their language should be called out as ‘qualia blind’.
The supporters of this camp also agree that there are no ‘hard mind body problems,’ and that the only problem is simply a result of sloppy epistemology of color and other intrinsic qualities, smells, and so on. We are simply still qualia ablind and don’t know the intrinsic colorness quality of anything. The prediction is that using traditional science in a non-qualia blind way will eventually uncover everything necessary about consciousness to completely take it all apart and put it back together in extremely amplified and computationally connected ways.
Representational Qualia Theory predicts that consciousness is composed of qualia or phenomenal qualities of something in our brain, as shown in this image by Steven Lehar. Objective observation of the stuff in our brain tells us nothing of these qualities, for the same reason that you can’t tell someone what redness is like, using only text. In reality, nobody knows the intrinsic colorness qualities of anything in the brain, nor of anything in physics for that matter. All scientific observation is currently done in a qualia blind way (using one abstract word for all things representing red). A good example is the work done by Jack Gallant. In order to discover and demonstrate which of all our descriptions of stuff in the brain is a description of redness, we simply need to understand what the supporters of this camp are saying, and what is described in our video. Only then can we do observations in a non qualia blind way.
Currently, nobody in science is getting this message, so if you understand this, and agree that it is important that we discover what redness is, please sign or support this camp (and get all your friends to do the same), or one of its sub camps, so that eventually physicists and neuroscientists will get the message. This is required to finally do science in a way that will not only tell us which of all our current descriptions of stuff is a description of redness, but also solve this so called ‘hard problem’ and prove which of all the sub camps is THE ONE.
Perception of things is done at a distance via chains of causal intrinsic properties. There are at least the following two sets of intrinsic properties which must be considered if one is not qualia blind:
For example, the most straightforward and easiest to falsify theory is Molecular Materialism. In fact, many people argue it is too simple, therefore already falsified. Though this and other arguments have not yet converted everyone, this falsifiability is the point. It predicts that something like our description of glutamate reacting in a synapse is a description of intrinsic redness. This is easily falsifiable for everyone, if anyone experiences redness without glutamate. Once that is falsified, we then substitute glutamate for the next most likely theory and jump to that camp until we find a necessary and sufficient set of descriptions that is redness which nobody can experimentally falsify. Non qualia blind reporting of experimental science ultimately corralling everyone into a definitively measured in real time scientific consensus. To say nothing of finally knowing the intrinsic colorness qualities of nature, and knowing if an AI is conscious or not.
Our understanding of “red” is limited to the unexamined conformity with the social convention of applying the label to objects that emit certain wavelengths of light. It is hard, if not impossible to find a specific definition of intrinsic “red” or “redness”.
The only difference between the two people in the above image is that one has a red/green inverter somewhere in their chain of perception. Even for those who agree that a redness quality has something to do with light, the question remains, “whose redness?”
Perception of intrinsic properties via our senses and scientific instruments is done through chains of different causal intrinsic properties. Perception is necessarily substrate independent. Each different representation in the chain must be interpreted, correctly, to the next link, to enable the substrate independence. Intrinsic qualities of conscious knowledge, like redness and greenness, which are the final result of the perception process aren’t perceived, they are directly apprehended. Objective information, like the label ‘red’, is some set of intrinsic properties being interpreted, with a dictionary, as something different. Subjective information, like redness, is just a qualitative fact that is directly apprehended. The intrinsic redness quality we know is our definition of the term ‘redness’.
If we objectively observed, from afar, the causal behavior of whatever it is, in our brain, that we directly apprehend as redness, what would we have? We would have an abstract objective description of redness behavior. But of course, the description of how redness behaves, would not be intrinsically red. The same is true for our description of glutamate, and our abstract descriptions of how glutamate behaves in a synapse. Our descriptions of glutamate behavior tells us nothing of what that glutamate behavior, itself could be qualitatively like, should it be directly apprehended. Is there anything wrong with assuming a conjecture like the following could be true, at least until it is falsified?
Functionalist camps predict redness and greenness arise from software that is functioning correctly, so if knowledge has redness and greenness qualities, which arise from the 1s and 0s of software functioning correctly, this would be considered conscious, unlike a mechanism that simply represented red knowledge with an abstract word like red, there being nothing that is intrinsically redness.
For example, if you ask a conscious mechanism a question like: “What is your redness like?” and if the objectively verifiable correct answer is something intrinsically phenomenal like: “My redness is like your greenness”, that is considered conscious. If the objectively observable accurate answer is “My knowledge is like the word red. It is abstracted away from any particular physical properties, so requires a dictionary.” this is not considered conscious.
We consider the notion of ‘elemental intrinsic quality’ to be the lowest level required to fully describe any conscious experience. For example, if the description of glutamate, reacting in a synapse, is objectively verified to be a description of subjective redness, you could of course break down glutamate into its constituent atoms, and these atoms further into sub atomic particles and so on. You would not be required to describe this subatomic level of detail to entirely reproduce elemental redness. This would be similar to the way that you don’t need to describe the atomic and subatomic makeup of each pixel of paint on a painting. You only need a sufficient set of pixel values, with enough depth to capture all possible color variants for each pixel.
We almost never experience elemental intrinsic qualities, without them being bound with other intrinsic qualities, memories, anticipations, and so on. Completely effing the ineffable nature of a composite qualitative experience would be difficult, as memories, motivations and such would also need to be reproduced and bound. A statement like “My redness is like your greenness.” is assumed to be only referring to elemental redness, not everything else that could normally be bound up with anyone’s experience of elemental redness.
In his paper on “Infonomy & Information” Deepak Loomba coins the term "infonomy.” This is a helpful concept when distinguishing between reality and knowledge of reality. The way our brain normally represents perception, is as if the strawberry is the initial cause of the perception process. But in reality, the big bang is the initial cause of the light landing on our retina, to say nothing of the increasingly smaller levels of physics, down to and beyond the quantum level, that make up the strawberry. All this is labeled as infonomy. Our perception system selects a subset target of all this to ‘perceive’ and represent with conscious knowledge. The strawberry out there is ‘infonomy’, our knowledge of the strawberry is ‘information’ representing only the important target part of infonomy.
Higher level cognitive ideas and thoughts, like additional knowledge of ourselves being aware of redness, are qualitatively different from colorness qualities, even more so than recalled memories of colors. But all these are still considered to be phenomenal like redness and greenness. For anything like this we can be conscious of, there must be something that is that piece of conscious knowledge. If we have conscious access to or awareness of knowledge, there must be some kind of computational binding enabling this access awareness to it.
The minimum required to be considered conscious is at least two values or pixels, each of which can have at least two different intrinsic qualities, like redness and greenness. These two pixels must be computationally bound, resulting in a “composite qualitative experience”. For example, a thermostat could be engineered to have a heat sensor wired to redness and greenness, rather than eyes sensing light. When the temperature got hot enough, the redness would change to greeness, resulting in the heater turning off, and visa versa. An equally functional mechanism that is only abstract could be a CPU with two registers, with enough discrete logic hardware to compare or bind these two registers with a computation operation like difference. Both mechanisms would be able to tell you if the two values were the same or not. One would be phenomenally conscious, the other would only be abstract. So, yes, according to this definition, you could engineer a conscious thermostat, for which everyone could know that it was like the redness a certain percentage of the human population uses to represent red things with. That is as soon as we find out what it is that has that intrinsic redness quality.
Some people wouldn’t want a conscious thermostat, thinking of the possibility for it to rise up against them. But the important difference between us and computers has nothing to do with complexity or intelligence. Thinking of a trivially conscious thermostat, with only two pixels of phenomenal knowledge, and no ability to ponder any attacks, proves intelligence isn’t necessary. The important difference is computers achieve their equivalent functionality running on abstract 1s and 0s, using dictionaries to enable them to run on any hardware. We, on the other hand, run directly on intrinsic qualities like redness and greenness. All a computer can ever have is an abstract word like ‘red’. Even a consciousness without enough complexity to represent all the letters in a word like ‘red’, can still know the intrinsic definition of ‘red’ or what it is like.
Almost all literature, today, uses sloppy terminology making it near impossible to clearly communicate about intrinsic qualities of qualia. They only use terms like ‘red’ or ‘color’. You can’t tell if they are talking about the intrinsic qualities of light, or the very different intrinsic qualities of our knowledge or qualia. We believe that people should start using better terminology. They should only use the term ‘color’ as a label for anything that reflects or emits a particular pattern of light. And they should use a different term like ‘colorness’ to refer to intrinsic properties of our conscious knowledge.
Otherwise most people will completely misinterpret what is being said, and remain qualia blind. Possible diversity of color should also be stressed, by pointing things out like: “my redness is like your greeness, both of which we call red.” And the fact that we can bridge the explanatory gap if one simply defines their terms, as is done in that statement. Whenever people are using sloppy terminology, when talking about perception like this, their language should be called out as ‘qualia blind’.
The supporters of this camp also agree that there are no ‘hard mind body problems,’ and that the only problem is simply a result of sloppy epistemology of color and other intrinsic qualities, smells, and so on. We are simply still qualia ablind and don’t know the intrinsic colorness quality of anything. The prediction is that using traditional science in a non-qualia blind way will eventually uncover everything necessary about consciousness to completely take it all apart and put it back together in extremely amplified and computationally connected ways.
Representational Qualia Theory predicts that consciousness is composed of qualia or phenomenal qualities of something in our brain, as shown in this image by Steven Lehar. Objective observation of the stuff in our brain tells us nothing of these qualities, for the same reason that you can’t tell someone what redness is like, using only text. In reality, nobody knows the intrinsic colorness qualities of anything in the brain, nor of anything in physics for that matter. All scientific observation is currently done in a qualia blind way (using one abstract word for all things representing red). A good example is the work done by Jack Gallant. In order to discover and demonstrate which of all our descriptions of stuff in the brain is a description of redness, we simply need to understand what the supporters of this camp are saying, and what is described in our video. Only then can we do observations in a non qualia blind way.
Currently, nobody in science is getting this message, so if you understand this, and agree that it is important that we discover what redness is, please sign or support this camp (and get all your friends to do the same), or one of its sub camps, so that eventually physicists and neuroscientists will get the message. This is required to finally do science in a way that will not only tell us which of all our current descriptions of stuff is a description of redness, but also solve this so called ‘hard problem’ and prove which of all the sub camps is THE ONE.
Perception of things is done at a distance via chains of causal intrinsic properties. There are at least the following two sets of intrinsic properties which must be considered if one is not qualia blind:
For example, the most straightforward and easiest to falsify theory is Molecular Materialism. In fact, many people argue it is too simple, therefore already falsified. Though this and other arguments have not yet converted everyone, this falsifiability is the point. It predicts that something like our description of glutamate reacting in a synapse is a description of intrinsic redness. This is easily falsifiable for everyone, if anyone experiences redness without glutamate. Once that is falsified, we then substitute glutamate for the next most likely theory and jump to that camp until we find a necessary and sufficient set of descriptions that is redness which nobody can experimentally falsify. Non qualia blind reporting of experimental science ultimately corralling everyone into a definitively measured in real time scientific consensus. To say nothing of finally knowing the intrinsic colorness qualities of nature, and knowing if an AI is conscious or not.
Our understanding of “red” is limited to the unexamined conformity with the social convention of applying the label to objects that emit certain wavelengths of light. It is hard, if not impossible to find a specific definition of intrinsic “red” or “redness”.
The only difference between the two people in the above image is that one has a red/green inverter somewhere in their chain of perception. Even for those who agree that a redness quality has something to do with light, the question remains, “whose redness?”
Perception of intrinsic properties via our senses and scientific instruments is done through chains of different causal intrinsic properties. Perception is necessarily substrate independent. Each different representation in the chain must be interpreted, correctly, to the next link, to enable the substrate independence. Intrinsic qualities of conscious knowledge, like redness and greenness, which are the final result of the perception process aren’t perceived, they are directly apprehended. Objective information, like the label ‘red’, is some set of intrinsic properties being interpreted, with a dictionary, as something different. Subjective information, like redness, is just a qualitative fact that is directly apprehended. The intrinsic redness quality we know is our definition of the term ‘redness’.
If we objectively observed, from afar, the causal behavior of whatever it is, in our brain, that we directly apprehend as redness, what would we have? We would have an abstract objective description of redness behavior. But of course, the description of how redness behaves, would not be intrinsically red. The same is true for our description of glutamate, and our abstract descriptions of how glutamate behaves in a synapse. Our descriptions of glutamate behavior tells us nothing of what that glutamate behavior, itself could be qualitatively like, should it be directly apprehended. Is there anything wrong with assuming a conjecture like the following could be true, at least until it is falsified?
Functionalist camps predict redness and greenness arise from software that is functioning correctly, so if knowledge has redness and greenness qualities, which arise from the 1s and 0s of software functioning correctly, this would be considered conscious, unlike a mechanism that simply represented red knowledge with an abstract word like red, there being nothing that is intrinsically redness.
For example, if you ask a conscious mechanism a question like: “What is your redness like?” and if the objectively verifiable correct answer is something intrinsically phenomenal like: “My redness is like your greenness”, that is considered conscious. If the objectively observable accurate answer is “My knowledge is like the word red. It is abstracted away from any particular physical properties, so requires a dictionary.” this is not considered conscious.
We consider the notion of ‘elemental intrinsic quality’ to be the lowest level required to fully describe any conscious experience. For example, if the description of glutamate, reacting in a synapse, is objectively verified to be a description of subjective redness, you could of course break down glutamate into its constituent atoms, and these atoms further into sub atomic particles and so on. You would not be required to describe this subatomic level of detail to entirely reproduce elemental redness. This would be similar to the way that you don’t need to describe the atomic and subatomic makeup of each pixel of paint on a painting. You only need a sufficient set of pixel values, with enough depth to capture all possible color variants for each pixel.
We almost never experience elemental intrinsic qualities, without them being bound with other intrinsic qualities, memories, anticipations, and so on. Completely effing the ineffable nature of a composite qualitative experience would be difficult, as memories, motivations and such would also need to be reproduced and bound. A statement like “My redness is like your greenness.” is assumed to be only referring to elemental redness, not everything else that could normally be bound up with anyone’s experience of elemental redness.
In his paper on “Infonomy & Information” Deepak Loomba coins the term "infonomy.” This is a helpful concept when distinguishing between reality and knowledge of reality. The way our brain normally represents perception, is as if the strawberry is the initial cause of the perception process. But in reality, the big bang is the initial cause of the light landing on our retina, to say nothing of the increasingly smaller levels of physics, down to and beyond the quantum level, that make up the strawberry. All this is labeled as infonomy. Our perception system selects a subset target of all this to ‘perceive’ and represent with conscious knowledge. The strawberry out there is ‘infonomy’, our knowledge of the strawberry is ‘information’ representing only the important target part of infonomy.
Higher level cognitive ideas and thoughts, like additional knowledge of ourselves being aware of redness, are qualitatively different from colorness qualities, even more so than recalled memories of colors. But all these are still considered to be phenomenal like redness and greenness. For anything like this we can be conscious of, there must be something that is that piece of conscious knowledge. If we have conscious access to or awareness of knowledge, there must be some kind of computational binding enabling this access awareness to it.
The minimum required to be considered conscious is at least two values or pixels, each of which can have at least two different intrinsic qualities, like redness and greenness. These two pixels must be computationally bound, resulting in a “composite qualitative experience”. For example, a thermostat could be engineered to have a heat sensor wired to redness and greenness, rather than eyes sensing light. When the temperature got hot enough, the redness would change to greeness, resulting in the heater turning off, and visa versa. An equally functional mechanism that is only abstract could be a CPU with two registers, with enough discrete logic hardware to compare or bind these two registers with a computation operation like difference. Both mechanisms would be able to tell you if the two values were the same or not. One would be phenomenally conscious, the other would only be abstract. So, yes, according to this definition, you could engineer a conscious thermostat, for which everyone could know that it was like the redness a certain percentage of the human population uses to represent red things with. That is as soon as we find out what it is that has that intrinsic redness quality.
Some people wouldn’t want a conscious thermostat, thinking of the possibility for it to rise up against them. But the important difference between us and computers has nothing to do with complexity or intelligence. Thinking of a trivially conscious thermostat, with only two pixels of phenomenal knowledge, and no ability to ponder any attacks, proves intelligence isn’t necessary. The important difference is computers achieve their equivalent functionality running on abstract 1s and 0s, using dictionaries to enable them to run on any hardware. We, on the other hand, run directly on intrinsic qualities like redness and greenness. All a computer can ever have is an abstract word like ‘red’. Even a consciousness without enough complexity to represent all the letters in a word like ‘red’, can still know the intrinsic definition of ‘red’ or what it is like.
Almost all literature, today, uses sloppy terminology making it near impossible to clearly communicate about intrinsic qualities of qualia. They only use terms like ‘red’ or ‘color’. You can’t tell if they are talking about the intrinsic qualities of light, or the very different intrinsic qualities of our knowledge or qualia. We believe that people should start using better terminology. They should only use the term ‘color’ as a label for anything that reflects or emits a particular pattern of light. And they should use a different term like ‘colorness’ to refer to intrinsic properties of our conscious knowledge.
Otherwise most people will completely misinterpret what is being said, and remain qualia blind. Possible diversity of color should also be stressed, by pointing things out like: “my redness is like your greeness, both of which we call red.” And the fact that we can bridge the explanatory gap if one simply defines their terms, as is done in that statement. Whenever people are using sloppy terminology, when talking about perception like this, their language should be called out as ‘qualia blind’.
The supporters of this camp also agree that there are no ‘hard mind body problems,’ and that the only problem is simply a result of sloppy epistemology of color and other intrinsic qualities, smells, and so on. We are simply still qualia ablind and don’t know the intrinsic colorness quality of anything. The prediction is that using traditional science in a non-qualia blind way will eventually uncover everything necessary about consciousness to completely take it all apart and put it back together in extremely amplified and computationally connected ways.