Actions

Epistemology

From Deliberative Democracy Institiute Wiki

Revision as of 00:34, 5 May 2013 by WinSysop (talk | contribs) (Explaining the Mechanisms of Knowledge)
framless

This page is a stub. It is not ready for publication and is used to aggregate information about a subject. You can add further reading and add information to the page. If you want to prepare this page for publication please consults with the creator of this page.
Tal Yaron 14:08, 20 February 2013 (IST)

framless

This page was writen by a non-English speeking writer. Please help us improve the quality of the paper.Tal Yaron 14:08, 20 February 2013 (IST)

The Foundations of Knowledge

The phenomenological cage

How can you know if you are not a brain in a vat?

Deliberation is a process of thoughtfully weighing options before making a decision. Yet in order to choose among the available options, we have to agree on the options available to us, and theןir outcomes. But many times we find ourselves in disagreement about the options or the ways the world behave. These disagreements are the results of differences in our understanding of how the world works. So in order to be able to better understand why we perceive the world differently, I suggest we have to understand how our knowledge is built, and why it is different from person to person. This Page will try to explain how knowledge is built and why we perceive the world differently.

To explain knowledge, I will suggest that knowledge is a thing that is being used to understand, explain, predict and manipulate the inputs that come from the senses.

Thouse inputs that come from our senses, whether outer senses like smell, vision, hearing, touch, wormth, etc. or from our inner senses like therst, hounger, love, hate, etc are all will be called the phenomena. Phenomena, with correspondance to Kant's philiosophy, is our senses expriences.

Through the usage of knowledge we try to understand our surroundings and inner feelings, but we have a an unobservable barriar to the surrounding or even our inner selves. We have no access to the surrounding itself. All our knowledge about the "surrounding" comes from our senses. We have pehnomena, but no access to the thing that creates the expreinces. For more than 2500 years of epistemology, nobody had found a reliable way to establish a relations between perception and the “surrounding”. To demonstrate the problem of the relations between knowledge and the surronding, we may use the thought experiment of the “brain in a vat”. In this thought experiment, you are asked to find a reliable way to know if you really exists as you perceive it, or you are actually a brain in a vat, which gets it's sensory inputs from a computer, that simulate the perceived world.

Till today nobody was able to find a reliable answer to this question. Philosophers sometime suggest that Hillary Putnaham had found a way, but his conclusions say otherwise. He concluded that we cannot distinguish between realty and virtual experience[1].

Therefore, we have to set for now an axiom that says:

We do not know the relation between our knowledge and the inner or outer-world”. Or in Kant's methodology, we cannot know the noumenon.

The only thing we can say is that we percive. How this perception is constructed, I will suggest later own in this paper. as a result of our inability to go beiond our perceptions, I will call this principle "The phenomenological cage principle".

Explaining the Mechanisms of Knowledge

The Phenomena

Kant described the inputs from the senses as “phenomena”.

Propensities in the phenomena

Knowledge is about tring to predict the phenoema. We try to predict by using an axiom. The axiom says that the phenomena are inputs that our senses gets from unknown sources (Which Kant named neumenon, or the thing in itself), and that while there is no neumneon, the phenomena will be random or blank, and that the nomenon when interacting with the senses, will create some propensities in the otherwise random or balnk phenoemna[2]. It will suggest that the neumena will create some order in the phenomena.

Phenomena Sets

I will also assume that the phenomena can be destinguished into specific sets. That there are specific smells (Smell

Memory

Similarity

The mind assume similarity between sequential phenomena or sensory inputs when the mind find that a set of phenomena has, by the way it precives, resmblance to other set of phenomena, all ready stored in his memory. We then can say that there were some order in phenomena, that was resamble to other order in the phenomena.

In a philospohical manner it is hard to define similarity in phenomena. For some solution you can go to Churchland[3].

When using neurons we can say that when signals from the same sets of receptors or from same set of sensory objects occur simultaneously, they will initiate an LTP[4] somewhere down the neural-networks, and will cause induction or LTP between the set of receptors or the set of sensory objects. In an optimal repetitions temporally sequence of similar inputs, the LTP will be strengthen.

See also the problem of identity

Simplicity and Induction

The relation between two similar phenomena might be very simple. They may be linked directly. For instance, A sound of a broken glass may be directly connected to a braking glass. But it can be connected in more complex way, for instance there may be some computer that create a breaking sound, whenever I see a breaking glass. Or this connection may be correct until some time, ans it may be wrong in some future time[5]. The reason to chose the most simple connection is because the number of available possible relations between a sound and a breaking glass are infinite. Therefore for reasons of effective storage, we will use the simplest solution available, which will be "All breaking glasses so far observed, has this specific sound".

Although this statement may be true it can not tell us nothing about the next occasion of breaking glass. If we will test it and we will find that next time we will hear a different sound, we might say that "All breaking glasses so far observed has this specific sound, while the last had another sound". By describing the relations between the phenomena in a descriptive mode, we may reach very fast to infinite numbers of description. So yet again for the sake of simplicity, we will try to describe the relations as inductions.

Inductions are relations, assumed to be constant regardless of time or place, between phenomena. In our example, we will say that "Always, when a glass brakes it has specific sound" ("broken glass sound"). Due to the Phenomenological cage principal, we are unable to check that this relation is always true. What we can do, is to try to test it in every occasion the one of the phenomena occur, and to see if it is still stand, or was the induction refuted[6]

When an induction is being refuted, we should try to suggest new relation, this time more complex, to describe the relations that will be true for every observation we have seen so far. This new induction will be tested again and again, and after every refutation, a new relation will be conjecture and will be tested, when every new induction should describe every observation we had so far.

We should strive to achive the simplest sets of inductions that describes erey observed relations between the phenomena. This set inductions (also called theories) will be called descriptive theories. If the sets of inductions also enable us to manipulate phenomena preciasly[7] it will be called forcasting theories. The inductions are like equations that describe phenomena in more economical way[8].

framless

This section needs more clarifications. Please explein:
Please better define how we mesure simplicity
Tal Yaron 08:45, 14 December 2012 (IST)

The Structure of Theories

Mental Objects

Senssory Mental Objects

construction of Mental Object by recurrence of several sensory stimuli simultaneously

Somtimes the brain finds several connections between diffrent sets of inputs, that come toghther. When several connections are reoccuring they form a cluster of inductions that form Mental Object (MO).For instance a specfic smell may come always when specific shape and specifice color and specific feeling appaer. A smell of an apple may come with a shape of an apple, a color of an apple and a taste of an apple. This cluster of connections will create for us a femiler object that when ever we see one, we will assume that all other coneection will appaer also. If we will also hear the same name ("apple") every time we will see some of the connected appreances, we will also give him a name: "apple"[9] that will be part of the mental object.

Refutation of a mental object:apple
One of the infnite ways to reconstract the MO:apple

Mental Objects are compused of inductions and as every induction in the phenomenological cage, they should be put to a test. Therefore MOs may change and reconstracted. For instance, all the apples we have seen so far were red. Sudenly I see a shape of an apple, I smell a smell of an apple and it has a feeling of an apple, but it's color is green and it's taste is diffrent. What is it? I have expreinced a refutation of my expected apperances by the old MO:apple. As in every induction, when it is refuted, we have to find new set of inductions that will describe all appreance we have seen so far. The posiblities to describe all phenomena is infinte, and again we will look for the simplest sets of connections that will describe all abservations. One simple rearngments will be to say that the old observations and the new observationare are two diffrent objects. Another simple reconstruction will be to say that all my observation relating this MO:apple are of the smae object but diffrent of kinds. In the former the connection to red and sweet (This is the taste of the red object), will be on mental object, while the the green and the soure will be another object. Both objects have the same shape and the the same touch and the same smell, but different colors and tastes. In the later, they are all apples, and apples have purticular shape, feeling and smell, but there is a kind of red-sweet appels and there is a kind of green-sour appels. of course there may be other ways to to reconstruct mental objects. We can say that all apples are red, but sombody painted and weetend this last observation of an apple, or this was an accident, freak of nature and so on.

To decide ifbetween alternatives we may need some more avidance. Further evidance may suggest that we should divied it to two diffrent mental objects. For instance if our theoris will be sufficiantly developed to the point when we recognize a "DNA" thus we may find that they have a DNA linage that is very diffrent, and because we defined DNA as an assntial proprtie of a leaving object, we may decide that they are two diffrent mental objects.

Deductions

We duduce by going through connected MOs

Mental Objects construct our understanding of the world. When we reason, we can go along the lines of connected MOs. For instance, humans have many inductions. We know that they have two legs, two arms, they leave up to 120 years, and in the meanwhile they breath, born and die, and many other inductions (mistakenly called proprties). We also know (or conjecture) that Socrates was human, and as such he has all the inductions of the MO:humans.

When we use the clasical sylogism "Socrats was human, all humens are mortal, therfore Socrates is mortal" we go along the lines that connect the MOs. By this mechanism we may say that Socrates was also born. If we want to speculate how old was his mother when she gave birth to Socrates we need some other inductions and mental object. we need to conjectures in what ege women can give birth (age 12-45), and that humans are born by womens and that that woman is his mother. Therefore we can conculde that Socrates was born when his mother was at the age of 12-45, and we therfore conclude that his mother is 12-45 years older then Socrates.

framless

This section needs more clarifications. Please explein:
More ilustrations will be very helpfull. If equations can be enterd, this will be awsome.
Tal Yaron 09:13, 14 December 2012 (IST)

Imaginary Mental Objects

Some times we have gaps between our mental objects. These gaps usually puzzle us and we look for an explanation or some intermediate MO that will closes the gap. Therefore we invent an unseen MO to bridge the gap. When we invent the bridging MO, we usually imagine it and we do not get if from phenomenological intervals (when we will further research, we will see that all MOs are imaginary mental objects), therefore we will call it Imaginary Mental Object or iMO, while the objects obtained from phenomenological intervals will call Sensory Mental Objects(sMO).

For instance, when the ancient looked upward and saw the sun travels through the skies, the where puzzled. The knew that every object that is traveling need something to pull or to thrust it, yet the sun seems to move without a mover. To solve it, they invented a "story" or a mental object that was missing. They told their hearers that the sun is pulled by invisible horses. The invisible horses where their iMO that bridged the gap.

We use iMOs more then we recognize naturally. All our theories and especially the scientific theories are built of networks of iMOs. We only seldom see directly our iMOs. For instance iMO:atoms was talked about for more then 200 years in the modern science. Yet not until recently people did not observe atoms. And even today when we "see" atoms it is through very complicated theories constucted into the scientific observation tools. We cannot observe atoms directly,and therefore atoms will stay iMOs.

People who do not engage epistemology are not awere of the extant of our usage of iMOs. Hume was probably the first thinker that showed the we take for granted things we do not see. He demonstrated in his book An Enquiry Concerning Human Understanding[10], that cause and effect are not to be observed directly, yet people use them as though they are real enteties. God is iMO, and also The force of gravity, and even the state is an iMO (Have any of you ever seen a state?).

Other Asspects of Epistemology

Bush and Mosteller Learning Curve

This section was mostly taken from Paul W. Glimcher1 paper in PNAS[11]

This very general idea of Conditional learning, was first mathematically formalized when Bush and Mosteller[12], 24) proposed that the probability of Pavlov's[13] dog expressing the salivary response on sequential trials could be computed through an iterative equation where

Bush and Mosteller eq1.jpg


In this equation, Anext_trial is the probability that the salivation will occur on the next trial (or more formally, the associative strength of the connection between the bell and salivation). To compute Anext_trial, one begins with the value of A on the previous trial and adds to it a correction based on the animal's experience during the most recent trial. This correction, or error term, is the difference between what the animal actually experienced (in this case, the reward of the meat powder expressed as Rcurrent_trial) and what he expected (simply, what A was on the previous trial). The difference between what was obtained and what was expected is multiplied by α, a number ranging from 0 to 1, which is known as the learning rate. When α = 1, A is always immediately updated so that it equals R from the last trial. When α = 0.5, only one-half of the error is corrected, and the value of A converges in half steps to R. When the value of α is small, around 0.1, then A is only very slowly incremented to the value of R.

What the Bush and Mosteller[14][15] equation does is compute an average of previous rewards across previous trials. In this average, the most recent rewards have the greatest impact, whereas rewards far in the past have only a weak impact. If, to take a concrete example, α = 0.5, then the equation takes the most recent reward, uses it to compute the error term, and multiplies that term by 0.5. One-half of the new value of A is, thus, constructed from this most recent observation. That means that the sum of all previous error terms (those from all trials in the past) has to count for the other one-half of the estimate. If one looks at that older one-half of the estimate, one-half of that one-half comes from what was observed one trial ago (thus, 0.25 of the total estimate) and one-half (0.25 of the estimate) comes from the sum of all trials before that one. The iterative equation reflects a weighted sum of previous rewards. When the learning rate (α) is 0.5, the weighting rule effectively being carried out is an exponential series, the rate at which the weight declines being controlled by α.

Bush and Mosteller eq2.gif

When α is high, the exponential function declines rapidly and puts all of the weight on the most recent experiences of the animal. When α is low, it declines slowly and averages together many observations, which is shown in Fig. 1.

Bush and Mosteller graph1.gif

Fig 1: “Weights determining the effects of previous rewards on current associative strength effectively decline as an exponential function of time”[16].


The Bush and Mosteller equation was critically important, because it was the first use of this kind of iterative error-based rule for reinforcement learning; additionally, it forms the basis of all modern approaches to this problem. This is a fact often obscured by what is known as the Rescorla–Wagner model of classical conditioning[17]>. The Rescorla–Wagner model was an important extension of the Bush and Mosteller approach to the study of what happens to associative strength when two cues predict the same event. Their findings were so influential that the basic Bush and Mosteller rule is now often mistakenly attributed to Rescorla and Wagner by neurobiologists.

References

<references>
  1. Putnam, H. (1981): "Brains in a vat" in Reason, Truth, and History, Cambridge University Press; reprinted in DeRose and Warfield, editors (1999): Skepticism: A Contemporary Reader, Oxford UP
  2. Popper, K. (1997). World of Propensities (p. 60). Thoemmes Press.
  3. Paul M. Churchland. (2012). Plato’s Camera: How the Physical Brain Captures a Landscape of Abstract Universals. MIT Press.
  4. Teyler, T. J., & DiScenna, P. (1987). Long-term potentiation. Annual review of neuroscience, 10, 131–61.
  5. Goodman, N., & Putnam, H. (1983). Fact, Fiction, and Forecast, Fourth Edition (p. 160). Harvard University Press.
  6. Popper, K. (2002). The Logic of Scientific Discovery (Routledge Classics) (p. 544). Routledge.
  7. Habermas, J. (1985). The Theory of Communicative Action, Volume 1: Reason and the Rationalization of Society (p. 465). Beacon Press.
  8. Ernst, M. (1986). On the Economical Nature of Physical Inquiry. Popular Scientific Lectures (pp. 186–214). Open Court.
  9. Wittgenstein, L. (1973). Philosophical Investigations (3rd Edition) (p. 250). Pearson.
  10. Hume D. (1748) An Enquiry Concerning Human Understanding
  11. Glimcher, P. W. (2011). Understanding dopamine and reinforcement learning: the dopamine reward prediction error hypothesis. Proceedings of the National Academy of Sciences of the United States of America, 108 Suppl (Supplement_3), 15647–54.
  12. Bush RR, Mosteller F (1951) A mathematical model for simple learning. Psychol Rev 58:313–323
  13. Pavlov IP (1927) Conditioned Reflexes: An Investigation of the Physiological Activity of the Cerebral Cortex (Dover, New York).
  14. Bush RR, Mosteller F (1951) A mathematical model for simple learning. Psychol Rev 58:313–323
  15. Bush RR, Mosteller F (1951) A model for stimulus generalization and discrimination. Psychol Rev 58:413–423.
  16. Schoenbaum G, Roesch MR, Stalnaker TA, Takahashi YK (2009) A new perspective on the role of the orbitofrontal cortex in adaptive behaviour. Nat Rev Neurosci 10:885–892.
  17. Rescorla RA, Wagner AR (1972) in Classical Conditioning II: Current Research and Theory, A theory of Pavlovian conditioning: Variations in the effectiveness of reinforcement and nonreinforcement, eds Black AH, Prokasy WF (Appleton Century Crofts, New York), pp 64–99.