Epistemology

I wrote:

The thing to do [about AI] is figure out what programming constructs are necessary to implement guesses and criticism.

Zyn Evam replied (his comments are green):

Cool. Any leads? Can you tell more? That's is what I have problems with. I cannot think of anything else than evolution to implement guesses and criticism.

the right answer would have to involve evolution, b/c evolution is how knowledge is created. i wonder why you were looking for something else.

one of the hard problems is:

suppose you:

  1. represent ideas in code, in a general way
  2. represent criticism in code (this is actually implied by (1) since criticisms are ideas)
  3. have code which correctly detects which ideas contradict each other and which don't
  4. have code to brainstorm new ideas and variants of existing ideas

that's all hard. but you still have the following problem:

two ideas contradict. which one is wrong? (or both could be wrong.)

this is a problem which could use better philosophy writing about it, btw. i'd expect that philosophy work to happen before AI gets anywhere. it's related to what's sometimes called the duhem-quine problem, which Popper wrote about too.

one of my own ideas about epistemology is to look at symmetries. two ideas contradicting is symmetric.

what do you mean by symmetries? how two ideas contradicting symmetric? could you give an example?

"X contradicts Y" means that "Y contradicts X". When two ideas contradict, you know at least one of them is mistake, but not which one. (Actually it's harder than that because you could be mistaken that they contradict.)

Criticism fundamentally involves contradiction. Sometimes a criticism is right, and sometimes the idea being criticized is right, and how do you decide which from the mere fact that they contradict each other?

With no additional information beyond "X and Y contradict", you have no way to take sides. And labelling Y a criticism of X doesn't mean you should side with it. X and Y have symmetric (equal) status. In order to decide whether to judge X or Y positively you need some kind of method of breaking the symmetry, some way to differentiate them and take sides.

Arguments are often symmetric too. E.g., "X is right because I said so" can be used equally well to argue for Y. And "X is imperfect" can be used equally well to argue against Y.

How to break this kind of symmetry is a major epistemology problem which is normally discussed in other terms like: When evidence contradicts a hypothesis, it's possible to claim the evidence is mistaken rather than the hypothesis. (And sometimes it is!) How do you decide?

So when two ideas contradict we know one of them at least is mistaken, but not which one. When we have evidence that seems to contradict a hypothesis we can never be sure that it indeed contradicts it. From the mere fact of contradiction, without additional information, we cannot decide which one is false. We need additional information.

Hypotheses are built on other hypotheses. We need to break the symmetry by looking at the hypotheses on which the contradicting ideas depend. And the question is: how would you do that? Is that right?

Mostly right. You can also look at the attributes of the contradicting ideas themselves, gather new observational data, or consider whatever else may be relevant.

And there are two separate questions:

  1. How do you evaluate criticisms at all?

  2. How do you evaluate criticisms formally, in code, for AIs?

I believe I know a lot amount about (1), and have something like a usable answer. I believe I know only a little about (2) and have nothing like a usable answer to it. I believe further progress on (1) -- refining, organizing, and clarifying the answer -- will help with solving (2).

Below I discuss some pieces of the answer to (1), which is quite complex in full. And there's even more complexity when you consider it as just one piece fitting into an evolutionary epistemology. I also discuss typical wrong answers to (1). Part of the difficult is that what most people believe they know about (1) is false, and this gets in the way of understanding a better answer.

My answer is in the Popperian tradition. Some bits and pieces of Popper's thinking have fairly widespread influence. But his main ideas are largely misunderstood and consequently rejected.

Part of Popper's answer to (1) is to form critical preferences -- decide which ideas better survive criticism (especially evidentiary criticism from challenging test experiments).

But I reject scoring ideas in general epistemology. That's a pre-Popper holdover which Popper didn't change.

Note: Ideas can be scored when you have an explanation of why a particular scoring system will help you solve a particular problem. E.g. CPU benchmark scores. Scoring works when limited to a context or domain, and when the scores themselves are treated more like a piece of evidence to consider in your explanations and arguments, rather than a final conclusion. This kind of scoring is actually comparable to measuring the length of an object -- you define a measure and you decide how to evaluate the resulting length score. This is different than an epistemology score, universal idea goodness score, or truth score.

I further reject -- with Popper -- attempts to give ideas a probability-of-truth score or similar.

Scores -- like observations -- can be referenced in arguments, but can't directly make our decisions for us. We always must come up with an explanation of how to solve our problem(s) and expose it to criticism and act accordingly. Scores are not explanations.

This all makes the AI project harder than it appears to e.g. Bayesians. Scores would be easier to translate to code than explanations. E.g. you can store a score as a floating point number, but how do you store an explanation in a computer? And you can trivially compare two scores with a numerical comparison, but how do you have a computer compare two explanations?

Well, you don't directly compare explanations. You criticize explanations and give them a boolean score of refuted or non-refuted. You accept and act on a single non-refuted explanation for a particular problem or context. You must (contextually) refute all the other explanations, rather have one explanation win a comparison against the others.

This procedure doesn't need scores or Popper's somewhat vague and score-like critical preferences.

This view highlights the importance of correctly judging whether an idea refutes another idea or not. That's less crucial in scoring systems where criticism adds or subtract points. If you evaluate one issue incorrectly and give an idea -5 points instead of +5 points, it could still end up winning by 100 points so your mistake didn't really matter. That's actually bad -- it essentially means that issue had no bearing on your conclusion. This allows for glossing over or ignoring criticisms.

A correct criticism says why an idea fails to solve the problem(s) of interest. Why it does not work in context. So a correct criticism entirely refutes an idea! And if a criticism doesn't do that, then it's harmless. Translating this to points, a criticism should either subtract all the points or none, and thus using a scoring system correctly you end up back at the all-or-nothing boolean evaluation I advocate.

This effectively-boolean issue comes up with supporting evidence as well. Suppose some number of points is awarded for fitting with each piece of evidence. The points can even vary based on some judgement of how importance each piece of evidence is. The importance judgement can be arbitrary, it doesn't even matter to my point. And consider evidence fitting with or supporting a theory to refer to non-contradiction since the only known alternatives basically consist of biased human intuition (aka using unstated, ambiguous ideas without figuring out what they are very clearly).

So you have a million pieces of evidence, each worth some points. You may, with me, wish to score an idea at 0 points if it contradicts a single piece of evidence. That implies only two scores are possible: 0 or the sum total of the point value of every piece of evidence.

But let's look at two ways people try to avoid that.

First, they simply don't add (or subtract) points for contradiction. The result is simple: some ideas get the maximum score, and the rest get a lower score. Only the maximum score ideas are of interest, and the rest can be lumped together as the bad (refuted) category. Since they won't be used at all anyway, it doesn't matter which of them outscore the others.

Second, they score ideas using different sets of evidence. Then two ideas can score maximum points, but one is scored using a larger set of evidence and gets a higher score. This is a really fucked up approach! Why should one rival theory be excluded from being considered against some of the evidence? (The answer is because people selectively evaluate each idea against a small set of evidence deemed relevant. How are the selections made? Biased intuition.)

There's an important fact here which Popper knew and many people today don't grasp. There are infinitely many theories which fit (don't contradict) any finite set of evidence. And these infinitely many theories include ones which offer up every possible conclusion. So there are always max-scoring theories, of some sort, for every position. Which makes this kind of scoring end up equivalent to the boolean evaluations I advocated in the first place. Max-score or not-max-score is boolean.

Most of these infinitely many theories are stupid which is why people try to ignore them. E.g. some of the form, "The following set of evidence is all correct, and also let's conclude X." X here is a completely unargued non sequitur conclusion. But this format of theory trivially allows a max-score theory for every conclusion.

The real solution to this problem is that, as Deutsch clearly explained in FoR (with the grass cure for the cold example), most bad ideas are rejected without experimental testing. Most ideas are refuted on grounds like:

  1. bad explanation

I was going to make a longer list, but everything else on my list can be considered a type of bad explanation. The categorizations aren't fundamental anyway, it's just organizing ideas for human convenience. A non sequitur is a type of bad explanation (non explanation). And a self-contradictory idea is a type of bad explanation too. And having a bad explanation (including none) of how it solves the problem it's supposed to solve is another important case. That gets into something else important which is understood by Popper and partly by Rand, but isn't well known:

Ideas are contextual. And the context is, specifically, that they address problems. Whether a criticism refutes an idea has to be evaluated in a particular context. The same idea (as stated in English) can solve one problem and fail to solve another problem. One way to approach this is to bundle ideas with their context and consider that whole thing the idea.

Getting back to the previous point, it's only ideas which survive our initial criticism (including doesn't blatantly contradict evidence we know offhand) that we take more interest in them and start carefully comparing them against the evidence and doing experimental tests. Testing helps settle a small number of important cases, but isn't a primary method. (Popper only partly understood this, and Deutsch got it right.)

The whole quest -- to judge ideas by how well (degree, score) they fit evidence -- is a mistake. That's a dead end and distraction. Scores are a bad idea, and evidence isn't the the place to focus. The really important thing is evaluating criticism in general, most of which broadly related to: what makes explanations bad?

BTW, what is an explanation? Loosely it's the kind of statement which answers why or how. The word "because" is the most common signal of explanations in English.

Solving problems requires some understanding of 1) how to solve the problem and 2) why that solution will work (so you can judge if the solution is correct). So explanation is required at a basic level.

So, backing up, how do you address all those stupid evidence-fitting rival ideas? You criticize them (by the category, not individually) for being bad explanations. In order to fit the evidence and have dumb conclusion, they have to have a dumb part you can criticize (unless the rival idea actually isn't so dumb as you thought, a case you have to be vigilant for). It's just not an evidence-based criticism (and nor should the criticism by done with unstated, based commonsense intuitions combined with frustration at the perversity of the person bringing an arbitrary, dumb idea into the discussion). And how do you address the non-evidence-fitting rival ideas? By rejecting them for contradicting the evidence (with no scoring).

Broadly it's important to take seriously that every flaw with an idea (such as contradicting evidence, having a self-contradiction, having a non sequitur, or having no explanation of how or why it solves the problem it claims to solve) either 1) ruins it for the problem context or 2) doesn't ruin it. So every criticism is either decisive or (contextually) a non-criticism. So evaluations of ideas have to be boolean.

There is no such thing as weak criticism. Either the criticism implies the idea doesn't solve the problem (strong criticism), or it doesn't (no criticism). Anything else is, at best, more like margin notes which may be something like useful clues to think about further and may lead to a criticism in the future.

The original question of interest was how to take sides between two contradicting ideas, such as an idea and a criticism of it. The answer requires a lot of context (only part of which I've covered above), but then it's short: reject the bad explanations! (Another important issue I haven't discussed is creating variants of current ideas. A typical reaction to a criticism is to quickly and cheaply make a new idea which is a little different in such a way that the criticism no longer applies to it. If you can do this without ruining the original idea, great. But sometimes attempts to do this run into problems like all the variants with the desired-traits-to-address-the-criticism ruin the explanation in the original idea.)


Elliot Temple | Permalink | Messages (0)

Valuing Criticism

I wrote to the Fallible Ideas discussion group:

this reminds me of a question: did you find many mistakes in Mises and others when reading them?

Zyn Evam replied (all green quotes):

not many :/

better discuss more!!!!!

and study critical thinking (a branch of philosophy) more!!

how should I do that? what is the best way?

write short posts to FI. daily.

there is no study guide or prepackaged life plan for this. no "learn this then this then this then this and then you're awesome" and you just follow the instructions. you have to lead yourself.

i can give examples of the kinds of things that are good to do. but don't just do my specific examples as if they were a curriculum.

you could talk about what you know and don't know already, and what problems you see with that and what you think is good about it, and ask what problems others see. you could talk about what if anything you think you might need to learn and why.

you could talk about what you think your problems in your life are and your current plans for improving and ask for any better ideas.

One problem I have is not writing much to FI, or writing sporadically, and then stopping. I believe I understand the value of others supplying criticism for my ideas, but I haven't integrated it much with my life. That is not enough. It is not like I feel urged to share my ideas so that others can find faults in them. I think that should be a thing to aim for. I should be excited about others pointing out I am wrong.

It helps to conceptualize "find faults" as "find opportunities for improvement".


It helps to value something highly. Some people value truth, but other values work too. Some people really want to win at video game or sports competitions and they form a good attitude to criticism to help them achieve that goal.

Valuing something highly handles layers of indirection better. If you care a ton about about D, that helps you care about A to help with B to help with C to help with D.

For example, consider a typical person who cares a little about their car and has no interest in paint. Then he won't want to learn about car detailing paints and brushes. He'll only do that if he feels pressured to by e.g. a highly visible scratch. But people who care a ton about their car often form some interest in car-related paint so they can improve their car. And the person who cares even more may learn about mixing custom paint or even manufacturing a new type of paint with different chemistry.

It helps to be interested in stuff in multiple ways. The guy who learns about paint chemistry and manufacturing generally either

1) already had some separate interest in science and business

or

2) he tried looking into them for his car. but once he got started on them, he found he liked them for some reasons independent of his car. so even if he stopped driving and sold his car, he still might continue with them.


It helps to slow down and pay more attention to your life. Try to be consciously aware of what you're doing and intentionally choose it according to some reasoning, rather than get "sucked in" to activities. If you can do that in general with your current activities, it will put you in a better position to make reasoned changes.

Don't try to change everything at once. If you can be more self-aware of your current activities, without changing them, that's a good step. Then you'll be in a better position to evaluate them and decide what, if anything, you actually want to change.


It helps to think about philosophical problems and connections in a regular basis, during your life. Like considering how philosophical issues are relevant to what you're doing and philosophical answers could help with it. You can do this intentionally if it doesn't come naturally to you. E.g. you can take regular 5 minute breaks to do it.

E.g. you could notice you're trying to do something difficult and you want it to work instead of not work, but you have a significant concern that it may not work. Then philosophy about errors is especially relevant. Is there a way to proceed so that error is impossible? Knowing the answer to that matters. If the answer is yes, it'd be good to find out the method. If the answer is no, then is there anything to be done about error besides fatalistically put up with it? etc

Or you could decide you need to learn a new skill for your project. Then philosophy about learning is relevant. Are there better and worse ways to learn? What causes some attempts to learn things to fail? How does one learn faster or better? That's all useful.

Are these kinds of things too abstract for you? You can concretize. The book Understanding Objectivism helps. You can find relevant sections if you search it for terms like "concretize", "concretizing" and "chewing".

Peikoff talked somewhere about his experiences learning from Ayn Rand. She'd tell him an idea, and then he'd go out in the world and notice it in a bunch of places and see it for himself and connect it to a bunch of concretes.


I thought my problem was finding the time. But it has more to do with preferences.

You have something like a prioritized list of stuff you want to do. You look at it and think the stuff above philosophy will take up all your time. You think if you could finish the short term stuff near the top of the list, then you'd have enough time for philosophy.

But you'll always add new things to the list. There's plenty of stuff you could do. So what matters most is how highly prioritized philosophy is. Raising philosophy's prioritization will make a bigger difference than freeing up some time by clearing some things off the list.

Specifically, if you prioritize philosophy above most incoming things being added to your list, then you'll do it often. As a loose approximation, you can think of the incoming new stuff to do as being on a bell curve with a mean of 100 priority and a standard deviation of 15. Then if you prioritize philosophy at 90, it doesn't have much chance. But if you prioritize philosophy at 130, then around 98% of the new additions to your list will be inserted below philosophy.


Elliot Temple | Permalink | Messages (3)

Peikoff Getting Parmenides Wrong

Understanding Objectivism by Leonard Peikoff:

What is the name for the type of person in philosophy who clings to concepts and says, in effect, “Facts may contradict my concepts, but if so, it’s tough on the facts”? A “rationalist.” Rationalism has dominated philosophy (at least the better philosophy) through the ages. Starting way back with Parmenides, who gave an argument as to why change is impossible, and then saw things change in front of his eyes, and said, “They’re not really changing, because that simply does not agree with my unanswerable argument.”

I generally agree with Peikoff's point. I'm going to criticize only the comment on Parmenides.

For context, Peikoff is no casual commenter on the history of philosophy. He's a teacher who's lectured on it. And he considers the lectures good enough to sell:

http://www.peikoff.com/courses_and_lectures/the-history-of-philosophy-volume-1-–-founders-of-western-philosophy-thales-to-hume/

A reader would reasonably expect Peikoff to know what he's talking about regarding Parmenides, and not to have made this statement carelessly. I think Peikoff's own position would be, "I'm familiar with Parmenides and I'm right" rather than, "You're being too picky and can't expect me to know much about Parmenides".

Now, from The World of Parmenides by Karl Popper:

Parmenides was a philosopher of nature (in the sense of Newton’s philosophia naturalis). A whole series of highly important astronomical discoveries is credited to him: that the Morning Star and the Evening Star are one and the same; that the Earth has the shape of a sphere (rather than of a drum of a column, as Anaximander thought). About equally important is his discovery that the phases of the Moon are due to the changing way in which the illuminated half-sphere of the Moon is seen from the Earth.[5]”

So Parmenides was a scientist who made empirical discoveries. He wasn't a rationalist who refused to look at the world.

And Parmenides' main work was about the conflict he found between appearance and reality. As a scientist, Parmenides discovered some ways that appearance and reality don't match. E.g. the Earth looks flat but is spherical. This conflict stood out to Parmenides and interested him, so he wrestled with it, tried to make sense of it, and wrote about it.

Popper proposes Parmenides first made an empirical discovery that the moon doesn't change. Then second, in trying to grapple with it, he came up with the idea that actually nothing changes. [1]

Discovering the difference between appearance and reality is a big deal. It's a hard problem. Early work on the matter wasn't up to modern standards and people got confused, but that doesn't make Parmenides anything resembling a modern rationalist.

So what Peikoff said about Parmenides is completely wrong.


[1] This idea is a lot better than it sounds, by the way. It has some similarities to e.g. modern spacetime theory. Time is tricky and commonsense ideas about time are wrong (see David Deutsch's books for details).

And there was a successor theory to Parmenides', which was that reality consists of atoms and the void (rather than just atoms with no empty space), and the atoms don't change but do move. So Parmenides' idea was fruitful, it helped discuss an important problem and lead to some better ideas.

Parmenides' idea may also have been related to Xenophanes' religious ideas which also have some value (rejecting anthropomorphic gods, monotheism, differentiating perfect/divine truth from fallible human knowledge similarly to Parmenides). Parmenides was a pupil of Xenophanes.


Elliot Temple | Permalink | Messages (0)

Elliot Temple | Permalink | Messages (0)

Churchill and Roosevelt Betrayed Hundreds of Thousands to Death

The Gulag Archipelago by Aleksandr Solzhenitsyn has a lot of stories about how evil the Soviet gulag system was. Below I quote a brief section about Western complicity in Soviet crimes, and about Churchill and Roosevelt in particular.

Context for reading the quote: The "Vlasov army" refers to Russian units in the German army in WWII. They turned against the Germans and saved Prague before the Soviet army arrived. (Soviet histories lie and take credit.) The "act of a loyal ally" refers to the Vlasov army betraying the Germans. And the Soviets routinely punished anyone who'd been taken prisoner by the Germans with 15 years in the gulag (which is preceded by torture). Many other people were executed. One reason is they didn't want people who knew about life in Europe to spread information about it in Russia. The Soviets were so unfairly cruel and murderous it's hard to believe if you haven't read about it. Many people would rather have killed themselves than be executed or tortured and imprisoned by the Soviets. Keep that context in mind when considering turning anyone over to the Soviets.

After saving Prague (bold added):

... the Vlasov army began to retreat toward Bavaria and the Americans. They were pinning all their hopes on the possibility of being useful to the Allies; in this way their years of dangling in the German noose would finally become meaningful. But the Americans greeted them with a wall of armor and forced them to surrender to Soviet hands, as stipulated by the Yalta Conference. In Austria that May, Churchill perpetrated the same sort of "act of a loyal ally," but, out of our accustomed modesty, we did not publicize it. He turned over to the Soviet command the Cossack corps of 90,000 men.

[This surrender was an act of double-dealing consistent with the spirit of traditional English diplomacy. The heart of the matter was that the Cossacks were determined to fight to the death, or to cross the ocean, all the way to Paraguay or Indochina if they had to ... anything rather than surrender alive. Therefore, the English proposed, first, that the Cossacks give up their arms on the pretext of replacing them with standardized weapons. Then the officers—without the enlisted men—were summoned to a supposed conference on the future of the army in the city of Judenburg in the English occupation zone. But the English had secretly turned the city over to the Soviet armies the night before. Forty busloads of officers, all the way from commanders of companies on up to General Krasnov himself, crossed a high viaduct and drove straight down into a semicircle of Black Marias, next to which stood convoy guards with lists in their hands. The road back was blocked by Soviet tanks. The officers didn't even have anything with which to shoot themselves or to stab themselves to death, since their weapons had been taken away. They jumped from the viaduct onto the paving stones below. Immediately afterward, and just as treacherously, the English turned over the rank-and-file soldiers by the trainload—pretending that they were on their way to receive new weapons from their commanders.

In their own countries Roosevelt and Churchill are honored as embodiments of statesmanlike wisdom. To us, in our Russian prison conversations, their consistent shortsightedness and stupidity stood out as astonishingly obvious. How could they, in their decline from 1941 to 1945, fail to secure any guarantees whatever of the independence of Eastern Europe? How could they give away broad regions of Saxony and Thuringia in exchange for the preposterous toy of a four-zone Berlin, their own future Achilles' heel? And what was the military or political sense in their surrendering to destruction at Stalin's hands hundreds of thousands of armed Soviet citizens determined not to surrender? They say it was the price they paid for Stalin's agreeing to enter the war against Japan. With the atom bomb already in their hands, they paid Stalin for not refusing to occupy Manchuria, for strengthening Mao Tse-tung in China, and for giving Kim Il Sung control of half Korea! What bankruptcy of political thought! And when, subsequently, the Russians pushed out Mikolajczyk, when Benes and Masaryk came to their ends, when Berlin was blockaded, and Budapest flamed and fell silent, and Korea went up in smoke, and Britain's Conservatives fled from Suez, could one really believe that those among them with the most accurate memories did not at least recall that episode of the Cossacks?]

Along with them, he also handed over many wagonloads of old people, women, and children who did not want to return to their native Cossack rivers. This great hero, monuments to whom will in time cover all England, ordered that they, too, be surrendered to their deaths.


Elliot Temple | Permalink | Messages (2)

Stop Saying Lies and Other People's Ideas

Maps of Meaning: The Architecture of Belief by Jordan Peterson

I started to hear a “voice” inside my head, commenting on my opinions. Every time I said something, it said something – something critical. The voice employed a standard refrain, delivered in a somewhat bored and matter-of-fact tone:

You don’t believe that.

That isn’t true.

You don’t believe that.

That isn’t true.

The “voice” applied such comments to almost every phrase I spoke.

I couldn’t understand what to make of this. I knew the source of the commentary was part of me – I wasn’t schizophrenic – but this knowledge only increased my confusion. Which part, precisely, was me – the talking part, or the criticizing part? If it was the talking part, then what was the criticizing part? If it was the criticizing part – well, then: how could virtually everything I said be untrue? In my ignorance and confusion, I decided to experiment. I tried only to say things that my internal reviewer would pass unchallenged. This meant that I really had to listen to what I was saying, that I spoke much less often, and that I would frequently stop, midway through a sentence, feel embarrassed, and reformulate my thoughts. I soon noticed that I felt much less agitated and more confident when I only said things that the “voice” did not object to. This came as a definite relief. My experiment had been a success; I was the criticizing part. Nonetheless, it took me a long time to reconcile myself to the idea that almost all my thoughts weren’t real, weren’t true – or, at least, weren’t mine.

All the things I “believed” were things I thought sounded good, admirable, respectable, courageous. They weren’t my things, however – I had stolen them. Most of them I had taken from books. Having “understood” them, abstractly, I presumed I had a right to them – presumed that I could adopt them, as if they were mine: presumed that they were me. My head was stuffed full of the ideas of others; stuffed full of arguments I could not logically refute. I did not know then that an irrefutable argument is not necessarily true, nor that the right to identify with certain ideas had to be earned.

wise, IMO.

ppl overreach by saying a bunch of crap instead of actually doing stuff right and thinking. (and if u recommend they slow down, they often bring up the issue that zero would be a bad amount to talk, too. and then you see them say something really careless they spent 2 minutes on. why can't they consistently spend, say, 5 minutes reviewing each of their posts -- more for really long ones, but don't do those anyway -- and send if everything looks good? that should easily get them a more medium result between rushed and nothing.)

a common, important tip for learning is: better to do something correctly, slowly, then speed up. don't go faster than you know what you're doing and try to fix the mistakes later. this applies to learning to touch type, learning video games, and also writing an FI reply.

Peterson also said in a video somewhere, something like: most of what people say is lies or other people's ideas. they don't have their own ideas or a self. they need to create that. i wonder if he's read The Fountainhead.*

in another video, Peterson said basically that people have been building up lies on top of lies on top of lies, for decades. that's why they have such difficult problems! that's why their lives are such a mess! it's layer and layers and layers of lies to untangle!


Elliot Temple | Permalink | Messages (3)

Teachable Subjects

in some sense, people teach chess, math, Spanish, grammar, history, programming, formal and symbolic logic, chemistry, physics, biology, how our government works, how to cook.

they have school classes on these things. they have books explaining it. i can remember being taught about these things in the past.

by contrast, i cannot remember anyone ever teaching me:

  • how to understand what a sentence means
  • how to figure out if an argument is true or false
  • what an argument IS (and how to decide if X is an argument for Y)
  • how to decide if something is a non sequitur
  • how to decide if X implies Y
  • how to read sentences that don't not have double, triple or quadruple negatives.

people do teach relevant things. reading what the words are is relevant to understanding a written sentence. understanding !!X is relevant to double negatives. but using it is kinda up to you. people also do teach you to avoid using double negatives, and warn they are confusing.

people teach syllogisms. but that's a bad way to think about most arguments. it can make things worse!

people get examples. lots. kids hear many sentences. and they get information about what the sentence was about. like the parent says something including "pancakes" and then 20 minutes later breakfast is served. repeatedly.

i think a lot of how people understand sentences is actually like that. they map the sentence to: "blah blah something about pancakes" and think " cool i understand him". and when you're 3, hey, that's pretty good! success! but when you're 13 or 33, that's bad. but the 33 yo actually interprets tons of stuff that way and muddles through life.

and one of the things that makes it hard is the people talking are as dumb as you. so they say some actual specific words. but they don't know what words mean all that well either. so all they even MEANT is "blah blah something about pancakes". or maybe they had something more complicated in their head, but it didn't match the words they used anyway. that's not universal (even with dumb people) but it's common. a lot of times, "blah blah something about pancakes" is about as good as you can do because the speaker actually didn't use the right words to communicate more. if you try to listen to the details of what he said, and interpret them, you'll just misunderstand him!

people are really bad at explaining or teaching anything they find "self-evident" or super super obvious. that's part of the issue.

Rand, Popper and DD are great at this stuff, but their books don't teach it. they explain more advanced stuff.


anyway, i was thinking there are these subjects people know how to teach (often quite badly). and subjects they basically don't teach at all. and i don't think the other ones are impossible to teach btw, people just don't know how much or don't do it or really really suck at it.

some of the stuff they don't teach is basically the kinda stuff that IQ is about. people don't know how to teach IQ. (btw there's also other stuff they don't like teaching like swear words or sex stuff).

in more Popperian terms people need to do steps like:

1) understand, conceptually, what the problem they are trying to solve is.

and ppl get that wrong all the time. and no one really teaches it as a general skill. or knows how to. it is taught in specific ways, like they'll teach you about a particular category of chemistry problem and how to think about it and what to do with it.

2) brainstorm solutions

how do you brainstorm? there's infinite things you could come up with. which ones should you? how do you know? this isn't really taught. isn't not very important though, just don't get stuck here. the initial brainstorming can be shitty ideas, and that gets fixed in step 4.

3) criticize the brainstormed stuff

here is where you have to actually figure out what kinda stuff to target. you both think of attributes the ideas/examples/solutions/whatever should not have for some reason ("let's not use an animal example b/c ppl are confused about how animals differ from humans. an inanimate object will be clearer") and also attributes it should have ("i want a solution that leaves me with at least as much money as i had before, so i'm not gonna do X or Y").

4) judgement, like which criticisms apply to which ideas.

this is partly hard just in a basic way. if trait Z is bad, ok, well, which ideas have trait Z and which don't? how do you figure that out if you don't already know it?

and it's partly hard in a more complicated way b/c you don't wanna just throw out a bunch of ideas b/c of a dumb criticism. you also need to be judging the criticisms too and making counter-criticisms. that's so complicated it actually kinda ruins my attempt to make this a linear step-by-step process. i just threw it in here.

5) brainstorm variants

so the initial brainstorming can just be rather random crap. that's fine. i don't think any healthy adult actually has much trouble with that, even if they can't really say how they do it.

but this part is harder. where you're coming up with ideas that meet criteria you had from (3). it's like, how do i change the solution to leave me with more money? what are some inanimate objects? (ok that one is easy, but some are harder).

one reason people get stuck on (2) is they already know some criteria of criticism. so they skip (2) and do (3) first, and then move on. that's fine. no problem at all. it just doesn't count as being stuck on (2) if they are really stuck on (5).

6) judge when to stop

when is the idea good enough? how much more should you think of criticism and brainstorm better solutions? this isn't taught. people guess wildly and sometimes make corrections to their policies (like they realize to spend more time than normal for important stuff).


btw, i have not taught it to you here. this is summarizing and describing it. it doesn't actually teach you how to actually do it. it gives you some hints from which you might figure it out yourself. and in some sense that's all we ever give students. but there are lots of topics where the hints are way better and include actual explanations of how to do something. my 6 point list is not what i'd consider teaching it in the usual sense. it's talking about it and it may be helpful, but it really leaves a lot up to you to figure out how actually do the things i mention.

and there's other stuff besides the 6 things on my list and the points earlier like about non sequiturs and figuring out what's an argument for what and how that works and what would and wouldn't be a counter argument. but like, these are basic things people are bad at, and it's the kinda thing that matters to IQ, and people don't teach it.

it's also hard to get ppl to try to learn it. people on FI want to do things like learn Objectivism when they can barely read sentences, you know? they don't sit there and go over the basics.

part of the issue is people can read sentences. 90% of the time! but it's like, even a little idea involves using some basic skills a bunch of times. 50 tiny little basic things might go into 1 idea. so even if you're 99% reliable and doing the basic stuff right, tons of your ideas will be wrong! you need to be able to do the basic stuff with a VERY good success rate or you get totally overwhelmed with errors when you try to build complex mental structures out of millions of basic components. but people generally don't like practicing stuff they get right over 80% of the time. they don't like trying to go from 99% to 99.999%. and besides being HIGHLY reliable at the basics, people also need to be FAST at them. if you're going to build complex mental structures out of a million little pieces, you better be able to do most of the little pieces in well under a second. but people also don't like practicing to get faster at stuff they are already fast at. they don't like trying to go from 2 seconds to .2 seconds to .02 seconds.


morality is another thing people are bad at teaching. there are people who are good at explaining it like Ayn Rand. but like, i watched Pinocchio (the old disney movie) and the cricket (his conscience for some reason) gives Pinocchio a lecture on morality. it's something about avoiding temptation and doing the right thing. it's completely incoherent and Pinocchio doesn't understand. part of the actual plot is this incoherent moralizing that is not understood. and then Pinocchio is immediately thrust out into the world to face temptations to do other stuff besides go to school. and the movie illustrates, in a magical, exaggerated way with 2 unrealistic examples, how nice sounding things can be dangerous and he should have resisted temptation and gone to school.

i think this fits children's actual experiences pretty well: incoherent verbal moral advice they don't understand at home, followed by being thrust into the world totally unprepared.

a lot of how the Bible teaches morality is with stories, too. and there are other old stories with moral content, like fairytales.


Elliot Temple | Permalink | Messages (12)

John Locke's Politics

Locke had some good ideas about political philosophy:

  • People have natural rights to life, liberty and property.
  • Natural rights make sense according to reason and are also God's will.
  • Individual rights limit what other citizens and the government can do to individuals.
  • The government and law should treat everyone equally. Everyone is politically equal.
  • The government has limited power, not arbitrary or unlimited power.
  • The government's legitimacy comes from the consent of the citizens who prefer the government over the state of nature (anarchy).
  • The reason people form a government is to help resolve conflicts and protect rights.
  • The people in the government should work to benefit society, not for personal gain.
  • If a government is bad enough, the people have a right to rebel against it.

But Locke had some very bad ideas about education.

This info is from my newsletter. Read the whole thing to find out about Locke's nasty education views, the Barbary pirates, and some political links.

And sign up for the free newsletter. (2-4 emails/month)


Elliot Temple | Permalink | Messages (0)