Archive 1Archive 3Archive 4Archive 5

Eliminated prevarication in logical expressions

I have made some minor revisions in the description of the experiment to eliminate the prevarication present in the descriptions of the logical expressions there. For instance, it is unnecessary, when substituting a value for a variable in an expression that is otherwise and generally held to be true, to add words to infer that the expression concerned may thereby become untrue (for reasons unknown). If the expression is true then using it with different values does not change that. IF the expression "People have hearts" is true, THEN substituting the value "Mary" or John" results in a true expression. It is redundant, and implies a fundamental but unreasoned doubt, to embellish the result into the form: "People have hearts and so the doctor argues that Mary has a heart". It is a true consequence of people having hearts that Mary also has a heart, unless e.g. Mary is a ship, has had her heart removed, etc. --LookingGlass (talk) 14:01, 4 December 2012 (UTC)

The following got deleted from the article back in December 2010. This should be undone.

  • An episode of the TV show Numb3rs uses the Chinese room as a model in the episode "Chinese Box".
  • Searle's Chinese room is referenced in the novel Blindsight by Peter Watts by characters dealing with an alien entity that is intelligent but lacks consciousness.

Additionally the section should mention that in Daniel Cockburn's 2010 film You Are Here Searle's Chinese room argument is enacted on screen. (source e.g. http://www.filmcomment.com/article/you-are-here) — Preceding unsigned comment added by 84.41.34.154 (talkcontribs)

These don't sound significant enough to warrant a mention; see wp:UNDUE: "An article should not give undue weight to any aspects of the subject but should strive to treat each aspect with a weight appropriate to its significance to the subject. For example, discussion of isolated events, criticisms, or news reports about a subject may be verifiable and NPOV, but still be disproportionate to their overall significance to the article topic." Are any of these really important to the reader's understanding of the subject? ErikHaugen (talk | contribs) 05:56, 16 July 2012 (UTC)
WP:UNDUE is about proportion and relative weight. Since this is a very long article and quite mature, it seems the most significant aspects of the subject may already be well covered. I'm not vouching for those particular pop culture examples above, but I think a few sentences on the Chinese room usage in pop culture would not overwhelm the significant weight already in the article. $0.25. --Ds13 (talk) 06:21, 16 July 2012 (UTC)
Well, maybe. Look at lion, for example—lions are pretty significant in culture, and cultural depictions/etc are important to understanding the topic. Maybe that is true of this subject as well and these examples are good ones that demonstrate that. I really don't think so, though. ErikHaugen (talk | contribs) 06:39, 16 July 2012 (UTC)

Many mansions - missing?!

Where's the discussion of the many mansions reply, which is probably the most famous counter-argument to the Chinese room? Raul654 (talk) 07:45, 15 April 2011 (UTC)

While I agree that the reply deserves to be discussed, it is not my impression that it is the most famous counter-argument. Can you back up that assertion? (I believe the "systems" reply is the most popular.) Looie496 (talk) 16:44, 15 April 2011 (UTC)
Well, I've heard of that (mansions) one and not the others, so I guess I was just assuming it's the most famous. But either way, it does deserve to be discussed. Raul654 (talk) 16:48, 15 April 2011 (UTC)
The "Many Mansions" reply should be covered.
In my view, it belongs in the section currently called redesigning the room, because Searle's counter-argument is the same: he says that the reply abandons "Strong AI" as he has defined it. The Chinese Room argument only applies to digital machines manipulating formal symbols. He grants that there may some other means of producing intelligence and consciousness in a machine. He's only arguing that "symbol manipulation" won't do it. ---- CharlesGillingham (talk) 06:35, 5 June 2011 (UTC)
  Fixed (sometime ago) ---- CharlesGillingham (talk) 19:01, 5 September 2013 (UTC)

Compare "Chinese room" to search results of "optimal classification"

Do a Google search on "optimal classification". Then compare results to understand limits, if any. — Preceding unsigned comment added by 71.99.179.45 (talk) 03:39, 3 November 2013 (UTC)

Um, what's your point? Looie496 (talk) 15:07, 3 November 2013 (UTC)

Actual explanation of Chinese room in first paragraph seems wrong/makes no sense

I am not a philosopher or philosophy student so I am hesitant to make this change on my own, but I could make no sense at all of these sentences: "It supposes that there is a program that gives a computer the ability to carry on an intelligent conversation in written Chinese. If the program is given to someone who speaks only English to execute the instructions of the program by hand, . . ." (after that I understand). A search brought me to the Stanford Encyclopedia of Philosophy page [1], which has a far clearer and substantially different explanation that does not involve a computer program or someone executing a computer program's instructions "by hand."

Again, I am so far from being an expert on this I can't bring myself to make the change, but I hope someone who knows what they are doing will consider revising for correctness and clarity. Kcatmull (talk) 20:21, 24 July 2013 (UTC)

Can you elaborate on what part of this is unclear? ErikHaugen (talk | contribs) 06:59, 29 July 2013 (UTC)
The explanation assumes that the reader knows that a "program" is a series of instructions - this could be worded more clearly, I'll take a stab at it. --McGeddon (talk) 09:28, 29 July 2013 (UTC)
No, that's not quite it, sorry I was not clear. I think the explanation as written brings computers in too early in the explanation. Here's how the explanation at Stanford (http://plato.stanford.edu/entries/chinese-room/ ) begins: "The argument centers on a thought experiment in which someone who knows only English sits alone in a room following English instructions for manipulating strings of Chinese characters, such that to those outside the room it appears as if someone in the room understands Chinese. The argument is intended to show that while suitably programmed computers . . ." etc. So the thought experiment (as I understand it) is comparing a computer to a PERSON sitting alone in a room with a set of instructions, etc. But the first sentence of this article mentions a COMPUTER being given such a set of instructions, which is not the thought experiment; and then the second sentence confusingly mentions a person ("someone"), so the whole thing is a hopeless tangle. I think better to describe the thought experiment first -- English-speaking person sitting in a room with instructions: do they actually speak Chinese?--and then show that this is (or isn't) the situation of computers. Does that make more sense? Kcatmull (talk)
This is tricky stuff, and it's easy when trying to improve wording to actually make it worse. Could you propose a specific wording that could be substituted for the original? Looie496 (talk) 16:24, 5 August 2013 (UTC)
I think there is a problem here. Was the lead always like this? Sadly, S himself sometimes seems to start off in this way. cf. Minds, Brains, and Science: "Imagine that a bunch of computer programmers ..." Then, later, "Well, imagine that you are locked in a room ...". I think we should have a go at it. Myrvin (talk) 19:03, 5 August 2013 (UTC)
However, Minds, Brains and Programs has: "the following Gedankenexperiment. Suppose that I'm locked in a room and given a large batch of Chinese writing. Suppose furthermore (as is indeed the case) that I know no Chinese, either written or spoken". So that doesn't start with the computer program. Myrvin (talk) 20:23, 5 August 2013 (UTC)
I suggest that the first part says what the argument is supposed to do. The next should have the person in the room. Then introduce the analogy with a computer program Myrvin (talk) 20:27, 5 August 2013 (UTC)
It could begin:

The Chinese room is a thought experiment presented by John Searle in order to challenge the concept of strong AI, that a machine could successfully perform any intellectual task that a human being can.[2] Searle writes in his first explication: "Suppose that I'm locked in a room and ... that I know no Chinese, either written or spoken". He further supposes that he has a set of rules in English that "enable me to correlate one set of formal symbols with another set of formal symbols." These rules allow him to respond, in written Chinese, to questions, also written in Chinese, in such a way that the posers of the questions - who do understand Chinese - are convinced that Searle can actually read and write Chinese, even though he cannot. Myrvin (talk) 20:52, 5 August 2013 (UTC)

The definition of "strong AI" above is not quite right. "Strong AI" is the idea that a suitably programmed computer could have a mind and consciousness in the same sense human beings do. Note that it is possible for a computer to perform any intellectual task without having a mind or consciousness. ---- CharlesGillingham (talk) 21:51, 5 August 2013 (UTC)
Actually neither of those is quite right. Strong AI as Searle defines it is the claim that mind and consciousness are merely matters of executing the right programs. That isn't the same as being a computer, because computers can do more than execute programs -- for example, they can cause pixels to light up on video screens. Looie496 (talk) 05:49, 6 August 2013 (UTC)
I stole the words, verbatim, from the strong AI article. We can agree on something better. I suppose Searle's def should be there since that's what he's contesting. Myrvin (talk) 06:53, 6 August 2013 (UTC)
How about:

The Chinese room is a thought experiment presented by John Searle in order to challenge the claims of strong AI (strong artificial intelligence). According to Searle, when referring to a computer running a program intended to simulate human ability to understand stories: "Partisans of strong Al claim that the machine is not only simulating a human ability but also (1) that the machine can literally be said to understand the story and provide the answers to questions, and (2) that what the machine and its program do explains the human ability to understand the story and answer questions about it."[2] In order to contest this view, Searle writes in his first description of the argument: "Suppose that I'm locked in a room and ... that I know no Chinese, either written or spoken". He further supposes that he has a set of rules in English that "enable me to correlate one set of formal symbols with another set of formal symbols," that is, the Chinese characters. These rules allow him to respond, in written Chinese, to questions, also written in Chinese, in such a way that the posers of the questions - who do understand Chinese - are convinced that Searle can actually understand the Chinese conversation too, even though he cannot. Similarly, he argues that if there is a computer program that allows a computer to carry on an intelligent conversation in written Chinese, the computer executing the program would not understand the conversation either. Myrvin (talk) 13:16, 6 August 2013 (UTC)

That seems fine to me, though I found Myrvin's first shot at it (above) quite a bit clearer. But this is a huge improvement over what's there now. I hope you make the change! And thank you! Kcatmull (talk) —Preceding undated comment added 20:08, 6 August 2013 (UTC)
I filled in a correct definition of Searle's strong AI, and removed the link to Kurzweill's strong AI. It is impossible to understand Searle's argument if you don't know what he's arguing against.
One more thought: I think that the key to understanding the argument is to let go of the anthropomorphic assumption that "consciousness" and "intelligent behavior" are the same thing. So it's important that we keep these separated for the reader wherever possible, and thus the distinction between Kurzweill's strong AI and Searle's "strong AI hypothesis" is essential. ---- CharlesGillingham (talk) 18:38, 11 August 2013 (UTC)


I would like to see something in the article that explains or challenges searle's "similarly", because this presupposes that he (or whomever is the person in the room later replaced by a computer) has no curiosity about the patterns he is observing amongst the symbols either presented to him or arranged by him, nor does it explain how the person (again, or the machine) might begin to recognise patterns & begin to get creative with the rules. I'm pretty sure that there's some evidence somewhere of this being how the human mind actually functions; that starting with basic sets of rules (such as language, social behaviour & so forth), the mind begins to recognise patterns & then to extrapolate, interpolate, & that this is what we call personality. I'm also quite sure- again, I'd need to dig a little deeper into 'new scientist' & 'wired' back-issues!- that we already have computer systems that are similarly capable of what we would anthropomorphically call 'learning' or 'adaptation'. this all is slightly away from searle's original thought experiment, but I think his thought-experiment dodges the business of either the person or the computer being capable of adaptive behaviour, that this adaptive behaviour itself may be evidence of something beyond a simple mechanism in the room, & that somewhere out there, there is or has been exactly this challenge levelled at searle's experiment. I'm going to keep looking. duncanrmi (talk) 22:18, 17 April 2014 (UTC)

Need to say that some people think the Chinese room is not Turing complete

See the discussion above, at #Unable to learn. To recap: I am pretty confident that Searle would say that the argument only makes sense if the room is Turing complete. But we need to research this and nail it down, because there are replies in the literature that assume it is not. I think this belongs in a footnote to the section Computers vs. machines vs. brains. ---- CharlesGillingham (talk) 21:37, 10 February 2011 (UTC)

I found one that is unambiguous: Hanoch Ben-Yami (1993). A Note on the Chinese Room. Synthese 95 (2):169-72: "such a room is impossible: the man won't be able to respond correctly to questions like 'What is the time'?" Ben-Yami's critique explicitly assumes that the rule-book is fixed, i.e. there is no eraser, i.e. the room is not Turing complete. ---- CharlesGillingham (talk) 18:30, 11 December 2011 (UTC)
The Chinese Room can't possibly be Turing complete, because Turing completeness requires an unlimited tape. The Chinese Room would be a finite-state machine, not a Turing machine. Looie496 (talk) 18:56, 11 December 2011 (UTC)
Ah, yikes, yes of course that's true. Note however that no computer in the real world has an infinite amount of tape either, so Turing complete machines can not exist. The article sneaks past this point. We could add a few more clauses of the form "given enough memory and time" to the article where this point is a problem. Or we could dive in with a paragraph dealing with Infinitely Large machines vs. Arbitrarily Large machines. Is this a hair worth splitting? ---- CharlesGillingham (talk) 19:13, 11 December 2011 (UTC)

Looie: I think I'm starting to agree with you that the "Turing completeness" paragraph needs to be softened and tied closer to Searle's published comments. Ideally, I'd like to be able to quote Searle that the Chinese room "implements" a Turing machine or that the man in the room is "acting as" a Turing machine or something to that effect. Then we can make the point about Turing completeness in a footnote or as part of the discussion in "Redesigning the room", where it's actually relevant.

The only thing I have at this point is this: "Now when I first read [Fodor's criticism], I really didn't know whether to laugh or cry for Alan Turing, because, remember, this is Turing's definition that we're using here and what Fodor in effect is saying is that Turing doesn't know what a Turing machine is, and that's a very nervy thing to say."[1] This isn't exactly on point, and worse, it isn't even published; it's just a tape recording of something he said. ---- CharlesGillingham (talk) 22:29, 25 February 2012 (UTC)

Forgive me but this discussion seems to me to entirely miss the point i.e. that the Chinese Room is a thought experiment. The practical issues of a real experiment only impact on a thought experiment if they somehow change some significant element of it not if they relate to practical issues. If it is factually incorrect to suggest that the turing machine test fails then reference to the turing test should be eliminated as it is insignificant to the thought experiment itself irrespective of whether or not Searle included it in his description. A thought experiment, at every step, includes the words such as: "theoretically speaking". In the Chinese room it is irrelevant whether the person does or does not have an eraser as the point is simply that a human could, "theoretically speaking", execute the instructions that a computer would by following the same program. It would simply take the human being far longer to do (so long that the person may die before they have sufficient time to answer even one question, but again this is irelevant to the experiment). LookingGlass (talk) 13:29, 4 December 2012 (UTC)
I don't think I fully understand your note, LookingGlass, so I don't want to put words in your mouth, but you appear to be conflating the "Turing test" with "Turing machines". They are completely different things. It is unfortunate, I guess, that they are both relevant to this subject, because it is easy to make this mistake. There is no such thing as "Turing machine test" as far as I am aware. The point in this section is whether the man-room system is a Turing-complete system, ie equivalent to a Turing machine. This section has nothing to do with the Turing test, as far as I can tell. The eraser is necessary for the system to be Turing-complete. Whether the man has an eraser is hugely important in some sense at least—if the man did not have an eraser then I don't think any computationalist would dare assert that the system understands anything let alone the Chinese language. :) This also helps us understand how wild and out in the weeds the CRA claim is, since as far as I'm aware we have no reason or even hint to imagine that there is some kind of computer more powerful than a Turing-equivalent machine, yet that is what Searle claims the brain is. To the question at hand, while I'm not aware of Searle ever mentioning Turing machines by name, I had always interpreted the language that he does use—eg "formal symbol manipulation" etc—as referring to "what computers do" ie Turing machines. That is after all the context in which Searle is writing, isn't it? I agree it would be nice to have a source in the article to back this up if there is one, but is there really much question about this? I'm not so sure the Ben-Yami paper is really saying what we're saying it's saying; it doesn't even mention Turing machines, for example. ErikHaugen (talk | contribs) 17:46, 4 December 2012 (UTC)
My apologies ErikHaugen. Please read my remarks with the word "machine" deleted. As far as I can determine, Searle's experiment is unaffected by the details of the situation. The point at hand is that a computer program can, in theory, be executed by a human being. LookingGlass (talk) 12:11, 6 December 2012 (UTC)
The point is relevant to the "replies" which this article categorizes as "redesigning the room". There are many such criticisms. Searle argues that any argument which requires Searle to change the room is actually a refutation of strong AI, for exactly the reason you state: it should be obvious that anyone can execute any program by hand, even a program which Strong AI claims is "conscious" or "sentient" or whatever. If, for some reason, you can't execute the program by hand and get consciousness, then computation is not sufficient for consciousness, therefor strong AI is false.
"Turing completeness" is a way to say this that makes sense to people who are trained in computer science. ---- CharlesGillingham (talk) 06:35, 8 January 2013 (UTC)
Many thanks Charles. Searle's argument seems beautifully elegant to me. LookingGlass (talk) 20:56, 8 January 2013 (UTC)

"The Real Thing"

I removed a reference to "The Real Thing" in the "Strong AI vs. AI research" and replaced it with "human like cognition". The original phrase was intended to refer to cognition, but in the context of the sentence could be easily misconstrued to refer to intelligence. I felt the distinction was worthy of clarification because in Searles' conception, computers do in fact have "real" intelligence under a variety of understandings of the term but lack the capacity for awareness of the use of that intelligence. Jaydubya93 (talk) 13:44, 16 January 2014 (UTC)

I agree with you, however, I think "a simulation of human cognition" could be construed as "human-like cognition" (as opposed to, say, Rodney Brook's "bug-like cognition"). In this reading, Searle's argument presumes that machines with "human-like cognition" are possible, so the sentence as you changed it doesn't quite work either. I changed it to explicitly mention "mind" and "consciousness" (i.e. "awareness", as you said above) so that the issue is clearer. ---- CharlesGillingham (talk) 19:59, 18 January 2014 (UTC)
No. Searle's argument does not presumes "...that machines with *human-like cognition* are possible" ; it presumes that "that machines with "human-like cognition" WILL BE possible", if the scientific mentality changes. Luizpuodzius (talk) 19:33, 24 December 2014 (UTC)

Odd placement of citations and notes

Why does this article have citations at the beginning of paragraphs instead of the ends. Is this some WP innovation? Myrvin (talk) 06:46, 6 July 2015 (UTC)

The footnotes you're referring to are actually supposed to be attached to the bold title, but changes in formatting over the years have moved them down a line and to the front of the paragraph. I should move them after the first sentence. ---- CharlesGillingham (talk) 04:52, 10 September 2015 (UTC)
  Fixed ---- CharlesGillingham (talk) 05:05, 10 September 2015 (UTC)

Where I agree and disagree with Searle

<Sorry, I have removed the original contents of this section. This talk page should only be used to discuss ways of improving the Wikipedia article based on reputable published sources. It is not a forum for discussing the topic.> Looie496 (talk) 12:35, 22 September 2015 (UTC)

What am I missing?

<Sorry, I have removed the original contents of this section. This talk page should only be used to discuss ways of improving the Wikipedia article based on reputable published sources. It is not a forum for discussing the topic.> Luizpuodzius (talk) 23:23, 23 September 2015 (UTC)

Richard Yee source

(ref 51, 59, 60) Department of Computer Science, University of Massachusetts at Amherst. His paper is published in Lyceum, which seems to be a free online journal published by the Saint Anselm Philosophy Club. Can't find much on him, apart from a few papers presented at workshops ("Machine learning: Proceedings of the Eighth International Workshop (ML91)"), "Abstraction in Control Learning".

Is there any reason to believe this person and his opinion are notable? His argument seems to be superfluous, imo, it doesn't add anything. We know that the instructions (the "rules table") are the program. The room has only one set of instructions: for Chinese. There's no talk about the room doing other languages ("changing the program"). In all aspects it is a general turing machine doing one specific task. Yet Yee brings up external programs and universal turing machines, only to say they are a red herring and "philosophical discussions about "computers" should focus on general Turing computability without the distraction of universal programmability.". A distraction he himself introduced.

We're not supposed to interpret sources, but it makes me wonder whether his paper/opinion is really more notable than for example a random blog entry or forum post discussing the subject. The point he makes would already be obvious to the reader: the rules table, not the person, is the program. Yet this is once again repeated in the systems reply section with an example, concluding: Yet, now we know two things: (1) the computation of integer addition is actually occurring in the room and (2) the person is not the primary thing responsible for it. This was first added here. Is it really necessary to explain it at such length? Ssscienccce (talk) 02:43, 7 October 2015 (UTC)

I agree. It's the naive "system" reply, with a little formal window dressing.
I think he's kind of missed the point about the room being a Turing machine -- the point is, digital computers can never be "more sentient" than the room-with-Searle, they can only be faster. The Chinese room is as sentient as you ever get. ---- CharlesGillingham (talk) 05:26, 14 December 2015 (UTC)

Hello fellow Wikipedians,

I have just modified one external link on Chinese room. Please take a moment to review my edit. If you have any questions, or need the bot to ignore the links, or the page altogether, please visit this simple FaQ for additional information. I made the following changes:

When you have finished reviewing my changes, please set the checked parameter below to true or failed to let others know (documentation at {{Sourcecheck}}).

This message was posted before February 2018. After February 2018, "External links modified" talk page sections are no longer generated or monitored by InternetArchiveBot. No special action is required regarding these talk page notices, other than regular verification using the archive tool instructions below. Editors have permission to delete these "External links modified" talk page sections if they want to de-clutter talk pages, but see the RfC before doing mass systematic removals. This message is updated dynamically through the template {{source check}} (last update: 5 June 2024).

  • If you have discovered URLs which were erroneously considered dead by the bot, you can report them with this tool.
  • If you found an error with any archives or the URLs themselves, you can fix them with this tool.

Cheers.—cyberbot IITalk to my owner:Online 22:48, 26 May 2016 (UTC)

Hello fellow Wikipedians,

I have just modified one external link on Chinese room. Please take a moment to review my edit. If you have any questions, or need the bot to ignore the links, or the page altogether, please visit this simple FaQ for additional information. I made the following changes:

When you have finished reviewing my changes, please set the checked parameter below to true or failed to let others know (documentation at {{Sourcecheck}}).

This message was posted before February 2018. After February 2018, "External links modified" talk page sections are no longer generated or monitored by InternetArchiveBot. No special action is required regarding these talk page notices, other than regular verification using the archive tool instructions below. Editors have permission to delete these "External links modified" talk page sections if they want to de-clutter talk pages, but see the RfC before doing mass systematic removals. This message is updated dynamically through the template {{source check}} (last update: 5 June 2024).

  • If you have discovered URLs which were erroneously considered dead by the bot, you can report them with this tool.
  • If you found an error with any archives or the URLs themselves, you can fix them with this tool.

Cheers.—InternetArchiveBot (Report bug) 15:01, 22 November 2016 (UTC)

Regarding the reply section

Why is the replies section not labeled as criticisms instead? I'm curious as it's not something I'm used to seeing on Wikipedia, as many of these "replies" are criticisms/arguments. Wiremash (talk) 11:06, 28 November 2016 (UTC)

Because Searle's famous paper in Behavioral and Brain Sciences referred to them as "replies". In this respect he was more or less following the structure of Turing's famous paper Computing Machinery and Intelligence, which proposed the Turing test -- although Turing used the term "objections". Looie496 (talk) 15:54, 28 November 2016 (UTC)

Interesting, seems like a clever choice of words to imply neutrality Wiremash (talk) 16:07, 29 November 2016 (UTC)

Need to say that some people think that Searle is saying there are limits to how intelligently computers can behave

Similarly, some people also lump Searle in with Dreyfus, Penrose and others who have said that there are limits to what AI can achieve. This also will require some research, because Searle is rarely crystal clear about this. This belongs in a footnote to the section Strong AI vs. AI research. ---- CharlesGillingham (talk) 21:37, 10 February 2011 (UTC)

He seems clear enough to me: he doesn't claim that there are limits on computer behavior, only that there are limits on what can be inferred from that behavior. Looie496 (talk) 23:24, 5 April 2011 (UTC)
Yes, I think so too, but I have a strong feeling that there are some people who have written entire papers that were motivated by the assumption that Searle was saying that AI would never succeed in creating "human level intelligence". I think these papers are misguided, as I take it you do. Nevertheless, I think they exist, so we might want to mention them. ---- CharlesGillingham (talk) 08:19, 6 April 2011 (UTC)
Is this the same as asking if computers can understand, or that there are limits to their understanding? What does it mean to limit intelligence, or intelligent behaviour? Myrvin (talk) 10:16, 6 April 2011 (UTC)
There is eg. this paraphrase of Searle: "Adding a few lines of code cannot give intelligence to an unintelliget system. Therefore, we cannot hope to program a computer to exhibit understanding." Arbib & Hesse, The construction of reality p. 29. Myrvin (talk) 13:19, 6 April 2011 (UTC)
I think that, even in this quote, Searle still holds that there is a distinction between "real" intelligence and "simulated" intelligence. He accepts that "simulated" intelligence is possible. So the article always needs to make a clear distinction between intelligent behavior (which Searle thinks is possible) and "real" intelligence and understanding (which he does not think is possible).
The article covers this interpretation. The source is Russell and Norvig, the leading AI textbook.
What the article doesn't have is a source that disagrees with this interpretation: i.e. a source that thinks that Searle is saying there are limits to how much simulated intelligent behavior that a machine can demonstrate. I don't have this source, but I'm pretty sure it exists somewhere. ---- CharlesGillingham (talk) 17:32, 6 April 2011 (UTC)
Oops! I responded thinking that the quote came from Searle. Sorry if that was confusing. Perhaps Arbib & Hesse are the source I was looking for. Do they believe that Searle is saying there are limits to how intelligent a machine can behave? ---- CharlesGillingham (talk) 07:34, 7 April 2011 (UTC)
See what you think CG. It's in Google books at: [2]. Myrvin (talk) 08:27, 7 April 2011 (UTC)
Reading that quote one more time, I think that A&H do disagree with the article. They say (Searle says) a computer can't "exhibit understanding". Russell and Norvig disagree (I think). They say (Searle says) even if a computer can "exhibit" understanding, this doesn't mean that it actually understands.
With this issue, it's really difficult to tell the difference between these two positions from out-of-context quotes. If the writer isn't fully cognizant of the issue, they will tend to write sentences that can be read either way. ---- CharlesGillingham (talk) 19:27, 12 April 2011 (UTC)

Hello fellow Wikipedians,

I have just modified one external link on Chinese room. Please take a moment to review my edit. If you have any questions, or need the bot to ignore the links, or the page altogether, please visit this simple FaQ for additional information. I made the following changes:

When you have finished reviewing my changes, you may follow the instructions on the template below to fix any issues with the URLs.

 Y An editor has reviewed this edit and fixed any errors that were found.

  • If you have discovered URLs which were erroneously considered dead by the bot, you can report them with this tool.
  • If you found an error with any archives or the URLs themselves, you can fix them with this tool.

Cheers.—InternetArchiveBot (Report bug) 19:51, 11 January 2018 (UTC)

neural computator

If I could produce a calculator made entirely out of human neurons (and maybe some light emitting cells to form a display), can I then finally prove humans are not intelligent :P ? Such a biological machine would clearly possess only procedural capabilities and have a formal syntactic program. That would finally explain why most people are A) not self-aware and B) not capable.

You people do realize that emergent behavior is not strictly dependent on the material composition of its components, but rather emerge from the complex network of interactions that emerge said components? Essentially the entire discussion is nothing more than a straw man. — Preceding unsigned comment added by 195.26.3.225 (talk) 13:23, 30 March 2016 (UTC)

Exactly!
Reducing to the absurd in another way, Searle's argument is like extending a neuron's lack of understanding to the whole brain. 213.149.61.141 (talk) 23:48, 27 January 2017 (UTC)
But, to respond to Searle, you have to explain exactly how this "emergent" mind "emerges". You rightly point out that there is no contradiction, but Searle's argument is not a reductio-ad-absurdum. The argument is a challenge: what aspect of the system creates a conscious "mind"? Searle says there isn't any. You can't reply with the circular argument that assumes "consciousness" can "emerge" from a system described by program on a piece of paper. ---- CharlesGillingham (talk) 21:58, 18 May 2018 (UTC)
We don't know what aspect of the system creates a conscious mind. This is true whether the system we are talking about is some hypothetical chinese room, or just a normal human brain. I'm not sure how Searle can claim confidently that there isn't anything in the system that is conscious. Sure it's unintuitive that a conscious system made of paper can arise, but individual neurons aren't any more conscious than slips of paper, and collections of neurons nevertheless manage to become a conscious system somehow. What's so special about them? Given that the paper and the neurons are processing the exact same information, why is it so implausible that consciousness can emerge from the paper in the exact same way as it can from the neurons?
This problem won't be satisfactorily resolved until we can finally answer the challenge, as you say, but the argument doesn't really get you anywhere. If you don't already believe neurons are special and are the only things that can possibly generate consciousness, the argument won't sway you in the least because the argument relies on that assumption.
It is ironic that Searle claims his detractors to be arguing circularly on the basis that their objections to the argument only work if we assume substrate-independant consciousness exists, when his argument also only works if we assume it does not. It's barely even an argument, more a convoluted statement of his own opinion. It's like saying something unsubstantiated, and then when someone asks you to back it up, claiming that they are arguing circularly because their asking you to back up your argument doesn't itself prove you wrong. — Preceding unsigned comment added by Mrperson59 (talkcontribs) 06:06, 14 January 2021 (UTC)

The Game (1961) is a philosophical argument

You deleted my text and wrote: "The Game is not an argument in the philosophy of mind, and this article is about Searle's argument, not one element of his thought experiment". This is incorrect, because The Game (1961) is "an argument in the philosophy of mind". Possibly you are not familiar with this style of writing of philosophical arguments --- it is called Socratic dialogue, originates from Greece and is popular in East Europe including Bulgaria and Russia. I am familiar with these facts because Bulgaria is neighboring country of Greece. Thus, The Game (1961) is the original argument in Russian that contains both the Chinese room and the China brain arguments. Obviously, the American philosophers have plagiarized The Game as they have not cited it. I do not know any of the two philosophers who are credited separately for the Chinese room (1980) and the China brain (1978) to have stated publicly that they know to read and understand Russian language. During the cold war between the Soviet Union and USA, ideas were stolen by both sides without referencing the other. This ended in the 1990s with the end of the Soviet Union. In the course of 20 years (from 1960s to 1980s) after the publication of Dneprov's work, his story The Game was split into two arguments, although in the original it was combined into a single experiment. I would advise you to go and read the whole story before making further edits. In case you do not understand Russian, here is the full tri-lingual version (Russian-English-Bulgarian) prepared by me (I understand perfectly Russian and have verified every word in all 3 languages) of Dneprov A. The game. Knowledge—Power 1961; №5: 39-41. PDF. A direct quote from the article shows that it is a philosophical argument: "I think our game gave us the right answer to the question `Can machines think?' We've proven that even the most perfect simulation of machine thinking is not the thinking process itself." The original scan of the Russian journal is added as appendix to my translation. The Game is published in the May 1961 issue, just when the Soviet Union beat the USA to send the first human in the outer space -- that is why Yuri Gagarin is on the journal cover. p.s. Please include back the material that you deleted. I would not object if you edit it to suit your own vision of neutrality. Danko Georgiev (talk) 17:46, 19 August 2021 (UTC)

Nevertheless, this is an article about Searle's argument, not about all arguments of this form. ---- CharlesGillingham (talk) 21:07, 21 August 2021 (UTC)
I added Dneprov's name to the first paragraph (along with the other precursors mentioned in the history section). Is this acceptable? You may have noticed that, in my last edit, I made sure the article no longer gave any precedence to Searle for this kind of critique. It just says that he is the author of this specific critique.
But I still don't agree that your edits were good idea. Since my previous answer was a bit glib, so I thought I would flesh it out.
My goal here is strictly editorial. Every article has a topic, and it's an editor's job to keep all the material in the article on topic. The title of this article is "Chinese Room", which is a specific thought experiment, not a general approach to these problems. "The Game" and the "Chinese Room" are two different critiques of AI. This article is about only one of them.
Everything in this article is specific to the "Chinese Room" version of this idea. Nothing in this article is about "the Game", or other similar arguments or thought experiments (except the two paragraphs in "History"). The Replies are all from people who replied to Searle's argument, not "The Game". The clarifications in the Philosophy and Computer Science were made by people trying to clarify Searle's version of the idea, not "The Game". Searle's Complete Argument is made by Searle, not by Dnepov. There is hardly a sentence in this article that isn't reporting some bit of academic dialog about Searle's version. None of this academic dialog was about Dneprov's version.
Perhaps American academics should have paid more attention to Dnepov's version. Like you, I have no idea if any of the participants in this 40 year long argument even knew the Dneprov version existed. Perhaps Dneprov's version deserved a bigger footprint in the philosophy of mind, or the history of philosophy. Perhaps one day it will receive that kind of attention. All that might be true. But:
We are merely editors. We report, we do not correct. We can't rewrite academic history from Wikipedia. This article summarizes the 10,000 pages of academic argument that his been published in reaction to Searle's version. That's what the article is. Thanks to you, now it mentions Dnepov as one of the people who had the same idea, quite a bit earlier. But Dneprov's version is not the topic of the article. ---- CharlesGillingham (talk) 21:41, 21 August 2021 (UTC)
Finally, would you object to moving this conversation to the talk page of Chinese Room? I think it belongs there. ---- CharlesGillingham (talk) 21:41, 21 August 2021 (UTC)
Dear Charles, thank you very much for the explanations of your thoughts on the editing process. I will not mind if you move the conversation to another talk page if you want. Because I work daily on scientific projects, my thought process is focused mainly on veracity of ideas or arguments. Consequently, I can give you reasons for or against some proposition, but I will leave it up to you to decide what to do. In regard to your general attitude to separate Dneprov's argument and Searle's argument, it is factually incorrect because the Chinese language is not essential to characterize Searle's argument! Consider this: The Chinese room makes no sense to 1.5 billion Chinese citizens who are native speakers of Chinese! All these Chinese people will perform the manipulations and they will understand Chinese, thereby invalidating Searle's conclusion! Only from this global viewpoint that Searle's argument is meaningless to 1.5 billion native Chinese (i.e. 1/3 of all people on Earth!), you can understand the importance of my original edit in Wikipedia where I stated that Searle "proposed Americanized version" of the argument. "Americanized version" means that it makes sense to Americans, but may not make sense to people living in other parts of the world. In particular, if Chinese philosophers want to teach Searle's Chinese room in their philosophy textbooks, the only way to achieve the goal intended by the author is to change the language to some language that they do not understand like the "American room". Now, after I have demonstrated to you that the word "Chinese" in the "Chinese" room is not defining the argument, it is clear that Dneprov's argument is exactly the same argument written for Russians and using Portuguese language that Russians do not understand. What matters from historical viewpoint and in terms of attribution is that it precedes by ~20 years Searle's publication. A philosophical argument is defined by the general idea of proof and the final conclusion. Dneprov uses (1) people to simulate the working of an existing 1961 Soviet computing machine named "Ural", which translates a sentence from some language A to another language B. (2) The people do not understand language A, neither before, nor after they perform the translation algorithm. (3) Therefore, the executing of the translation algorithm does not provide understanding. Final conclusion intended by Dneprov using the technique of Socratic dialogue (i.e. the words are spoken by the main story character "Prof. Zarubin") is that machines cannot think. Searle's argument is identical except for the number of people involved in the translation (1 or many it does not matter) and the specific choices of languages A and B. My general attitude to contribute to Wikipedia is that the English version is for all mankind, and not only for Americans. Therefore, articles should be written from a country neutral perspective and sensitivity to the fact that most Wikipedia users are non-native English speakers across the globe. p.s. If you find a version of Searle's argument written by someone before 1961, then definitely send me a copy of the original text and I will advocate for the attribution of "Chinese room" to the first person who clearly formulated steps (1), (2) and (3) in the argument. Danko Georgiev (talk) 05:57, 22 August 2021 (UTC)
I still think you're misunderstanding our role here.
This is an encyclopedia article about a historically notable conversation in philosophy. The historical conversation has already happened. We can only report what the conversation was. We can't report what the conversation should have been.
We don't write what we, ourselves, think is true. We report what notable philosophers said. We don't try to put words into their mouth, or rebuild the subject for them. We try to explain what they said, who they said it to, and what the context was, without putting our own spin on it.
So it actually doesn't matter if you're right about this -- if Dnepov was making exactly the same argument. This article can't give the impression that the entire field of philosophy says the same things you do, even if you're right. Again, none of the other people cited in this article think they were writing about Dneprov. Harnad, Dennett, Dreyfus, Chalmers, McGinn, all of them --- they weren't writing about Dneprov. That might not even have heard of Dneprov. They were writing about Searle.
That's what we have to report -- what they thought. We can't take into account what we think. What we think has no place in Wikipedia. So you don't need to keep arguing for this -- no matter how good you're argument is, it doesn't effect the 'editorial' issue.---- CharlesGillingham (talk) 06:49, 23 August 2021 (UTC)
Dear Charles, I think you have a problem with understanding the meaning of jargon (technical words) used in philosophy, which means that you should probably refrain from editing articles on philosophical subjects. The sense in which the term "argument" is used in philosophy is the same as proof, not as conversation. The intended meaning (semantics) of the expression "Chinese room argument" is the same as "Chinese room proof" and NOT as "Chinese room conversation"! P.S. I have just seen the new revision of the article and do agree with it. A sentence in the introduction on the history of the proof was all that was needed to make things right. Danko Georgiev (talk) 15:45, 23 August 2021 (UTC)
To other wiki editors: The main issue that I had with the article is the removal of the text on the first documented discoverer Anatoly Dneprov of the Chinese room argument (proof/theorem). I do not object that Searle has come up (possibly independently) with the same proof, and that he is widely mis-credited as the originator of the argument. However, now we have a documentary record of Dneprov's 1961 work and high quality scan of the Soviet journal. So, this historic fact should be mentioned in the wiki article. For the type of wiki edit that I am proposing, I give a concrete example with the Pythagorean theorem, which bears the name of Pythagoras. For a long time, it was taught in Europe that Pythagoras discovered the general theorem, while Babylonians only knew of some special cases called Pythagorean triples (like 3,4,5). However, we now have historic evidence that Indians in the 8th to 6th centuries BCE already knew the Pythagorean theorem as it is recorded in Baudhāyana Sulbasūtra. This information on earliest proof is included in the wiki article on the Pythagorean theorem. The Chinese room argument is a proof or theorem providing negative answer to the question whether machines can think. Dneprov's proof is published in 1961, which is almost two decades earlier than Searle's proof in 1980. Therefore, Dneprov's work has to be mentioned in the article with respect to historic attribution. Dneprov was a Soviet physicist and working on cybernetics. His publication was intended to prove that machines cannot think to anyone interested in science. Danko Georgiev (talk) 16:19, 23 August 2021 (UTC)

Just to put this to rest -- Dneprov is now credited (in the history section) with having come up with the exact same argument 20 years earlier, but the article as a whole is still about Searle's version and the replies to it. ---- CharlesGillingham (talk) 09:27, 11 September 2021 (UTC)

Brain replacement scenario

I don't particularly see how this (or any other reply involving an artificial brain) actually is a reply to the Chinese Room, a thought experiment involving a set of instructions and a processor. More specifically though, the quote from Searle in this section is certainly wildly out of context; it's framed here as Searle's prediction of why brain replacement could not result in a machine consciousness, but in Rediscovery of the Mind, the source of the quote, Searle brings up brain replacement to illustrate the difficulty in inferring subjective mental states from observable behaviour, and this description of brain replacement resulting in gradually shrinking consciousness is alongside two alternative scenarios in which consciousness is entirely preserved (one in which behaviour is also preserved, and one in which it is not). It's not a prediction, or a reply to a reply to the Chinese Room argument — except perhaps thematically. LetsEditConstructively (talk) 14:14, 8 November 2021 (UTC)

I think we should cut the Searle quote. Or we could remove the word "predicts" and give a more accurate context of the quote, if you're up for it.
However, I think we should leave the scenario in, as one more item in the list, even if Searle himself wouldn't call it a "reply". Three reasons:
  • Russell and Norvig (the leading AI textbook) brings it up while discussing the Chinese Room, so we are justified in bringing it up here.
  • The scenario is on topic. You said that Searle brings it up while discussing "The difficulty in inferring subjective mental states from observable behaviour". This is a key issue in the Chinese Room argument. (Some would say it's the only issue. Colin McGinn, for one.)
  • The scenario is a legitimate problem for the CRA, at least as much as the other replies in this section. We know that: (1) Brains have subjective mental states. (2) A brain replaced by tiny, parallel digital machinery is equivalent to a digital computer running a program. (3) Searle thinks that the Chinese Room shows that it is impossible for a digital computer running a program to have subjective mental states. So, if Searle is right, then (1+2+3) consciousness must be present at the start of the procedure and absent at the end. It's hard to say exactly how and when it disappears, and it's easy to imagine that it just doesn't disappear and that Searle is wrong. It's an argument from intuition, like most of the other replies.
That's my two cents. Let me know if you agree. ---- CharlesTGillingham (talk) 07:47, 12 July 2022 (UTC)
I have incorporated your thoughts into the article ---- CharlesTGillingham (talk) 01:41, 2 August 2022 (UTC)
I probably should have left the first sentence off my initial post, because my main concern was how the quote from Searle was incorporated into the section. I think the new text/footnotes are much better, thank you!
It isn't relevant to the article but I have to ask: do we know (2)? They might be computationally equivalent, in that the same operations can be performed on either one (or on a very large piece of paper, for that matter), but they certainly aren't physically equivalent. Treating them as the same thing seems like begging the CRA's question. LetsEditConstructively (talk) 12:51, 1 February 2023 (UTC)
I suppose I mean that they aren't equivalent structurally, rather than physically. An artificial brain's ability to converse in Chinese is presumably intrinsic to the arrangement of its neurons rather than depending on the program its neurons execute, while a program that converses in Chinese (even one that functions by simulating neurons) should be able to do so regardless of the specifics of the hardware it's run on. Anyway this is getting fairly forumish so I'll drop it. LetsEditConstructively (talk) 23:00, 1 February 2023 (UTC)

Spelling nitpicking: minuscule "r" for room? Or uppercase R?

I've noticed that for "Chinese Room argument" there are many representaions in this article, spelling-wise. "Chinese Room", "Chinese room", "Chinese Room Argument", "Chinese room argument" and "Chinese Room argument". Now some of these mentions are directly quoted from research papers and might be represented in their correct spelling, whereas the article in general uses "Chinese room" or "Chinese room argument" with a minuscule r for room. Not wanting to stir things up I left things as they are, however I do wonder if you think the usage is consistent as-is (I think maybe it is not), or if anyone can decide because they read all the relevant referenced articles that they indeed are (not me, its been too long since I read Searle). Any pointers? Or too much meta already? Vintagesound (talk) 20:08, 25 May 2021 (UTC)

I would support either convention. "Chinese Room Argument" is the name of the argument. The "Chinese room" is something that appears in the thought experiment. But I imagine that there isn't a super reliable source. What does the Stanford Encyclopedia of Philosophy do? ---- CharlesGillingham (talk) 08:07, 16 August 2021 (UTC)
Good question, but the SEP article is a mess. It starts of using 'Chinese Room Argument' but later slips into 'Chinese Room argument'; it mostly refers to the room as 'the room' but sometimes refers to it as 'the Chinese Room'.
So I think Wikipedia is on its own, and your first instinct seems best -- 'Chinese Room Argument', 'Chinese room', 'room'. Ncsaint (talk) 12:25, 18 February 2023 (UTC)
  1. ^ http://plato.stanford.edu/entries/chinese-room/
  2. ^ a b Searle 1980.