Reading carefully the basic criticisms were:
"The Singularity is kind of the Mayan Calendar for people who aren't credulous scientific illiterates - it's an apocalyptic scenario that gets them feeling all tingly."
"Yeah, I know what it is. I just don't have "faith" it will come to fruition."
Aside from being substance-free, they are somewhat abusive. The misuse of the word "faith" is also disingenouous in my opinion.
RAMA was aware of the novelty of the concept and patiently explained. The follow-up criticisms were:
"Of course the Singularity is a '
possibility.' " Scare quotes are an argument? Actually, conceding a singularity is a possibility automatically renders all comparisons with the Rapture null and void. The guy that coined the phrase comparing them in practice rather tends to assume much of the Singularity argument.
"So, they're placing
faith on the fact that their extrapolation is correct?"
Equivocating on the meanings of "faith" is still disingenuous.
"The singularity is about as certain as the rapture. In other words, it's not. There is no guarantee that technology will advance at an accelerated rate especially in all areas. You need to temper your optimism with some realism."
The Rapture is impossible. The Singularity is not. The comparison is a false one.
"Ah, but what I
asked was proof for the statement: 'The real fantasy here is linear thinking in technology which is demonstratably false' - that is, a disproof of alternatives to accelerationism, one that makes it clear that these alternatives are as credible as geocentrism or Flat Earthers. It is not enough for John Smart to have a credible narrative; his must be the
only credible narrative."
Moore's Law is exponential, which proves that linear thing in technological futurism is demonstrably false. A general or global accelerationism in technology is also going to be exponential because the interlinks between the various technological advancements will grow exponentially. Whether it is going to be a logistic curve, or whether the exponent will be 0.000 1 is another issue of course.
"Science is not a religion."
This is a criticism of the Singularity? Rather, it is a criticism of the critics who keep equating the Singularity with superstition.
"Until then it's as likely as the Rapture, which you can also come talk to me about when it's history and fact."
The Rapture is impossible, a stupid and crazy idea, while the Singularity, despite my personal disagreements, is not.
"Humans are "unique" in the sense that we don't know of any species like ourselves, but not unique in the sense something like us couldn't exist elsewhere or be created artificially."
The Singularity proponents quite agree. Why then the accusations of being superstitious nitwits?
"It takes exponentially more digital computing power to simulate an analog brain than a genuine analog brain requires."
And Moore's Law (and other such "laws" about the advancement of computing technology) are exponential. Yes, there is every reason to think the brain can be simulated within a relatively short period.
"Come on, those kinds of arguments [against the Singularity] are so silly as to not even need rebutting."
Since this is so, why the insistence on abusing the proponents as religious?
"Where's the exponential growth in AI? I rest my case."
Exponential growth is very slow at first. The relatively belated appearance of extraordinarily rapid growth in a brief period is precisely the disjunction that prompts the name "Singularity." The only real basic criticism of the AI aspect of the Singularity lies in demonstrating some real reason why exponential increases in computing power cannot by brute calculation produce AI. That's the problem for the critics here, they can't make such an argument. They argue only from personal incredulity.
"Kurzweil in particular often seems driven by father issues rather than imagination, much less science."
Genetic fallacy, ad hominem and just generally substituting bile for logic.
"...since
human beings have shown no evidence of being able to write the kinds of software that would make strong A.I. possible, it's necessary for evangelists to posit a magical moment at which computers will somehow begin to write their own software and create their own successors.
There is no reason based in evidence to expect this to happen soon, if ever.
So, what is this assumption based on? Faith. Wishful thinking. Nothing more, no observations drawn from history or the real world.
Attempting to use the applicability of Moore's law to computer hardware as a starting assumption and basis for extrapolating a similar exponential growth and evolution in processes of
a different sort and order exposes the essential laziness in the thinking..."
As you can see, most of this is repetitive, a lot of it misusing the word "faith" yet again. I've italicized the heart of the criticism, which is first, that we haven't achieved strong AI as yet. Anything besides linear extrapolation does not serve as a reason, etc. In other words, all the abuse is is justified by assuming the argument. It also says we can't extrapolate from historically attested exponential advances because AI is "a different sort and order..." No one has disagreed that we don't have "strong" AI, whatever that may mean. This is from the same poster who pointed out the disagreement among AI researchers about what they're trying to achieve! The argument here contradicts that: If AI doesn't necessarily mean simulating human awareness, then being a "different sort and order" is entirely irrelevant. Not only is the argument pointless, it is unsupported here.
"As others have pointed out, it's worth considering that biological evolution (the touchstone model here) has not, despite a head start of billions of years, stumbled onto the "algorithms" which would support exponentially accelerating change of this kind (despite the self-evident utility of such for the adaptation and survival of living forms)."
The problem with the argument from personal incredulity is what you don't know. The immune system is a living example of the type of a self-changing algorithm. It is not at all clear that intelligence is useful for adaptation and survival, despite the claim it is "self-evident." After all, humanity has only managed to survive about 200 000 yrs, and the average for species is 1 000 000 yrs. Get back to me later on this one. As for exponential growth in general, it leads to boom-and-bust cycles or turns out to be a logistic growth. The first is not self-evidently useful. The denial of exponential growth also denies the logistic curve, even though we find that in nature.
"You know what's fueled the past couple hundred years of human advancement? Fossil fuels."
Sounds like the beginning of an argument against the Singularity as a guaranteed Heaven on Earth.
"The assumption that simulating a human brain will result in generalized AI is also totally faulty, mainly because we so poorly understand how the interactions of neurons and chemicals results in the properties we see as intelligence and consciousness. How are you going to simulate something you don't even understand?"
Actually it is possible to simulate something you don't understand. Since exponential growth in computing power is historically established, then brute force simulation of the brain may produce some sort of AI, despite the looney assertions there is no reason to think so. What you need is an actual argument is not just that no exponential growth in computing power could possibly simulate the brain, but an argument that simulating the brain
cannot produce AI.
There hasn't been any offered except the mystical insistence that human-style awareness is essential to AI and that human-style awareness is uncopiable, period. This isn't an argument.
"What they're really positing is that we don't have to understand how to do [programming for AI,] because "understanding" is just about to blossom forth from the machines themselves."
They're positing that brute force simulation will produce some sort of AI. It is insisted by "emulation" isn't the same thing, but that is more mysticism than an argument. The exponential growth in computing power establishes the reasonableness of the argument. The notion that AI must be a human-style awareness is a tacit assumption of this criticism. It is still an unsupported assumption. Also, it does not address the Singularity advocates who are not convinced that human-style awareness is necessary. For those, the progress already made constitutes steps toward AI as machine, not human, intelligence.
There is no magic moment when "intelligence" emerges, not in biology nor in computers.
"The problem with "downloading minds" isn't one of "taking umbrage" but that it's the equivalent to "bottling unobtainium" - you've got a verb acting there on a noun that stands in for something with poorly or undefined characteristics and which there's little reason to think actually exists as an entity."
This is so confused it's hard to believe it was offered as a criticism. If the "mind" is not a metaphysical entity (I entirely agree)
then it is in principle simulable. In other words, this "criticism" inadvertently concedes the point. Also, this poster, somehow, knows this "poorly or undefined" something that might not even exist is still, somehow, identifiably a "different sort or order..." from machine computation. This nonsense has nothing to do with why someone should care about a simulation of his or her mind.
"What emerges in human beings is mainly the ego construct, that internal model which as you say is self-aware: it imagines itself as an entity moving within the context of environment and holds to the delusion that it more or less orders and controls events.
A lot of AI research used to concentrate on replicating the ego - it's what would really be examined by the Turing test. Is self-awareness actually intelligence?"
Precisely why the Singularity can happen without duplicating human-style awareness, not a criticism.
"We could have a long discussion about the 'ego' aspect--what some call a 'user illusion.' It is the aspect of the human mind we understand least, but I think the one most essential to duplicate in order to have a machine expand beyond its own programming, purely of its own volition."
Why? Did our minds expand beyond our ancestors purely of our own volition? The mysticism is getting very deep.
"Creativity may be one aspect of human intelligence we can't emulate simply because it can't be quantified. You could emulate intuition with a random number generator but then it wouldn't be intelligence."
Why not? Does it make sense to say that an emulation isn't what it emulates, even though it emulates it?
"Then there is also the fact that if you don't understand something, you certainly can't improve it and that's the whole point of singularity."
Refuted by trial and error. And with exponential growth in computation, trial and error is
huge.
"Adaptability. [and human-style consciousness] Without these two aspects, a machine intelligence would eventually face a problem it can't solve or a question it can't answer."
Human intelligence faces problems it can't solve and questions it can't answers. This seems to insist that the Singularity must produce a man-made God or it isn't the Singularity. In debating terms, I think this is called "moving the goalposts."
Reading carefully, it's obvious that the basic criticisms are confused and mystical and at best no more than "Gaw!"
[FONT=Arial]
[FONT=Arial][/FONT][/FONT]