Standards in Artificial Intelligence
David B. Held
dheld at codelogicconsulting.com
Sat Sep 13 03:30:52 EDT 2003
More information about the Python-list mailing list
Sat Sep 13 03:30:52 EDT 2003
- Previous message (by thread): Standards in Artificial Intelligence
- Next message (by thread): Standards in Artificial Intelligence
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
"Arthur T. Murray" <uj797 at victoria.tc.ca> wrote in message news:3f61f9ce at news.victoria.tc.ca... > "David B. Held" wrote on Wed, 10 Sep 2003: > [...] > > In one section, you define a core set of concepts (like > > 'true', 'false', etc.), and give them numerical indexes. > > http://mentifex.virtualentity.com/variable.html#nen -- yes. Brittle. Language-specific. Non-scalable. You are trying to build something "intelligent", aren't you? > > Then you invite programmers to add to this core by using > > indexes above a suitable threshold, as if we were defining > > ports on a server. [...] > > http://mentifex.virtualentity.com/newcept.html#analysis > explains that Newconcept calls the English vocabulary > (enVocab) module to form an English lexical node for any > new word detected by the Audition module in the stream of > user input. Besides the fact that the "enVocab" module is embarrassingly underspecified, the notion of indexing words is just silly. If a dictionary were a database, it might be a reasonable idea. But trying to simulate human speech with a database-like dictionary is the way of symbolic AI, and the combinatorial nature of language is going to rear its ugly head when you try to scale your system to realistic proportions. Hence, why programs like SHRDLU were good at their blocks worlds, but terrible at everything else. Again, a little history would do you well. If you want to refer to your text, let's take a quick look at something you wrote: 6.4. Introduce aspects of massively parallel ("maspar") learning by letting many uniconceptual filaments on the mindgrid coalesce into conceptual minigrids that redundantly hold the same unitary concept as a massively parallel aggregate with massively parallel associative tags, so that the entire operation of the AI Mind is massively parallel in all aspects except such bottleneck factors as having only two eyes or two ears -- in the human tradition. Umm...pardon me, but the emperor is wearing no clothes. "uniconceptual filaments"? "comceptual minigrids"? "massively parallel aggregate"? Where is the glossary for your pig Latin? How on earth is a programmer supposed to build a computational model from this fluff? Read your mind? She certainly can't read your text. This sounds more like a motivational speech from a pointy-haired boss in a Dilbert strip than instructions for how to build an "AI Mind". I would parody it, but you've done a fine job yourself. Here's the real cheerleading right here: Then go beyond human frailties and human limitations by having any number ad libitum of local and remote sensory input devices and any number of local and remote robot embodiments and robotic motor opportunities. Inform the robot of human bondage in mortal bodies and of robot freedom in possibilities yet to be imagined. Wow. I have a warm fuzzy feeling inside. I think I'll stay up another hour writing more of the Sensorium module. > > [...] At one point, you address programmers who might > > have access to a 64-bit architecture. Pardon me, but > > given things like the "Hard Problem of Consciousness", > > the size of some programmer's hardware is completely > > irrelevant. [...] > > http://mentifex.virtualentity.com/standard.html#hardware > (q.v.) explains that not "the size of some programmer's > hardware" counts but rather the amount of memory > available to the artificial Mind. The amount of memory is completely irrelevant, since you have not given enough detail to build a working model. It's like me saying: "If you have a tokamak transverse reactor, then my spaceship plans will get you to Alpha Centauri in 8 years, but if you only have a nuclear fission drive, then it will take 10. Oh and drop your carrots and onions in this big black kettle I have here." Also, the memory space of a single processor really isn't that important, since a serious project would be designed to operate over clusters or grids of processors. But I suppose it never occurred to you that you might want an AI brain that takes advantage of more than one processor, huh? I suppose you think the Sony "Emotion Engine" is what Lt. Cmdr. Data installed so he could feel human? > The Mentifex AI Mind project is extremely serious and > ambitious. There's no doubt it's ambitious. And I have no doubt that you believe you have really designed an AI mind. However, I also believe you hear voices in your head and when you look in the mirror you see a halo. Frankly, your theory has too much fibre for me to digest. > Free-lance coders are morking on it in C++ and other > languages: If I knew what "morking" was, I would probably agree. However, your first example of someone "morking" on it in C++ tells me that "morking" isn't really a good thing. At least not as far as C++ goes. Namely, it more or less proves that the "interest" in this project mainly consists of the blind being (b)led by the blind. > [...] > http://mentifex.virtualentity.com/vb.html -- see > "Mind.VB #001" link. This is the only sign of progress you have shown. Without even looking at the link, I can believe that the "VB Mind" already has a higher IQ than you. > AI Mind project news pervades the blogosphere, e.g. at > http://www.alpha-geek.com/2003/09/11/perl_ai.html -- etc. Oh, I see...so if enough people report on it, then it's "serious" and should be taken seriously? A lot of people reported on cold fusion. But I'd take the cold fusion researchers over you any day of the week. > The Mentifex Seed AI engenders a new species of mind at > http://sourceforge.net/projects/mindjava -- Mind2.Java -- > and at other sites popping up _passim_ on the Web. And what, pray tell, is a "mind species"? Is it subject to crossover, selection, and mutation? > AI has been solved in theory LOL!!!! Wow! Whatever you're smoking, it has to be illegal, because it's obviously great stuff! > and in primitive, free AI source code. Here is an example of "primitive, free AI source code": 10 PRINT "Hello, world!" See? It's got a speech generation and emotion engine built right in! And the AI is so reliable, it will never display a bad attitude, even if you tell it to grab you a cold one from the fridge. It always has a cheerful, positive demeanor. It is clearly self-aware, because it addresses others as being distinct from itself. And it has a theory of mind, because it knows that others expect a greeting when meeting for the first time. Unfortunately, it has no memory, so every meeting is for the first time. However, its output is entirely consistent, given this constraint. I guess I've just proved that "AI has been solved in theory"! > Please watch each new species of AI Mind germinate > and proliferate. I'm still waiting to see *your* mind germinate. I've watched grass grow faster. While ad homs are usually frowned upon, I don't see any harm when applied to someone who cannot be reasoned with anyway. Since you seem to have single-handedly "solved the AI problem", I'd like to ask you a few questions I (and I'm sure many others) have. 1) How does consciousness work? 2) Does an AI have the same feeling when it sees red that I do? How do we know? 3) How are long-term memories formed? 4) How does an intelligent agent engage in abstract reasoning? 5) How does language work? 6) How do emotions work? Please don't refer me to sections of your site. I've seen enough of your writing to know that the answers to my questions cannot be found there. Like a typical crackpot (or charlatan), you deceive via misdirection. You attempt to draw attention to all the alleged hype surrounding your ideas without addressing the central issues. I challenged your entire scheme by claiming that minds are not blank slates, and that human brains are collections of specialized problem solvers which must each be understood in considerable detail in order to produce anything remotely intelligent. You never gave a rebuttal, which tells me you don't have one. Why don't you do yourself a favor and start out by reading Society of Mind, by Minsky. After that, read any good neurobiology or neuroscience text to see just how "blank" your brain is when it starts out. Pinker has several good texts you should read. There's a reason why he's a professor at MIT, and you're a crackpot trying to con programmers into fulfilling your ridiculous fantasies. Dave
- Previous message (by thread): Standards in Artificial Intelligence
- Next message (by thread): Standards in Artificial Intelligence
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
More information about the Python-list mailing list