It looks more like chicken and the egg problem. There should be some kind of internal tension to assert that culture restrains some kinds of research methodologies: it might be that what we see in the field is exactly what field wants to produce.
Mathematics tends to be detached from reality, and CS is not much different in its ability to create its own problems. You don't need computers to create models which you'll then extensively study, but surely CS opens more possibilities for such activity. Mathematics thrives on formal rules, and computers enable much more rigorous checking and discovery of much richer structures acting according to precisely defined rules; you can't fool a machine, and machine can cheaply perform all tedious bookkeeping. Rigor required for papers differs significantly for different applications, depending on its applicability. There's a pretty continuous spectrum from theoretical to applied research. As times goes on, some of the applied stuff gets converted to theoretical form. So, I think that problem is in absence of community that would produce research with the flavor we need to make progress in AGI. It's not suppressed, it just doesn't exist. Getting it started should be concerned with actions which can actually be performed, not with specific properties of its non-existent state. On 9/29/07, Robert Wensman <[EMAIL PROTECTED]> wrote: > > About why it is difficult to create complex system that balances several > functions, as opposed to a system with just one single function (assuming > that the former would be important for AGI): To find a reason for this > difficulty, I would like to point in a different direction, namely at the > scientific method itself as currently applied. > > It is my understanding, contemporary computer science works along a certain > pattern, that I believe is inspired much by mathematical research. A > researcher within computer science is supposed to write papers that follow a > pattern such as: > > "Present a formally well specified hypothesis, prove formally that the > hypothesis is correct." > > At this point, many would probably look at me in a strange way and say "so > what? that is how science is supposed to work! And for good reasons!". But > if we examine this pattern more carefully, we can see that it causes some > problems when it comes to a subject such as AGI. > > First of all, if we assume general intelligence to be a vague concept, it is > difficult to base any well specified hypothesis about how it is supposed to > work. There are many reasons why we should assume general intelligence can > never be anything more than a vague concept, one is complexity. If a concept > is complex enough, the description of such a concept becomes too lengthy for > it to be consistently reproduced in scientific communication. > > Secondly, if we cannot present a well formed hypothesis, making a formal > proof of its correctness is futile. Even experiments would be of little use, > as there is no well defined hypothesis to make comparisons against. How > about conclusions such as: > > "I designed my system in this specified way, and now after some > experimenting with it, we judge it to be a little bit smarter than our > previous design" > > There is no well specified hypothesis, there is no formal proof that the > hypothesis is correct. And because such research doesn´t fit the standard > pattern of good research, I guess scientists are forced to follow narrower > paths where provability is more easy. Like: > > "Using adaption algorithm A is proven by experiments to be 30% faster than > adaptation algorithm B when trying to learn a function F" > > I guess this boils down to a age old conflict of computer science, namely > whether computer science is about finding eternal facts, like in > mathematics, or whether computer science is about engineering and finding > methods that seems to work better, whether it can be proven formally or not. > > I believe one reason computer science is good at digging holes, is because > computer scientists looks with envy at, and wants to be like mathematicians. > Digging out eternal truths, so true that they would even be true in any > logically possible world, seems like a very noble prospect. Since computer > scientists have the freedom to invent the systems they study, it is tempting > to think that a lot of things they invent has the single purpose of being > something that is easy to state truths about, rather than something that is > useful or meaningful. > > I have seen something that I think is an example of this. I don't know how > many hundreds papers have been written about bottom-up parsing for compiling > of programming languages. Judge by my surprise when I started working at my > current company, and I found out that they used nothing else than recursive > descent for their massive system. Guess what? It worked lightning fast, and > in fact, the recursive descent was necessary for some syntax extension > features their language had. Their language had no such things as "reserved > words". So from my experience I see no use at all for all these papers about > bottom up parsing, and it makes me think, why were they written? Because the > world really needed bottom up parsing in the last decades, or because it was > easy to make theories on the subject? Maybe there will be some protests on > this specific example, but isn't it likely that a lot of systems designed by > computer science are designed so to make good theories? > > Another example is how the scientific community has moved away from > programming language design. Designing a programming language is similar to > building an AGI in this abstract sense, namely it is about making a balanced > combination of different functionality to try to obtain some vaguely defined > improvement, such as readability and structural beauty. It will not make > good research, in the current understanding of the word. > > This is also why I put little faith in the current computer science > community when it comes to vague research such as AGI. Surely enough, some > professors who has reached a certain status might have enough freedom to do > what they want. But PhD students who enter the community will feel a lot of > pressure to produce "solid proven facts" rather than "interesting > speculations", or "things that seem to work better". This forces the > community as a whole down the road towards narrow AI. > > Maybe there is hope if computer scientists tries to be a little bit less > like mathematicians, and dares to let in a little bit of the psychological > vagueness in their paper writing jargon. By that I do not mean to encourage > any kind of Freud-like incoherent crackpot theories, but just the kind of > vagueness that is associated with any kind of complex engineering, like > "this system seems to be better than that system", or "it seems this design > could benefit a certain capability" etc. Maybe an increased focus on AGI > would encourage such a development. > > /Robert Wensman > > > > > These are not clearly separable things. One of the reasons many > > people do the system synthesis and balanced approximations so badly > > is because they tend to use minor variations of the same function > > representations they would use when playing with those functions in > > isolation. The assumption that a particular set of functions are > > only expressible as a particular narrow form can frequently make it > > impossible to synthesize a useful system because the selected form > > imposes limits and tradeoffs specific to its form in practice that > > are not required to achieve equivalent function. > > > > A lot of computer science tends to be like this in practice ( e.g. the > > ever ubiquitous balanced tree). > > > > Cheers, > > > > J. Andrew Rogers > > > > ----- > > This list is sponsored by AGIRI: http://www.agiri.org/email > > To unsubscribe or change your options, please go to: > > http://v2.listbox.com/member/?& > > > > ________________________________ > This list is sponsored by AGIRI: http://www.agiri.org/email > To unsubscribe or change your options, please go to: > http://v2.listbox.com/member/?& -- Vladimir Nesov mailto:[EMAIL PROTECTED] ----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=8660244&id_secret=47999791-00fced
