Jim,

You somehow completely misunderstood what I was attempting to communicate.

Continuing...
On Mon, Nov 9, 2015 at 5:35 PM, Jim Bromer <[email protected]> wrote:

> Steve,
> I understand what you are saying (even though the details of what you
> actually write are mostly irrelevant to me).
>

... which probably explains why you so misunderstood what I was
communicating.


> I have done many thought experiments about the AI problem.


Then please post some of them, e.g. what you might say, and what you might
expect in response.


> Your
> challenge in your last email is not something that is new to me, I
> have thought about that exact question many times. However, I finally
> figured out that instead of spending a lot of time trying to get you
> on a higher plane of discussion I should use the opportunity to my
> best advantage and try to explain my theories again.
>

... and MY point was that theories be damned, that there is only one known
way of extracting meaning from text at a practical rate of speed. If you
can't adapt THAT, then your quest is truly hopeless.

>
> So the program does not need multiple IO facilities in order to
> understand some things.


I never said it did. I was addressing the ways that things can NOT work.
That can't be simply turned around to say that doing something else WILL
work.


> Why don't I think that multiple IO modalities
> are necessary for advancements in AI? I get that it would make a truly
> advanced AI program more sophisticated but trying to use sensors and
> robotics at this time would just make the contemporary programming
> more difficult. The data coming in through video, for example, needs
> to be analyzed before it can be used by an AI program. Stronger AI
> can't just take input from multiple IO modalities and automatically
> turn current AI into Strong(er) AI.
>
> But there is another side of the opinion that multiple IO modalities
> are necessary for Stronger AI. The most obvious one is that much or
> our human experience is based on multiple IO so of course a true AGI
> program would have to be adaptable to different IO situations. (I
> already get that.) However, the next question down is whether
> different senses and IO are absolutely necessary for any AGI program.
>

My point was that EXPERIMENTATION is absolutely necessary to sort out the
countless possible realities that spring forth from textual explanations.
It is a bit difficult to do experimentation from inside a box.


> To make another effort to explain this point of view I will say once
> again in a new way that without certain advancements in AI no number
> if IO modalities are going to make the program work the way we would
> want to. So to highlight my area of interest in another way let me
> just restate that I want to do some study of how the AI part of the AI
> program can be improved in order to get it to work the way that I
> think it needs to work.
>
> I don't believe that current AI / AGI methods have Conceptual
> Integration methods that are sophisticated enough.
> I believe that my theories on Conceptual Integration, while being
> composed of methods that are mundane, should be powerful enough to
> deal with many of the contemporary problems. My methods will not
> produce true human-like AGI but I believe I can produce AI that human
> beings will be able to communicate with using language that is much
> closer to natural human language. That does not mean that the program
> will be able to understand anything that is readable, but that it will
> be able to work with knowledge that it has acquired and use a more
> natural kind of language to communicate about it. Does that help you
> to better understand what I am interested in?
>

I think everyone here agrees that presently known methods do NOT lead to
AGI, regardless of how far you might extend them with better hardware, etc.

>
> I have talked about ambiguity in language many times in this group. I
> am surprised that you don't remember that I did so.


Perhaps you missed my thread about disambiguation being a really bad idea?


> I have also talked
> about the anaphoric-like reference problem.
>
> Simple ambiguity is a human problem. For instance, you thought I was
> one of the guys who was promoting the theory that a language-based AI
> program would just be able to learn from the internet. (I am not one
> of those guys.) That was an ambiguity-type problem. It was also a
> reference problem. You misunderstood what I was referring to when I
> was talking about AI / AGI. It can be resolved by communication using
> natural language unless there is something interfering with that
> communication.
>
> However, the problem with natural language AI is that the potential
> for massive ambiguity and referential confusion makes the AI problem
> seem intractable. I believe that I have some good theories to deal
> with that - but at the same time I realize that I am going to run up
> against the knowledge complexity problem very quickly.
>
> So I will try to explain what I want the program to do after I have
> taken a day or two to plan my response carefully and write it out as
> simply as I can.
>

OK. I'll look for that.

*Steve*
===================

>
> Jim Bromer
>
>
> On Mon, Nov 9, 2015 at 3:10 AM, Steve Richfield
> <[email protected]> wrote:
> > Jim,
> >
> > OK, maybe a simple thought experiment will help. Presume for a moment
> that
> > the system you imagine exists here and now. You have just turned it on.
> Now,
> > start typing (or talking) to make it do what you want it to do:
> >
> >
> >
> >
> >
> > Steve
> >
> > On Sun, Nov 8, 2015 at 11:58 PM, Jim Bromer <[email protected]> wrote:
> >>
> >> Steve,
> >> I do not understand how you can misunderstand what I am talking about
> >> so badly since I have posted hundreds if not thousands of messages on
> >> groups like this.
> >> I am not planning on getting a computer program to learn by reading
> >> the internet with the expectation of having it achieve human like
> >> abilities. The supposition that that is what I am talking about
> >> strikes me as nonsensical.
> >>
> >> Unfortunately, I probably will not be able to do much at all but I
> >> hope to at least experiment with a higher learning mechanism that, for
> >> example, can be taught through conversation. In order to achieve this
> >> I am planning on using a parallel referent language which will be
> >> designed to shape knowledge about some things in ways that would be
> >> like rough prototypes of the kinds of ideas that I have in mind for an
> >> AI / AGI program.
> >>
> >> I agree that in order to 'understand' text the program would need to
> >> build models about reality. However, I am in total disagreement with
> >> the idea that an AI program would need to have some sort of embodiment
> >> to understand text about the world. Such thinking strikes me as
> >> absurdly naïve. The reason is that I believe there are very basic
> >> parts of AI methods that are almost totally missing in contemporary AI
> >> programs and the fantasy that if you could just slap some robotics
> >> onto one of these primitive AI designs then the program would
> >> magically be more capable of understanding is somewhere near the
> >> height of absurdity.
> >>
> >> So I want to try to improve on the AI programming by working on a
> >> design that would be able to integrate simple knowledge or knowledge
> >> about simple things together. I don't think deep learning is good
> >> enough for this because deep learning neural networks have a such a
> >> crude way of going about this integration of related ideas. Of course,
> >> I am aware that there are other problems with complexity that are
> >> difficult to work around. I have been trying to find a polynomial time
> >> solution to SAT for the past few years because I thought that would
> >> help me with complexity. However, I haven't been able to make any real
> >> progress on that and I just don't see how the human mind could
> >> possibly be using p=np in some way given the propensity of 'organic'
> >> processes to generate unique manifestations of very irregular
> >> patterns.
> >>
> >> So back to the main topic of my thought. Interactions with the sensory
> >> world of human beings is not necessary for AI because the information
> >> that comes in from those senses would have to be integrated well to
> >> produce 'understanding' just as information that comes in through the
> >> keyboard would need to be. It is my belief that smart conceptual
> >> integration is an outstanding problem in contemporary AI and it is a
> >> problem I can work on.
> >>
> >> Unfortunately I won't be able to do much work on the problem. Perhaps
> >> I squandered too much time writing in groups like this and working on
> >> my p=np? project. On the other hand, I hadn't really settled down on
> >> what I think is the best way to approach the problem until the last
> >> few years. But I would at least like to try to do something.
> >> Jim Bromer
> >>
> >>
> >> On Sun, Nov 8, 2015 at 5:16 PM, Steve Richfield
> >> <[email protected]> wrote:
> >> > Jim,
> >> >
> >> > You seem to be suffering from a malady that many others on this forum
> >> > appear
> >> > to be suffering from. The primary symptom is the belief that simple
> >> > observation of HIGHLY noise-ridden text can lead machines to useful
> >> > epiphanies.
> >> >
> >> > The problem is that pretty much everything people write is, in a word,
> >> > wrong. Overbroad statements (and natural languages can NOT precisely
> >> > bracket
> >> > statements), statements made based on unstated presumptions (and no
> one
> >> > can
> >> > list all their presumptions), statements made based on faulty models
> >> > (and
> >> > ALL models are faulty, as every physicist knows), confused meanings or
> >> > words
> >> > (most words have multiple meanings), lies made for economic or other
> >> > gain,
> >> > etc., etc., etc. It is difficult to find ANY clear statements that are
> >> > unquestionably accurate in all of their potential meanings.
> >> >
> >> > OK, so suppose you accept the above and simply want to do the best you
> >> > can.
> >> > Still, you can NOT approach the capabilities of an average person,
> >> > because
> >> > we have a real world in which to test the many possible
> interpretations
> >> > of
> >> > what we read, whereas a machine can only accept, reject, or assign a
> >> > probability with NO other information.
> >> >
> >> > I used to believe as you do - until 2001 when I became very sick. It
> >> > took me
> >> > 3 months of every conscious minute to figure out what was wrong and
> what
> >> > might be done about it. Sure I got most of my information from books
> or
> >> > the
> >> > Internet, but there remained several very different models in which
> this
> >> > all
> >> > fit, which I resolved with phone calls to key researchers, and a
> little
> >> > of
> >> > my own primary experimentation to sort the flyshit from the pepper.
> >> > Still, I
> >> > remained unsure until what should work did work to cure my condition.
> >> >
> >> > When I went back to figure out how the Internet might be reorganized
> to
> >> > reduce my search from 3 months to a few minutes, the structure of
> >> > DrEliza
> >> > emerged, and later my patent.
> >> >
> >> > The problem is that to usefully solve problems you need MODELS, yet
> >> > natural
> >> > language (and most human thinking) predates this concept and only
> >> > provides
> >> > INFORMATION. Sometimes you can construct a model from information, but
> >> > this
> >> > is RARE. Making models requires highly qualified geniuses who can get
> >> > their
> >> > arms around entire fields and synthesize models that fit ALL known
> >> > observations. I have done this in several narrow areas, but it will
> take
> >> > a
> >> > LOT more of this to transform the Internet to a model-based system
> from
> >> > which an AI/AGI can usefully address problems of all types.
> >> >
> >> > Further, natural language is highly granular - there are far more
> things
> >> > that you can NOT say than things you CAN say, so people without even
> >> > thinking about it round to the nearest syntactically expressible
> meaning
> >> > in
> >> > EVERYTHING they say or write. This "rounding" completely destroys any
> >> > ability to construct accurate models.
> >> >
> >> > Some languages have weak workarounds to this, e.g. German with its
> >> > concatenated words, or Arabic where they alter spellings for emphasis,
> >> > but
> >> > these measures only slightly reduce their granularity.
> >> >
> >> > All in all, each person here must either find a way to jump from
> >> > information
> >> > to models, or abandon this quest. Sure, information can help people
> >> > solve a
> >> > problems whose solution has already been stated, but there are already
> >> > plenty of experts around who do this quite well. It is the UNsolved
> >> > problems
> >> > that are interesting, and it is these UNsolved problems that can NOT
> >> > EVER be
> >> > solved by automated means based on people's writings.
> >> >
> >> > How can I say not EVER when writings continue to accumulate? Because
> >> > society's problems also continue to accumulate, so as people find
> >> > solutions
> >> > to past unsolved problems, even more new problems emerge to replace
> >> > them.
> >> >
> >> > I hereby proclaim your apparent quest to be theoretically unachievable
> >> > for
> >> > the many reasons outlined above. Sure it would be of astronomical
> value,
> >> > like the methodology to change lead into gold that so many people put
> so
> >> > much effort into, but why waste your time unless/until you can find
> SOME
> >> > way
> >> > around the above-listed barriers.
> >> >
> >> > Steve
> >> > =====================
> >> >
> >> > On Sun, Nov 8, 2015 at 9:00 AM, Jim Bromer <[email protected]>
> wrote:
> >> >>
> >> >> Steve,
> >> >> It will take me some time to reply carefully so let me respond to
> >> >> something I feel strongly about.
> >> >>
> >> >> >>And because it is not an all encompassing
> >> >> language of communication it could be used to test the 'emergence' of
> >> >> insight that could arise if enough preparatory work had been done,
> >> >> even if I haven't figured out how that could be done without the
> >> >> artificial referent language.
> >> >> >>
> >> >>
> >> >> >There is a VAST chasm between being able to define language
> >> >> > constructions
> >> >> > and meanings, and "insight".
> >> >> >
> >> >>
> >> >> I believe there is a vast chasm between 'simple associations' or
> >> >> 'simple
> >> >> correlations' or associations derived from 'neural networks' and
> >> >> conceptual
> >> >> integration. Sophisticated artificial conceptual integration would
> make
> >> >> 'insight' feasible and simple examples across a wide range of subject
> >> >> matter
> >> >> should arise fairly quickly. But since AI programs are only capable
> of
> >> >> the
> >> >> simplest examples of 'insight' then declarations about the chasm
> >> >> between AI
> >> >> and 'insight' are expected. So I totally disagree with you about
> this.
> >> >> I
> >> >> feel that your feelings about this are historically accurate but have
> >> >> little
> >> >> to do with the potential near-future. As I say, I do not recall
> hearing
> >> >> about an AI program that is capable of learning via conversation
> except
> >> >> for
> >> >> extremely simple domains. I feel that I have a solution for this
> >> >> problem but
> >> >> the trial and error process of getting from where I am now and where
> I
> >> >> think
> >> >> I can get is so overwhelming a challenge that my decision to use the
> >> >> artificial referent para-language makes a sense.
> >> >>
> >> >> Jim Bromer
> >> >>
> >> >> On Sun, Nov 8, 2015 at 11:19 AM, Steve Richfield
> >> >> <[email protected]> wrote:
> >> >>>
> >> >>> Jim,
> >> >>>
> >> >>> FINALLY - SOMEONE who wants to discuss PRACTICAL implementations of
> >> >>> TAI/TAGI.
> >> >>>
> >> >>> Continuing...
> >> >>> On Sun, Nov 8, 2015 at 7:14 AM, Jim Bromer <[email protected]>
> >> >>> wrote:
> >> >>>>
> >> >>>> After I wrote that message I realized that I had tried to start
> >> >>>> discussions about an artificial language that could be used to
> shape
> >> >>>> a
> >> >>>> general AI program before. Many of these discussions were side
> >> >>>> tracked
> >> >>>> when people started talking about Esperanto or about lambda
> calculus
> >> >>>> based artificial languages and stuff like that. That is not what I
> am
> >> >>>> thinking of.
> >> >>>
> >> >>>
> >> >>> You mean, having syntax like:
> >> >>>
> >> >>> When that I write "xxxx" I mean "yyy".
> >> >>>
> >> >>> to define idioms, for more subtle things like:
> >> >>>
> >> >>> Consider that when I write "," I may mean ";".
> >> >>>
> >> >>> which expresses potential alternative interpretations of future
> >> >>> writings?
> >> >>>>
> >> >>>>
> >> >>>> The artificial language could be used with video or audio or other
> >> >>>> kinds of IO environments, but I would use it along side of an
> attempt
> >> >>>> to get the AI program to learn to use a natural language.
> >> >>>
> >> >>>
> >> >>> I did a VERY similar thing in a FORTRAN/ALGOL/BASIC compiler I once
> >> >>> wrote
> >> >>> for Remote Time-Sharing Corp. It started out as a very simplistic
> >> >>> metacompiler, to which I fed it a description of a more capable
> >> >>> metacompiler, in which language I fed it a description of an
> >> >>> optimizing
> >> >>> metacompiler.
> >> >>>
> >> >>> This could easily be done in a language like English, where a
> >> >>> rule-driven
> >> >>> system like I have been discussing here has rules whose function is
> to
> >> >>> introduce new rules.
> >> >>>
> >> >>>>
> >> >>>> One of the
> >> >>>> dreams of old AI was that if you started instructing the program to
> >> >>>> learn using the artificialities of some kind of language it would
> >> >>>> eventually have enough information for genuine learning to emerge.
> >> >>>
> >> >>>
> >> >>> The think that seems to be the killer here is erroneous learning of
> >> >>> various sorts. Superstitious learning is theoretical unavoidable.
> Once
> >> >>> you
> >> >>> get something erroneous into such a system, it becomes
> >> >>> difficult/impossible
> >> >>> to get it out. A VERY simple demonstration comes in trying to use
> >> >>> Dragon
> >> >>> NaturallySpeaking's speech input to correct its errors in your
> >> >>> dictation. As
> >> >>> you would expect it makes errors in trying to correct the errors,
> and
> >> >>> this
> >> >>> often compounds to overwhelm any hope of setting things right.
> >> >>>
> >> >>> Add to that not knowing exactly what a computer got wrong, or even
> >> >>> being
> >> >>> able to recognize that the computer got something wrong, and you can
> >> >>> see how
> >> >>> difficult/impossible it is to correct wrongly "learned" rules.
> >> >>>
> >> >>>>
> >> >>>> This never really worked. Why not? Partly because computers were
> not
> >> >>>> powerful enough in the old days
> >> >>>
> >> >>>
> >> >>> And still aren't - unless you use my patented LFU methodology.
> >> >>>
> >> >>>>
> >> >>>> and, in my opinion, AI researchers had
> >> >>>> not appreciated the necessity of sophisticated data integration
> >> >>>> methods for some reason. (Old computer systems might one day be
> shown
> >> >>>> to have been potentially powerful enough to run some future program
> >> >>>> but they were not powerful enough to entertain the trial and error
> >> >>>> process that would have been required using experimental programs
> of
> >> >>>> the day.
> >> >>>
> >> >>>
> >> >>> The advantage in LFU is about the same as the advantage of a modern
> PC
> >> >>> over an old vacuum tube clunker, so yes, they could have done a LOT
> >> >>> more way
> >> >>> back then.
> >> >>>
> >> >>> The "cycle time" of an IBM-709 computer was 12 microseconds, and
> most
> >> >>> instructions took two cycles - one to access and interpret the
> >> >>> instruction,
> >> >>> and one to access and operate on the operand.
> >> >>>
> >> >>>>
> >> >>>> For example, with better conceptual integration methods a
> >> >>>> future efficient AI program might be used on an old computer system
> >> >>>> just to show that it could be run on it.)
> >> >>>
> >> >>>
> >> >>> No, except for a few in the Computer Museum's display in Cupertino
> >> >>> they
> >> >>> have all been melted down for their scrap metal, and the Museum
> won't
> >> >>> turn
> >> >>> them back on.
> >> >>>>
> >> >>>>
> >> >>>> So the artificial referent language would not be a complete
> language
> >> >>>> (of communication) like Esperanto wants to be. And it would not be
> a
> >> >>>> logically sound language like lambda calculus wants to be. It could
> >> >>>> be
> >> >>>> used to establish referents from the IO data environment. It would
> >> >>>> need to be capable of denoting a distinction between how those data
> >> >>>> objects can be used. For example in natural language there is an
> >> >>>> important distinction between syntax and semantics. So if I used
> this
> >> >>>> referent language with a natural language IO then one of the
> >> >>>> artificialities would be to distinguish syntactic relations from
> >> >>>> semantic relations. On the other hand, this distinction is not
> always
> >> >>>> necessary, desired or clear cut. To explain this, many (or maybe
> >> >>>> most)
> >> >>>> (what I think are) desirable syntactic relations are based on some
> >> >>>> semantic conditions. But then again there is no reason not to
> design
> >> >>>> the artificial language to be able to represent relations that are
> >> >>>> mixes of semantics and syntax.
> >> >>>
> >> >>>
> >> >>> Leaving a stupid computer to untangle such messes is probably a
> >> >>> mistake.
> >> >>> However, it would be fairly easy to provide a mechanism for people
> to
> >> >>> specify such things.
> >> >>>>
> >> >>>>
> >> >>>> As I see it, the main problem with language based AI has been the
> >> >>>> lack
> >> >>>> of a really good conceptual integration solution.
> >> >>>
> >> >>>
> >> >>> This broad statement could be said about ANYTHING people haven't yet
> >> >>> seen
> >> >>> a way to make work - like AGI.
> >> >>>>
> >> >>>>
> >> >>>> One of the reasons I write to groups like this is that I want to
> get
> >> >>>> some ideas about how an idea might work.
> >> >>>
> >> >>>
> >> >>> Same here.
> >> >>>
> >> >>>>
> >> >>>> But when I wrote about an
> >> >>>> artificial para-language before I wasn't really sure it I even
> wanted
> >> >>>> to use it. I finally have come to the conclusion that it makes a
> lot
> >> >>>> of sense. I can use it to speed up tests about my AI/AGI theories
> but
> >> >>>> then I could also test those theories with more relaxed
> instructions.
> >> >>>> So the artificial para-referent language would not a all
> encompassing
> >> >>>> language of communication, it would not be a logically sound
> language
> >> >>>> other than to denote semantic and syntactic references and
> relations
> >> >>>> based on mixes of semantic and syntactic references. It could also
> >> >>>> denote relations that I think would be important to a text-based
> >> >>>> AI/AGI program. Because the logic of the method would not be tight
> >> >>>> and
> >> >>>> a contradicting case would not (always) lead to an artificially
> >> >>>> reported error, the AI methods would have to do some learning for
> >> >>>> itself. So the para-referent language would not sidetrack the whole
> >> >>>> effort because if the AI methods have to have the potential to
> >> >>>> exhibit
> >> >>>> some genuine learning. And because it is not an all encompassing
> >> >>>> language of communication it could be used to test the 'emergence'
> of
> >> >>>> insight that could arise if enough preparatory work had been done,
> >> >>>> even if I haven't figured out how that could be done without the
> >> >>>> artificial referent language. The benefit is that I could use it to
> >> >>>> test and develop my AI theories. I am really excited by this idea
> >> >>>> this
> >> >>>> time.
> >> >>>
> >> >>>
> >> >>> There is a VAST chasm between being able to define language
> >> >>> constructions
> >> >>> and meanings, and "insight".
> >> >>>
> >> >>> Steve
> >> >>> ======================
> >> >>>>
> >> >>>> Jim Bromer
> >> >>>>
> >> >>>>
> >> >>>> On Sat, Nov 7, 2015 at 10:22 PM, Jim Bromer <[email protected]>
> >> >>>> wrote:
> >> >>>> > I was just working on my latest p=np? idea and I hit up against
> >> >>>> > method
> >> >>>> > that is either in np or is otherwise extremely inefficient. So I
> >> >>>> > have
> >> >>>> > to come to the conclusion that the human mind is not capable of
> SAT
> >> >>>> > in
> >> >>>> > p.
> >> >>>> >
> >> >>>> > So then how do we figure how to deal with so many complicated
> >> >>>> > situations? Of course I still don't know because so many
> situations
> >> >>>> > seem similar to a SAT problem. The mind must be able to detect
> many
> >> >>>> > different things that are going on at once or which might be
> useful
> >> >>>> > to
> >> >>>> > recall from memory to deal with a situation. But still, there is
> >> >>>> > nothing in my own introspective analysis of my thinking which
> looks
> >> >>>> > anything like a p=np process.
> >> >>>> >
> >> >>>> > So what is wrong with AI? One thing that AI has been consistently
> >> >>>> > lacking is the ability to learn through conversation. My feeling
> is
> >> >>>> > that this is not just a problem with communication but a learning
> >> >>>> > problem as well. In other words AI is not able to truly learn
> >> >>>> > except
> >> >>>> > in a few special cases. Most of those special cases are examples
> of
> >> >>>> > narrow AI but there are others where the learning that takes
> place
> >> >>>> > isn't necessarily like other narrow AI but where the domain of
> >> >>>> > learning is so restricted that it is narrow in the sense that the
> >> >>>> > applicability of the method is limited.
> >> >>>> >
> >> >>>> > Then I started thinking of an artificial language which can refer
> >> >>>> > to
> >> >>>> > situations or objects in the IO data environment and which can be
> >> >>>> > used
> >> >>>> > to instruct a program as it is running. I think this is an
> unusual
> >> >>>> > idea.
> >> >>>> >
> >> >>>> > One of the characteristics about programming methods that seem to
> >> >>>> > catch on with programmers is that they can be used in a very
> simple
> >> >>>> > manner and in more complicated programming. I think an artificial
> >> >>>> > language which could be used to instruct a computer to notice
> >> >>>> > objects
> >> >>>> > in the IO data environment and which could also be used to refine
> >> >>>> > those instructions using this artificial language with the
> >> >>>> > references
> >> >>>> > that it had previously established has a lot of potential. And it
> >> >>>> > can
> >> >>>> > help us become more clear about what is needed to make better AGI
> >> >>>> > programs.
> >> >>>> > Jim Bromer
> >>
> >>
> >> -------------------------------------------
> >> AGI
> >> Archives: https://www.listbox.com/member/archive/303/=now
> >> RSS Feed:
> https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac
> >> Modify Your Subscription: https://www.listbox.com/member/?&;
> >> Powered by Listbox: http://www.listbox.com
> >
> >
> >
> >
> > --
> > Full employment can be had with the stoke of a pen. Simply institute a
> six
> > hour workday. That will easily create enough new jobs to bring back full
> > employment.
> >
> > AGI | Archives | Modify Your Subscription
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>



-- 
Full employment can be had with the stoke of a pen. Simply institute a six
hour workday. That will easily create enough new jobs to bring back full
employment.



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to