Peter's comments, as usual, are provocative. I have comments on each
of his two parts.
The first part and comment concerns the quote from David Larkin, which
leads to the following question (in my terms, as I understand the
problem): "Is there any practical difference between (1) using a
bivalent first-order logic, whose count of predicates needed to model
natural language has an upper-bound not equal to aleph null, but
instead equal to the count of positive real numbers, and (2) using a
multi-valued first-order logic with at most aleph null predicates?"
That is, should we force the potentially infinite classification
categories of (some? all?) predicates to be expressed syntactically by
inventing additional predicates (e.g., "a little bald," "a little more
bald," "somewhat bald," * ) or should we build the degrees into the
semantics, using multi-valued degree-of-truth matrices?
The best general answer, I think, is that is all depends on what we're
up to in a particular application, and on which system will help us do
it better. In some tasks, it is more useful to use bivalent
approximations and to multiply predicates, while for other tasks it is
more useful to simplify the syntactic models and to complicate the
semantics. And there may be formal reasons for preferring one or the
other as a default system. My evidence for thinking that this
case-by-case approach is the "best" general answer is that this is
exactly what we do in the real world.
My second comment concerns Peter's posed problem about which system is
more helpful in deciding whether " 'roller skates' are 'motor
vehicles' for purposes of an ordinance prohibiting the operation of a
motor vehicle in a park?" My quick comment is that Peter's problem is
not correctly worded: the word "are" should be changed to "should be
considered." Most of law (including a great deal of what we call
"factfinding") is concerned with governmental use of words to invent
and enforce the use of other words, and thereby to justify
governmental action. The law does not use words merely to describe the
world.
But Peter asks a probing question about his posed problem: "Is it
useful to think in terms of shifting sets (rough sets) or varying
degrees of membership in a set (fuzzy set)? If not, why not?" So let
me address this last question more completely.
One of the most useful language games we have invented depends upon
society's creating and enforcing the use of "observational predicates"
* predicates whose correct usage is warranted entirely by direct
individual perceptions of (what we call) observable objects. Examples
are "red," "loud," "sharp," and "seventeen." By creating such
predicates and enforcing their "correct" usage in elementary school
and everyday life, we generate a fairly uniform word usage within a
linguistic culture. Such uniformity is caused by a high degree of
homogeneity in our physiology of perception (thanks to evolution) and
a high degree of homogeneity in the physical behavior of light
absorption and reflection. Once we create and police the use of such
"observational predicates," it also becomes useful to develop and test
models that predict "our use" of some observational predicates given
"our use" of other observational predicates. This only works because
we can agree in advance, with a high degree of consensus and
independently of the model, how we will describe or classify the event
outcomes of our trials. With respect to such predicates, we can then
debate and test our predictive models, including our logic and
algebraic models. That is, we can usefully (some would say
"meaningfully") ask the first question above (David Larkin's).
But Peter's predicates are not so "well-behaved" * that is, we are not
"well-behaved" when we use them. So any model-validation effort seems
pointless (useless) from the outset. Of course, law is not simply (in
the sense of "only") linguistic politics. Factfinding in law involves
both observational predicates and many other types of predicates,
whose usage is designed and policed through various societal
processes. Many of these linguistic patterns and many of these
processes are highly resistant to formal modeling. I say "highly
resistant" because it is not always easy to identify the practical
problems being addressed, and therefore not easy to determine whether
the use of particular formal models will help or hinder the solution
of those problems. One practical but general problem on which we might
want to make headway, however, is how to combine observational
predicates and other types, and provide decisional rules for
linguistic usage on "ultimate issues." But even here, we have to look
at our legal problems on a more case-by-case basis, to see what might
be useful.
On this, perhaps, most of us would agree: the "rule of law" entails a
preference (other things being equal) for rule-based, consistent
linguistic usage, so that similar cases are treated similarly, and
future governmental action is justifiable by past governmental
process. But in view of the vast number of different legal problems,
the diverse nature of legal predicates, and the ever-present politics
of governance, any progress with formal systems is almost inevitably
"local" * advanced, at most, problem by problem, and inference-type by
inference-type. And there's a good chance that with Peter's specific
problem ("roller skates" and "motor vehicles"), the most we can ever
get, because the most we ever even want, is rule by fiat.
Happy New Year to all!
Vern.
* * * * *
Vern R. Walker, Ph.D., J.D.
Professor of Law
Hofstra University School of Law
121 Hofstra University
Hempstead, New York 11549
Tel. (516) 463-5165
Fax (516) 463-4962
URL: <people.hofstra.edu/faculty/vern_r_walker>
>>> "Peter Tillers" <[EMAIL PROTECTED]> 12/30/03 7:01 PM >>>
One or two members of the list have made points like the following one by
David Larkin:
>With standard Bayesian techniques it is possible to reason
probabilistically about variables with non-binary domains. For example, in a
standard belief network one might find a node labeled "Temperature" which
can take on the values "Low", "Medium", and "High". A fuzzy-logic based
interpretation of the variable would amount to a simple relabeling. The node
could be renamed, "The object is hot", taking a truth value of "0", "0.5",
or "1". Mathematically, there would be no difference. You would say,
according to bivalent logic, "The object is hot" can only be true or false,
and we can only reason probabilistically about the chance that it will be
true or false, and therefore partial truth cannot be addressed. But standard
belief networks are not bivalent (i. e., the variables are not binary), but
in general multi-valent. Vague and perception-based degrees of truth, such
as the temperature example given above, are commonly used to describe the
states of variables.<
I have a question (that I penned some time ago) for members of this list who
share David Larkin's sentiment:
How would I use the procedure of assessing the probabilities of multivalent
variables to determine the probability (of the proposition or hypothesis)
that "roller skates" are "motor vehicles" for purposes of an ordinance
prohibiting the operation of a motor vehicle in a park? Would you propose
that one of the hypotheses whose probability I am to assess is "roller
skates have the property of being a 'motor vehicle' only to a small extent"?
(I am taking "motor vehicle" as a primitive property or attribute, rather
than as one that is decomposable into subordinate properties or attributes.)
If this formalism were to show us anything, we would need to figure out what
events or evidence or symptoms etc. we should use in expressions such as
P(E|"'roller skates' have small amount of the property 'motor vehicle'")
Most legal scholars (if they understood probability theory well, that is)
and some probability theorists would be tempted to conclude that the concept
of a conditional probability cannot readily be used to address uncertainties
of the sort found immediately above. If so, can we use a different kind of
"formalism"? Is it _conceivable_ that we could do so? If so, what would that
way of thinking be like? Is it useful to think in terms of shifting sets
(rough sets) or varying degrees of membership in a set (fuzzy sets)? If not,
why not?
***
In the world of accounting there are often alternative ways of describing
the same financial state of affairs. The descriptions one gets often seem
different; but they are often not inconsistent, and sometimes they amount to
paraphrases, different but equivalent ways of saying the same thing.
Sometimes (but not always) the differences between probability theorists and
fuzzy set theorists strike me as being like that. In response to formulation
x -- e.g., think of this uncertainty as "degree of truth" -- there is the
response: e.g. #1: we can instead think of this problem, or we can instead
describe this problem, as a situation in which we are trying to assess our
uncertainty about a continuum of possible states, we can instead speak of
the probabilities of continuous variables; e.g. #2: we can instead speak of
the probabilities of non-continuous but multiple disjoint possible states.
Speaking as a non-mathematician: sometimes such alternative ways of
describing uncertainty seem to me to be, at bottom, "the same", and if I
were to prefer one way of talking over another, I would want to have some
_reasons_ for talking about a particular kind of uncertainty one way rather
than another way. Although I think there are in fact some advantages
(sometimes) to using the grammar of conditional probability to talk about
hypotheses about legal norms -- in this regard I am in a small minority
among legal scholars --, I also cannot escape the sense that the grammar of
conditional probability just does not "get at" some of the types of
uncertainty that are found in legal reasoning (and in other contexts as
well). Think of the underlying intuition of rough sets, the notion that some
words or classifications seem to flip and shift, that they have fluctuating
perimeters: How does one get at this type of uncertainty, effectively
capture it, by talking in the language of conditional probabilities? I do
not deny that it could be "done" if one tried "hard enough" -- standard
probability theory is protean: it can be interpreted in an immense variety
of ways -- but in the end I think I would still wonder if deployment of
notions such as conditional probabilities, likelihood ratios etc. would be
the best way to think about words that seem to have "rough edges" or to
about the appropriate categorization of things that seem to fall into two or
more categories that to some extent seem disjoint to some degree but not
strictly or completely disjoint.
Since I am not a fuzzy or rough set theorist, I could not begin to prove
that my doubts (such as those mentioned immediately above) about the
relative fertility of the standard probability calculus in some contexts are
justified or that fuzzy sets or rough sets make it possible to think in a
more orderly and more productive manner about uncertainties such as those
mentioned in the preceding paragraph. This is why I (desperately) need the
help of logicians and mathematicians and probability theorists etc. But if
you were to advise me, I would want you to keep this in mind: I would not
think you have made a good defense of your preferred theory -- your
preferred theoretical apparatus -- merely by showing that your theory or
apparatus can "handle" some problem that a rival theory also claims to
handle. The question, it seems to me, is which theory does a _better_ job of
"handling" the problem at hand.
Nothing I say here, of course, answers the last question. Here I am not even
_attempting_ to answer the last question. Furthermore: I confess that I do
not know the answer to the last question. I only know that conditional
probability talk sometimes just doesn't seem to get at the kind of
uncertainty that I sometimes see in law. (The reverse is also the case: talk
about fuzzy and rough sets also sometimes just doesn't seem to fully get at
the kind of uncertainty that I think I sometimes see in law [unless, of
course, we are forced to agree with the thesis that the standard probability
calculus is just a special case in fuzzy sets].)
When considering the general question I pose in this message, please keep in
mind that I am a legal professional and that lawyers and legal professionals
are very interested in "word wars" -- in battles over the meaning of words
and in _arguments_ about the meaning or scope of legal classifications. The
way a word or classification is conventionally used is part of the kind of
argument that a lawyer often makes. But only part. Lawyers appeal to reasons
other conventional usage when they argue about the "proper" or "appropriate"
or "correct" interpretation of some legal phrase or classification. (Much
depends on the context in which a lawyer is operating. When lawyers draft
contracts, they are more likely to emphasize conventional {legal}
interpretation. When they litigate, they are likely to emphasize
conventional usage considerably less.)
Peter T
P.S. I suppose that a discussion list -- even one as distinguished as
this -- is not the appropriate venue for a discussion of the implication of
the principle of non-contradiction for the question of the (logical)
coherence of fuzzy set-based arguments. (When I have time -- but I am
running out of time! -- I will reconsider this basic(?) question in my own
autodidactic way. [There is a crasser but more trenchant way to describe my
efforts at self-education. The adjective I have in mind is a hyphenated
expression that begins with "h" and ends with "d".])
****
Peter Tillers http://tillers.net
Professor of Law
Cardozo School of Law
55 Fifth Avenue
New York, New York 10003
U.S.
(212) 790-0334 [EMAIL PROTECTED]
FAX (212) 790-0205
------- End of Forwarded Message