Re: Unicode is so flawed that 7 or 8 bit encoding is not an issue

2002-03-21 Thread Masataka Ohta

James;

  As you could have seen, on IETF mailing list, Harald and I have, at
  least, agreed that, if you use unicode based encoding, local context
  (or locale) must be carried out of band
 
 Few will disagree (including me) that using Unicode to do localization is
 almost impossible without locale context.

Huh? No one said such a thing.

What is agreed is that, to use unicode, it must be supplied out of
band local context.

So, could you explain how the context is supplied with current IDN spec?

 But I am touching on a sensitive topic about what you considered as I18N 
 L10N and what the rest of the world thinks.

Totally irrelevant.

  But, anyway, the discussion so far in IETF list is enough to deny IDN.
 
 I am not aware there was a IETF wide last call on IDN yet.

To deny IDN, IESG can simply say IDN is disbanded. There is no last
call necessaruy.

  I have found a theory to deny PKI and to explain why public key
  cryptograpy is not so polular dispite the efforts of ISO, which will
  destory entire business of a company or companies owing several
  important TLDs.
 
 Very interesting. Could you share the theory, or privately if you prefer?

I will post it later if I have time.

Masataka Ohta




Re: Unicode is so flawed that 7 or 8 bit encoding is not an issue

2002-03-21 Thread James Seng

  Few will disagree (including me) that using Unicode to do localization
is
  almost impossible without locale context.

 Huh? No one said such a thing.

 What is agreed is that, to use unicode, it must be supplied out of
 band local context.

In that case, I disagree with to use unicode, it must be supplied out of
band local context.

You only need out of band local context to use Unicode if you are doing
localization.

 So, could you explain how the context is supplied with current IDN spec?

Internationalization as defined in POSIX, means making software location
neutral. Making a program specific to any particular language or culture or
codeset is known as localization.

I in IDN is Internationalization.

  I am not aware there was a IETF wide last call on IDN yet.

 To deny IDN, IESG can simply say IDN is disbanded. There is no last
 call necessaruy.

Ah of course IESG can. But sorry, you dont represent IESG.

   I have found a theory to deny PKI and to explain why public key
   cryptograpy is not so polular dispite the efforts of ISO, which will
   destory entire business of a company or companies owing several
   important TLDs.
 
  Very interesting. Could you share the theory, or privately if you
prefer?

 I will post it later if I have time.

Please do. Thanks.

-James Seng




Re: I don't want to be facing 8-bit bugs in 2013

2002-03-21 Thread Meritt James

You want to be facing 8-bit bugs in 2002?  I recommend reconsideration
of priorities.

-- 
James W. Meritt CISSP, CISA
Booz | Allen | Hamilton
phone: (410) 684-6566




RE: Netmeeting - NAT issue

2002-03-21 Thread Tony Hain

Aaron Falk wrote:
 I think one can make the case that having border protection may
 prevent a DOS attack from consuming interior network resources and
 allowing interior hosts to communicate amongst themselves.

And if your interior network resources are less than 10x your external
resource, you have an unusual network indeed. Yes it may be more
convenient to have the border deal with DOS, but is it *required* as
Noel asserted?

 We've
 recently had some fierce DOS attacks on our ISP but I'm still able to
 run NFS without a problem. This is a good thing.

NFS  'good thing' are a matter of personal opinion. In any case if NFS
has trouble running when it has less than 90% of the interior resource,
one might have to question which set of packets should be defined as the
DOS.

Tony








RE: Netmeeting - NAT issue

2002-03-21 Thread J. Noel Chiappa

 From: Tony Hain [EMAIL PROTECTED]

 it may be more convenient to have the border deal with DOS, but is it
 *required* as Noel asserted?

First, there's good idea, required, and *required*. It's *required*
that your computer have a test-and-branch instruction to be a Turing machine.
It's not *required* that it have a jump instruction, but all computers do -
so it're pretty much required in a machine architecture. An example of good
idea would be lots of registers - most machines have that, but not all. Etc,
etc.

I never said it was *required* to have some security functions at the border
- merely that it was likely to happen for a variety of reasons (e.g. policy
enforcement in a large organization). I think my meaning was somewhere
between good idea and required (as defined above) - I don't know exactly
where.


Second, when I made my statement about security alone demands that we be
able to move some functionality to a 'site border router', or some such, I
was speaking of security stuff in general, not DoS protection in particular.

I think there are different kinds of DoS attacks (I actually created a
taxonomy of DoS attacks for a research effort I'm involved with), and I expect
that different DoS attacks will need different mechanisms to handle them (i.e.
in the most efficient and robust manner - bearing in mind the old adage that
and engineer is someone who can do for $1 what any fool can do for $5). I
suspect that some might be at the borders, some might be at the servers - but
that's my intuition.

But DoS is still a very limited corner of security.

Noel




Moving Towards UTF8 vs ASCII(ACE) Forever

2002-03-21 Thread Edmon Chung

An underlying question we must ask ourselves from all the discussions that
have sprung up every now and then is:
Do we wish to

1. eventually move the DNS towards UTF8/16 OR
2. do we want to stay with ASCII(ACE) for the rest of our lives?

If the answer is 1. then the IDN solution should take it into concern.  If
it is 2. then perhaps the IETF should revise some of its
internationalization policies.

I can see an arguement coming that says there is no draft on this, but if we
dont even get this fundamental question figured out, how can there be a
collective or meaningful effort towards a solution that makes most sense??

Edmon

PS. Perhaps we should have a show of hands or something about the real
DIRECTION that PEOPLE want to see with the internationalization of the
Domain Names.




Re: [idn] Re: I don't want to be facing 8-bit bugs in 2013

2002-03-21 Thread Mark Davis

 Unicode is not usable in international context.
...

It would not be worth replying to these threadworn and repeated
assertions by Mr. Ohta, except that some members of this list may not
be that familiar with Unicode. Clearly Unicode is being used
successfully in a huge variety of products in international contexts.

For more information on the CJK repertoire (which is the part Mr. Ohta
objects to), see http://www.unicode.org/unicode/faq/han_cjk.html.

Mark
—

Γνῶθι σαυτόν — Θαλῆς
[For transliteration, see http://oss.software.ibm.com/cgi-bin/icu/tr]

http://www.macchiato.com

- Original Message -
From: Masataka Ohta [EMAIL PROTECTED]
To: Robert Elz [EMAIL PROTECTED]
Cc: [EMAIL PROTECTED]; [EMAIL PROTECTED]; [EMAIL PROTECTED];
[EMAIL PROTECTED]
Sent: Wednesday, March 20, 2002 07:58
Subject: [idn] Re: I don't want to be facing 8-bit bugs in 2013


 Kre;

| IDNA does _not_ work, because Unicode does not work in
International
| context.
 
  This argument is bogus, and always has been.   If (and where)
unicode
  is defective, the right thing to do is to fix unicode.

 Unicode is not usable in international context.

 There is no unicode implementaion work in international context.

 Unicode is usable in some local context.

 There is some unicode implementaion work in local contexts.

 However, the context information must be supplied out of band.

 And, the out of band information is equivalent to charset
 information, regardless of whether you call it charset or
 not.

  So, stop arguing against unicode (10646) - just fix any problems
it has.

 Fix is to supply context information out of band to specify which
 Unicode-based local character set to use.

 With MIME, it is doable by using different charset names for
 different local character set.

 See, for example, RFC1815.

 As for IDN, it can't just say use charset of utf-7 or use charset
 of utf-8.

 IDN can say for Japanese, use charset of utf-7-japanese.

 Or, if you insist not to distinguish different local character sets
by
 MIME charset names, IDN can say use charset of utf-7, but, for
 Japanese, use Japan local version of utf-7 and somehow specify
 how a name is used for Japanese.

 Anyway, with the fix, there is no reason to prefer Unicode-based
 local character sets, which is not widely used today, than existing
 local character sets already used world wide.

 Masataka Ohta







Re: Unicode is so flawed that 7 or 8 bit encoding is not an issue

2002-03-21 Thread James Seng
[Note: IDN WG list removed]

 As you could have seen, on IETF mailing list, Harald and I have, at
 least, agreed that, if you use unicode based encoding, local context
 (or locale) must be carried out of band

Few will disagree (including me) that using Unicode to do localization is
almost impossible without locale context.

But I am touching on a sensitive topic about what you considered as I18N 
L10N and what the rest of the world thinks.

 But, anyway, the discussion so far in IETF list is enough to deny IDN.

I am not aware there was a IETF wide last call on IDN yet.

 I have found a theory to deny PKI and to explain why public key
 cryptograpy is not so polular dispite the efforts of ISO, which will
 destory entire business of a company or companies owing several
 important TLDs.

Very interesting. Could you share the theory, or privately if you prefer?

ps: btw, what effort of ISO are you referring?

 So, don't bother to say that there are so many so-called-international-
 but-actuallly-local domain names registered.

Huh? Since when this was ever a factor in IETF consideration?

-James Seng


Re: [idn] WG last call summary

2002-03-21 Thread tedd

On Mar 19, D. J. Bernstein [EMAIL PROTECTED] wrote:

  Go sell a Greek user an ``internationalized domain name'' with a delta,
  Pete. Then tell him that most of his correspondents will see the delta
  as incomprehensible gobbledygook rather than a delta. See what he says.

Okay, I'm not Greek, but as the registered owner of .com, I'll tell you*.

I expect that some windoze users will display it (.com) as a box dot
com; mac users will display it as a delta dot com; and everyone else
will display it as a bq--eida.com (RACE), or dq--uuyg.com (DUDE), or
zq---h9g.com (AMZ_C), or punycode, or whatever -- AND, they will
continue to do so until they have the appropriate software on their
machines to handle the conversion. That's what I expect.

On the other hand, I do not expect that the IDN will eventually
produce something that will transparently produce the appropriate
glyphs on millions of users machines. In my mind, that's simply not
possible because I think it requires that every machine on this
planet have a complete copy of UNICODE installed and available.

Granted, I am clueless about a lot of this IDN stuff and I may be
hubris (as Mr. Paul Hoffman once informed me) in my offering my
opinion to the Gods of the Internet, but I do think I see the
distinction between which side of the modem one is on and what
responsibility lies in each camp.

OK, scenario 1:

You tell him that although it's gobbledygook to people without greek
alphabet support, it will still work. It's not convenient, but it WILL work.
Guaranteed. For his business colleagues and friends in Greece, who DO have
the latest and greatest software, it will display as a delta. His ISP hasn't
had to upgrade, and everybody in the world can use his domain - eventually
they will see it as a delta as well, but for now they see it as an encoded
string they can still use no problem.
-snip
If you were IT director of a large firm, and you had a choice as to which to
roll out, which would you choose?

--
Paul Robinson

I would vote for scenario 1. Besides, IMHO (in my hubris opinion)
that's the only way it's going to work.

tedd

* The Greek version of delta registers a internal error -- it's the
Math language set that provides an acceptable delta.
-- 
http://sperling.com




Re: I don't want to be facing 8-bit bugs in 2013

2002-03-21 Thread Robert Elz

Date:Thu, 21 Mar 2002 00:57:18 +0859 ()
From:Masataka Ohta [EMAIL PROTECTED]
Message-ID:  [EMAIL PROTECTED]

Otha-san

  | Anyway, with the fix, there is no reason to prefer Unicode-based
  | local character sets, which is not widely used today, than existing
  | local character sets already used world wide.

Let's assume that local char sets, and an explicit indication of which
to use is adequate for this purpose (as Harald has said, and I agree,
that's not sufficient for all purposes, but for stuff like domain names,
file names, etc, I suspect it is).

Then, let's take all the MIME names of those charsets, and number them,
0, 1, 2, 3, ... (as many as it takes), that's a 1::1 mapping, after all
the mime charset names are in ascii, we wouldn't want to keep that.

Then, rather than labelling whole names with a charset, we'll label
every individual character, so if, by some strange chance, ascii (or 8859-1)
happened to be allocated number 0, then 'A' would be 0-65.
This way we can mix and match names with characters from any random
character set (it may be a little less efficient than one label per name
but that's OK, assume we'll pay that price for the flexibility).

No, we'll just list all the possible characters in all the char sets,
with their char set numbers attached, in one big table.

What we have at that point is (essentially) 10646 - unicode.   Just up
above you said this works (throwing in some redundant labelling cannot
cause it to stop working, nor can altering the labels from ascii to binary
numerals).

Of course, a large fraction of the char sets that we just listed in a big
table contain ascii as a subset,so we end up with dozens of versions of
A (it's in 8859-1 8859-2 ... tis-620 (Thai) ...).  All of those A's are
in all respects identical, keeping them will do no more than cause
interoperability problems.   So let's go through and squash all the duplicates.
and then renumber everything to be a bit more compact (the renumbering is
a 1::1 operation that alters nothing important).

Having done that, we have exactly 10646 (or can have, if we picked the
right character sets for our initial big list of everything, gave them
the appropriate char set numbers, and compressed the number spaces in
just the right way).

Again, you said above, this works.

The only place in all of this where there's any possibility of a problem is
in the squash all the duplicates - if some characters that are duplicates
aren't removed, or some that aren't are treated as if they are.  If this
happened, then there'd be a bug in the actual character tables (too many, or
too few) - but this doesn't alter the principle of the thing.

If there is a bug like this (which I am not able to judge) then someone
should get it fixed.

Whether or not there is a bug, the unicode/10646 approach is clearly the most
flexible way, and is perfectly adequate for, labelling things using
whatever characters anyone wants to use - internationally or locally.

There is simply no way to rationally claim that a local char set, plus label,
is adequate and unicode is not.

kre






Looking ahead to IETF 54

2002-03-21 Thread Ole J. Jacobsen


As a frequent visitor to Japan, I am planning to put together some kind of
guide and put it on a web page somewhere, more on that later.

Meanwhile, please mark your calendar for:

1. Akihabara (electronics town) Geek Tour, Sunday July 14. On Sundays, the
Main Drag aka Chuo Dori is closed to cars and Akihabara will be full of
people shopping for the latest gadgets. Meet at entrance to Landmark
shopping center at 11am. We will return before 5pm in time for the welcome
reception.

2. Minato Mirai Hall Fisk Pipe Organ Demo. Monday, July 15, 6pm. Details
to follow.

http://www.city.yokohama.jp/me/mmhall/about/pipe-e.html

When you plan your travel, consider using Osaka (KIX) as you entry point
instead of Tokyo. If you do, you should stop in Kobe or Osaka and visit
the world's longest suspension bridge:

http://www.hsba.go.jp/bridge/e-akasi.htm

You'll also have easy access to Kyoto.

Then take the bullet train to Yokohama. Get a Nozomi if you can, they
are the fastest.

See you in Yokohama!

Ole

Ole J. Jacobsen
Editor and Publisher
The Internet Protocol Journal
Office of the CTO, Cisco Systems
Tel: +1 408-527-8972
GSM: +1 415-370-4628
E-mail: [EMAIL PROTECTED]
URL: http://www.cisco.com/ipj






Re: Moving Towards UTF8 vs ASCII(ACE) Forever

2002-03-21 Thread Keith Moore

 Do we wish to
 
 1. eventually move the DNS towards UTF8/16 OR
 2. do we want to stay with ASCII(ACE) for the rest of our lives?

it's a false dichotomy.  

first, there are probably some purposes for which identifiers 
(including DNS names) should stay ASCII (and not even 10646
encoded as ASCII) for the forseeable purpose.  there are other
purposes for which IDNs are useful.

second, ASCII for the rest of our lives is a mischaracterization.
IDNA allows applications to accept and present IDNs in native
form, without requiring all applications and infrastructure to
upgrade before IDNs can be used.   IDNA is designed to maximize
the rate at which IDNs can be deployed.

users don't care whether IDN queries are encoded on the wire.
they only care about whether IDNs work for them.  IDNA puts the
user in a position to determine whether IDNs work for him -
a UTF-8 everywhere approache forces him to wait for everybody 
else to adopt IDNs before he can use them.

Keith




Re: [idn] Moving Towards UTF8 vs ASCII(ACE) Forever

2002-03-21 Thread John Stracke

 Provably false: well-coded applications know the limitations of domain
 names, and do not even attempt to make requests for non-ASCII names.

First of all, I disagree with the well-coded part because I believe a
well-coded application will do the dns request as is and allow the dns
response to determine the validity of a domain name.

Nope.  You can almost certainly give your user a better error message if 
you validate the input yourself.

For that matter, a well-designed application will not even make it 
possible to enter anything but ASCII in an input field (whatever) for a 
domain name.  Basic rule of usability: making user mistakes impossible is 
better than catching them after the fact.

Secondly, if this is
the case, then the plugin should be intercepting the input to the app and
not the output.  It is still a possibility.

No, because that input is highly application-specific.

/\
|John Stracke|Principal Engineer |
|[EMAIL PROTECTED]   |Incentive Systems, Inc.|
|http://www.incentivesystems.com |My opinions are my own.|
||
|Diplomacy: The art of letting someone else have your way|
\/




Re: [idn] Moving Towards UTF8 vs ASCII(ACE) Forever

2002-03-21 Thread D. J. Bernstein

John Stracke writes:
 For that matter, a well-designed application will not even make it 
 possible to enter anything but ASCII in an input field (whatever) for a 
 domain name.

That's incredibly bad design.

You're violating the basic principles of information hiding articulated
by Parnas in the early 1970s. Instead of isolating the name-existence
decision in one place (the DNS server's database), you're spreading the
decision across a huge number of programs on a huge number of machines.
Changing a decision becomes extremely expensive.

We're seeing the economic consequences of this in BIND's res library.
That library is the most widely used UNIX DNS-lookup mechanism, and is
one of the largest sources of UTF-8 failures; see http://pi.cr.yp.to. It
has to be fixed on a huge number of machines.

(Note that this cost, together with the other costs of making UTF-8 IDNs
work, is only a tiny fraction of the costs of making IDNA work.)

If you think that the 8-bit problems in res are an example of people
agreeing with your design ideas, you're mistaken. The change was made in
a panic in 1996, when CERT announced that several programs had security
flaws based on careless use of DNS PTR results. Of course, anyone who
thinks about the problem for a moment can see that unusual DNS A _input_
has no relevance to the security issue, but people in a security-related
panic often don't stop to think about the damage they're causing.

 Basic rule of usability: making user mistakes impossible is 
 better than catching them after the fact.

You obviously aren't achieving that goal. You're catching typos if and
only if they involve non-ASCII characters. What about ASCII typos? What
is the basis for your assertion that the no-such-name message should, as
a matter of UI design, be different between these two situations?

Even worse, what about ASCII typos that produce valid domain names?

Basic rule of usability: Have the computer copy the data so that the
user doesn't have an opportunity to make a mistake. Saves time, too.

---D. J. Bernstein, Associate Professor, Department of Mathematics,
Statistics, and Computer Science, University of Illinois at Chicago




7 bits forever!

2002-03-21 Thread D. J. Bernstein

Keith Moore writes:
 IDNA is designed to maximize the rate at which IDNs can be deployed.

I hereby declare that cs.utk.edu is an IDN. We will now spend twenty
years trying to convince all application programs to display it as the
international picture of an ostrich with his head in the sand.

Look, Keith, we've deployed an IDN in zero time! Infinite rate! You
didn't even have to touch your DNS records!

Boy, I'm glad that, when you faced the much smaller problem of non-ASCII
subject lines back in 1991, you and your buddies decided to ``maximize
the rate'' of deployment by inventing your own encoding mechanisms,
rather than giving in to the demands of 8-bit transparency. Oh, sure, we
still have quite a few display failures eleven years later, and users in
Europe are constantly sending raw ISO 8859-1 in violation of your 7-bit
requirement, and sendmail still has problems with UTF-8, and now we're
facing years of IDNA pain followed by years of IDNA-MAILBOX pain, but
you were able to _instantaneously_ deploy Quoted-Printable back in 1991,
and that's what matters! Brilliant!

Even better, your incisive design decisions mean that billions of people
will be able to continue using their 36-bit computers, which, as you
know, are incapable of processing character data in units other than 7
bits, because 36 is not divisible by 8, but it is divisible by 7, um,
well, except for the remainder of 1, but who's counting? The point is,
Keith, THANK YOU! You've made all our lives so much better! I hope
someday you can travel to Taiwan and meet some of your fans.

---D. J. Bernstein, Associate Professor, Department of Mathematics,
Statistics, and Computer Science, University of Illinois at Chicago




Re: Netmeeting - NAT issue

2002-03-21 Thread james woodyatt

On Thursday, March 21, 2002, at 06:15 PM, [EMAIL PROTECTED] wrote:
 Of course, there is the possibility that if they were totally honest,
 and marketed their devices as Enabling appliances for selected Internet
 services that they'd STILL make money (and then you'd have no one to
 blame).

Please read the warning below and keep NAT products away from kids and 
children who are not old enough to understand the risks of network 
address translation:

INTERNET ARCHITECTURE BOARD WARNING: Using Network Address
Translators Causes Loss of Routing Transparency and may Complicate
Application Protocol Interoperability in the Internet.

Thank you!


--
j h woodyatt [EMAIL PROTECTED]
please network responsibly.