Hi Barry. Thanks for your response. I believe we share the same goals here
(the "what"). Where I think we need to focus the discussion then is on the
mechanisms to achieve that (the "how"). Let me fill in the details, as I see
them, by responding to a few things you wrote:
If you do this by profiling, we need to get to a point where two things are
true: (1) the profile chain has ended, and what's left is well defined, and (2)
I have to be able to, *from the specs and the protocol alone*, determine what
profile will be used. I don't see how this protocol gives me any way to
determine what to send or what to expect to receive. And if an out-of-band
understanding is required for that, that doesn't interoperate.
I completely agree with you that the profile chain has to end, with what's left
being well-defined and that from the protocol specs alone, one can determine
how to interoperate.
Now, the way we usually handle the need for "different kinds of things" is to
have different fields for each, or to have one field tagged so that it's
self-defining (as URIs have a scheme that says what to do with them). If the
Audience field might look like this in one application:
urn:ietf:params:banana
...and like this in another
abra:cadabra
...where the first is understood to be a URI and the second is something else,
then please explain how you and I can each write applications to that?
You can write applications to that by having the profile chain end, and with
the contents of the Audience field being completely specified somewhere in the
profile chain being used. Also, I'll observe that we are using the "tagged
field" approach that you mention in the assertions specs, using the defined tag
values urn:ietf:params:oauth:grant-type:saml2-bearer,
urn:ietf:params:oauth:client-assertion-type:saml2-bearer,
urn:ietf:params:oauth:grant-type:jwt-bearer, and
urn:ietf:params:oauth:client-assertion-type:jwt-bearer to declare the token
type and use of that token type. (The OpenID Connect profile also uses a
"tagged field" which is an OAuth scope value of "openid" to dynamically declare
to the OAuth implementation that the OpenID Connect profile is being used.
Other profiles may similarly indicate their usage through different scope
values.)
I am proposing that there must be a way for someone writing an application to
know what to use in these fields that will work with your application (or will
work with a server in the same way as your application does, or whatever)
*without* having to go to the documentation of your application to figure it
out.
I agree with what I think you mean, but possibly not with how you're saying it.
Using the TCP analogy again, in fact, to understand the contents of the TCP
stream for port 25, one has to go to the documentation of the application
communicating on port 25. In this case, that documentation is RFC 821 and its
successors. SMTP is a profile of TCP that further specifies the contents of
the data being exchanged. An analogous situation exists when using OAuth
Assertions.
I'll also observe that the working group is also working on a specification
that enables an OAuth client to dynamically register itself with the
Authorization Server (draft-ietf-oauth-dyn-reg) and that that registration does
declare information about what profile is being used, as John Bradley pointed
out in his response. That's a key piece of the whole solution to enable
interoperable implementations.
So using your Ukrainian dolls analogy, yes, the OAuth Assertions spec and the
OAuth SAML2 Profile and OAuth JWT Profile specs are dolls inside other dolls -
not the outer doll. That's by design, and not a spec defect, at least as I see
it. We already do have mechanisms for dynamically declaring what profile is
being used, and we are using them.
I agree with Stephen that we should let this conversation run for a while to
make sure everyone comes to a common understanding, but ultimately, I hope that
you'll withdraw your DISCUSS, because, in fact, interoperable implementations
can be written by reading the specs used alone.
Best wishes,
-- Mike
-----Original Message-----
From: [email protected] [mailto:[email protected]] On Behalf Of Barry
Leiba
Sent: Sunday, February 17, 2013 9:38 PM
To: Mike Jones
Cc: Stephen Farrell; [email protected]; [email protected]
Subject: Re: [OAUTH-WG] oauth assertions plan
OK, I have some time to respond to this on a real computer.
Let's look at the general mechanism that oauth provides, using one use case:
A client asks an authorization server for authorization to do something.
The authorization server responds with an authorization token, which the client
is required to understand.
We have talked about three kinds of tokens: bearer tokens, MAC tokens, and
assertion tokens.
How does a client know what kind of token it will get from a particular
authorization server?
How does a server that supports multiple token types know what kind of token it
should give to a particular client?
Now, suppose that the server decides to give back an assertion:
How does the server know whether to give a SAML assertion or a JWT assertion?
How does the client know which it's going to get?
And now you're saying that even if everyone knows they're going to get an
assertion token, and specifically a JWT assertion, the semantics of particular
fields in those tokens are undefined. If you have different meanings
("different kinds of things", as you've said) for the Audience field, how is
the server supposed to communicate which meaning *it* is using? How is there
any assurance that a client will understand it in the same way?
These are the sorts of things I'm concerned about, and this is what I mean by
saying that it's like a Ukrainian doll: you open the oauth doll and find the
token doll; you open the token doll and find the assertions doll; you open the
assertions doll and find the JWT doll; you open the JWT doll and find the
Audience field doll... at what point do the dolls end and interoperablity
begins?
I understand that you want certain things to be handled differently by
different applications, and I'm fine with that in principle. But the other
part of the principle is that I have to be able to write an application that
interoperates with yours *purely by reading the specs*.
If you do this by profiling, we need to get to a point where two things are
true: (1) the profile chain has ended, and what's left is well defined, and (2)
I have to be able to, *from the specs and the protocol alone*, determine what
profile will be used. I don't see how this protocol gives me any way to
determine what to send or what to expect to receive. And if an out-of-band
understanding is required for that, that doesn't interoperate.
Now, the way we usually handle the need for "different kinds of things" is to
have different fields for each, or to have one field tagged so that it's
self-defining (as URIs have a scheme that says what to do with them). If the
Audience field might look like this in one application:
urn:ietf:params:banana
...and like this in another
abra:cadabra
...where the first is understood to be a URI and the second is something else,
then please explain how you and I can each write applications to that?
For your specific questions:
> Barry, are you proposing that we require that the Audience contain a
> specific data format tailored to a particular application of
> assertions? If so, what format are you proposing, and for which
> application of assertions? Likewise, are you proposing that the
> Subject field contain a particular data format tailored to a particular
> application? And also the Issuer field?
I am proposing that there must be a way for someone writing an application to
know what to use in these fields that will work with your application (or will
work with a server in the same way as your application does, or whatever)
*without* having to go to the documentation of your application to figure it
out.
That's why I say that as I see it, it's not an issue of MTI. I'm not saying
that I want you to require that any particular thing be implemented. I'm
saying that both sides need to be able to know which variations *are*
implemented. How else do you get interoperability?
Your TCP/ports example is a perfect one to show how this works: the assignments
for port numbers are *exactly* to create the interoperability I'm talking about
here. TCP doesn't say anything about the protocol ("profile", perhaps) that's
used over it. But we've defined that port 110 is used for POP and 143 is used
for IMAP and 25 is used for SMTP, and so on. So I know that if my IMAP
application wants to talk with your IMAP server, I can accomplish that by using
port 143. If you decide to run your IMAP server on port 142 instead, or if you
run your POP server on port 143, we will not interoperate.
But all I need to know is (1) how to do IMAP, (2) how to do it over TCP, and
(3) what port to use... and I can build an IMAP client that will work with any
IMAP server that follows the same specs.
Can you do that here? Please explain how.
Barry
_______________________________________________
OAuth mailing list
[email protected]
https://www.ietf.org/mailman/listinfo/oauth