Rfc | 1636 |
Title | Report of IAB Workshop on Security in the Internet Architecture -
February 8-10, 1994 |
Author | R. Braden, D. Clark, S. Crocker, C. Huitema |
Date | June 1994 |
Format: | TXT, HTML |
Status: | INFORMATIONAL |
|
Network Working Group R. Braden
Request for Comments: 1636 ISI
Category: Informational D. Clark
MIT Laboratory for Computer Science
S. Crocker
Trusted Information Systems, Inc.
C. Huitema
INRIA, IAB Chair
June 1994
Report of IAB Workshop on
Security in the Internet Architecture
February 8-10, 1994
Status of this Memo
This memo provides information for the Internet community. This memo
does not specify an Internet standard of any kind. Distribution of
this memo is unlimited.
Abstract
This document is a report on an Internet architecture workshop,
initiated by the IAB and held at USC Information Sciences Institute
on February 8-10, 1994. This workshop generally focused on security
issues in the Internet architecture.
This document should be regarded as a set of working notes containing
ideas about security that were developed by Internet experts in a
broad spectrum of areas, including routing, mobility, realtime
service, and provider requirements, as well as security. It contains
some significant diversity of opinions on some important issues.
This memo is offered as one input in the process of developing viable
security mechanisms and procedures for the Internet.
Table of Contents
1. INTRODUCTION .................................................. 2
2. OVERVIEW ...................................................... 4
2.1 Strategic and Political Issues ........................... 4
2.2 Security Issues .......................................... 4
2.3 DNS Names for Certificates ............................... 7
3. FIREWALL ARCHITECTURE ......................................... 9
3.1 Introduction ............................................. 9
3.2 Application-Layer Firewalls .............................. 11
3.3 IP-Layer Firewalls ....................................... 12
4. SECURE QOS FORWARDING ......................................... 21
4.1 The Requirement for Setup ................................ 21
4.2 Securing the Setup Process. .............................. 22
4.3 Validating an LLID ....................................... 24
4.4 Dynamics of Setup ........................................ 28
4.5 Receiver-Initiated Setup ................................. 30
4.6 Other Issues ............................................. 30
5. AN AUTHENTICATION SERVICE ..................................... 35
5.1 Names and Credentials .................................... 36
5.2 Identity-Based Authorization ............................. 37
5.3 Choosing Credentials ..................................... 38
6. OTHER ISSUES .................................................. 39
6.1 Privacy and Authentication of Multicast Groups ........... 39
6.2 Secure Plug-and-Play a Must .............................. 41
6.3 A Short-Term Confidentiality Mechanism ................... 42
7. CONCLUSIONS ................................................... 44
7.1 Suggested Short-Term Actions ............................. 44
7.2 Suggested Medium-Term Actions ............................ 46
7.3 Suggested Long-Term Actions .............................. 46
APPENDIX A -- Workshop Organization .............................. 48
Security Considerations .......................................... 52
Authors' Addresses ............................................... 52
1. INTRODUCTION
The Internet Architecture Board (IAB) holds occasional workshops
designed to consider long-term issues and strategies for the
Internet, and to suggest future directions for the Internet
architecture. This long-term planning function of the IAB is
complementary to the ongoing engineering efforts performed by working
groups of the Internet Engineering Task Force (IETF), under the
leadership of the Internet Engineering Steering Group (IESG) and area
directorates.
An IAB-initiated workshop on the role of security in the Internet
Architecture was held on February 8-10, 1994 at the Information
Sciences Institute of the University of Southern California, in
Marina del Rey, California. This RFC reports the results of the
workshop.
In addition to the IAB members, attendees at this meeting included
the IESG Area Directors for the relevant areas (Internet, Transport,
Security, and IPng) and a group of 15 other experts in the following
areas: IPng, routing, mobility, realtime service, and security (see
Appendix for a list of attendees). The IAB explicitly tried to
balance the number of attendees from each area of expertise.
Logistics limited the attendance to about 30, which unfortunately
meant that many highly qualified experts were omitted from the
invitation list.
In summary, the objectives of this workshop were (1) to explore the
interconnections between security and the rest of the Internet
architecture, and (2) to develop recommendations for the Internet
community on future directions with respect to security. These
objectives arose from a conviction in the IAB that the two most
important problem areas for the Internet architecture are scaling and
security. While the scaling problems have led to a flood of
activities on IPng, there has been less effort devoted to security.
Although some came to the workshop eager to discuss short-term
security issues in the Internet, the workshop program was designed to
focus more on long-term issues and broad principles. Thus, the
meeting began with the following ground rule: valid topics of
discussion should involve both security and at least one from the
list: (a) routing (unicast and multicast), (b) mobility, and (c)
realtime service. As a basis for initial discussion, the invitees
met via email to generate a set of scenarios (see Appendix)
satisfying this ground rule.
The 30 attendees were divided into three "breakout" groups, with each
group including experts in all the areas. The meeting was then
structured as plenary meetings alternating with parallel breakout
group sessions (see the agenda in Appendix). On the third day, the
groups produced text summarizing the results of their discussions.
This memo is composed of that text, somewhat rearranged and edited
into a single document.
The meeting process determined the character of this document. It
should be regarded as a set of working notes produced by mostly-
autonomous groups, containing some diversity of opinions as well as
duplication of ideas. It is not the output of the "security
community", but instead represents ideas about security developed by
a broad spectrum of Internet experts. It is offered as a step in a
process of developing viable security mechanisms and procedures for
the Internet.
2. OVERVIEW
2.1 Strategic and Political Issues
Despite the workshop emphasis on architectural issues, there was
considerable discussion of the real-politik of security.
For a number of years, the IETF, with IAB backing, has worked on
developing PEM, which provides email security with a great deal of
functionality. A question was repeatedly raised at the workshop:
why has user acceptance of PEM been slow? A number of answers to
this question were suggested.
(a) High-quality implementations have been slow in coming.
(b) The use of a patented technology, the RSA algorithm, violates
social conventions of the Internet.
(c) Export restrictions dampen vendor enthusiasm.
(d) PEM currently depends upon a certificate hierarchy for its
names, and certificates form a new and complex name space.
There is no organizational infrastructure in place for creat-
ing and managing this name space.
(e) There is no directory infrastructure available for looking up
certificates.
The decision to use X.500 has been a complete failure, due to
the slow deployment of X.500 in the Internet. Because of UDP
packet size restrictions, it is not currently feasible to
store certificates in the DNS, even if the DNS were expanded
to hold records for individual email users.
It seems probable that more than one, and possibly all, of these
reasons are at work to discourage PEM adoption.
The baleful comment about eating: "Everything I enjoy is either
immoral, illegal, or fattening" seems to apply to the cryptography
technology that is required for Internet security.
2.2 Security Issues
Almost everyone agrees that the Internet needs more and better
security. However, that may mean different things to different
people. Four top-level requirements for Internet security were
identified: end-to-end security, end-system security, secure QOS,
and secure network infrastructure.
A. End-to-End Security
One requirement is to support confidentiality, authentication
and integrity for end-to-end communications. These security
services are best provided on an end-to-end basis, in order
to minimize the number of network components that users must
trust. Here the "end" may be the end system itself, or a
proxy (e.g., a firewall) acting on behalf of an end system.
For point-to-point applications, the workshop felt that
existing security techniques are well suited to support
confidentiality, authentication and integrity services
efficiently. These existing techniques include symmetric
encryption applied on an end-to-end basis, message digest
functions, and key management algorithms. Current work in
these areas in the IETF include the PEM and Common
Authentication Technologies working groups.
The group favored a strategic direction for coping with
export restrictions: separate authentication from privacy
(i.e., confidentiality). This will allow work to proceed on
authentication for the Internet, despite government
restrictions on export of privacy technology. Conversely, it
will allow easy deployment of privacy without authentication,
where this is appropriate.
The workshop explored the implications of multicasting for
end-to-end security. Some of the unicast security techniques
can be applied directly to multicast applications, while
others must be modified. Section 6.2 contains the results of
these discussions; in summary, the conclusions were:
a) Existing technology is adequate to support
confidentiality, authentication, and integrity at the
level of an entire multicast group. Supporting
authentication and integrity at the level of an
individual multicast source is performance-limited and
will require technology advances.
b) End-to-end controls should be based on end system or
user identifiers, not low level identifiers or locator
information. This requirement should spawn engineering
work which consists of applying known key distribution
and cryptographic techniques.
B. End-System Security
Every host has its own security defenses, but the strength of
these defenses depends upon the care that is taken in
administering them. Careful host security administration
means plugging security holes in the kernel and applications
as well as enforcing discipline on users to set good (hard to
crack) passwords.
Good security administration is labor-intensive, and
therefore organizations often find it difficult to maintain
the security of a large number of internal machines. To
protect their machines from outside subversion, organizations
often erect an outer security wall or "perimeter". Machines
inside the perimeter communicate with the rest of the
Internet only through a small set of carefully managed
machines called "firewalls". Firewalls may operate at the
application layer, in which case they are application relays,
or at the IP layer, in which case they are firewall routers.
The workshop spent considerable time on the architecture of
firewall routers. The results are contained in Section 3.
C. Secure QOS
The Internet is being extended to provide quality-of-service
capabilities; this is the topic called "realtime service" in
the workshop. These extensions raise a new set of security
issues for the architecture, to assure that users are not
allowed to attach to resources they are not authorized to
use, both to prevent theft of resources and to prevent denial
of service due to unauthorized traffic. The resources to be
protected include link shares, service classes or queues,
multicast trees, and so on. These resources are used as
virtual channels within the network, where each virtual
channel is intended to be used by a particular subset or
"class" of packets.
Secure QOS, i.e., protection against improper virtual channel
usage, is a form of access control mechanism. In general it
will be based on some form of state establishment (setup)
that defines authorized "classes". This setup may be done
via management configuration (typically in advance and for
aggregates of users), or it may be done dynamically via
control information in packets or special messages (typically
at the time of use by the source or receiver(s) of the
flow/data). In addition to state establishment, some form of
authentication will be needed to assure that successive
packets belong to the established class. The general case to
be solved is the multicast group, since in general the
multicast problem includes the two-party case as a subset.
The workshop developed an approach to the secure QOS problem,
which appears in Section 4 below.
D. Secure Network Infrastructure
Network operation depends upon the management and control
protocols used to configure and operate the network
infrastructure, including routers and DNS servers. An attack
on the network infrastructure may cause denial-of-service
from the user viewpoint, but from the network operators'
viewpoint, security from attack requires authentication and
integrity for network control and management messages.
Securing the routing protocols seems to be a straightforward
engineering task. The workshop concluded the following.
a) All routing information exchanges should be
authenticated between neighboring routers.
b) The sources of all route information should be
authenticated.
c) Although authenticating the authority of an injector of
route information is feasible, authentication of
operations on that routing information (e.g.,
aggregation) requires further consideration.
Securing router management protocols (e.g., SNMP, Telnet,
TFTP) is urgent, because of the currently active threats.
Fortunately, the design task should be a straightforward
application of existing authentication mechanisms.
Securing DNS is an important issue, but it did not receive
much attention at the workshop.
2.3 DNS Names for Certificates
As noted in Section 2.1, work on PEM has assumed the use of X.509
distinguished names as the basis for issuing certificates, with
public-key encryption. The most controversial discussion at the
workshop concerned the possibility of using DNS (i.e., domain)
names instead of X.509 distinguished names as (at least) an
interim basis for Internet security.
The argument in favor of DNS names is that they are simple and
well understood in the Internet world. It is easy for a computer
operating in the Internet to be identified this way, and users who
receive email on such machines already have DNS mailbox names. In
contrast, introducing X.509 distinguished names for security will
add a new layer of names. Most importantly, there is an existing
administrative model for assigning DNS names. There is no
administrative infrastructure for assigning X.509 distinguished
names, and generating them may be too complex for early
acceptance. The advocates of DNS names for certificates hope that
using DNS names would encourage the widespread use of security in
the Internet. It is expected that DNS names can be replaced later
by a more capable naming mechanism such as X.509-based
certificates.
The basic argument against DNS names as a basis for security is
that they are too "weak". Their use may lead to confusion in many
instances, and this confusion can only grow as more organizations
and individuals attach to the Internet. Some commercial email
systems employ numeric mailbox names, and in many organizations
there are uncertainties such as whether "bumber@foo.edu" belongs
to Bill Umber or Tom Bumber. While it is feasible to make DNS
names more descriptive, there is a concern that the existing
infrastructure, with millions of short, non-descriptive names,
will be an impediment to adoption of more descriptive names.
It was noted that the question of what name space to use for
certificates is independent of the problem of building an
infrastructure for retrieving those names. Because of UDP packet
size restrictions, it would not be feasible to store certificates
in the DNS without significant changes, even if the DNS were
expanded to hold records for individual email users.
The group was unable to reach a consensus on the issue of using
DNS names for security; further discussion in the Internet
community is needed.
3. FIREWALL ARCHITECTURE
3.1 Introduction
A firewall may be used to isolate a specific connected segment of
Internet topology. When such a segment has multiple links to the
rest of the Internet, coordinated firewall machines are required
on all the links.
Firewalls may be implemented at different layers in the protocol
stack. They are most commonly implemented at the application
layer by forwarding (application) gateways, or at the IP
(Internet) layer by filtering routers. Section 3.2 discusses
application gateways. Section 3.3 concerns Internet-layer
firewalls, which filter IP datagrams entering or leaving a
security perimeter.
The general architectural model for a firewall should separate
policy, i.e., determining whether or not the requester of a
service should be granted access to that service, from control,
i.e., limiting access to resources to those who have been granted
access.
3.1.1 The Use for Firewalls
Firewalls are a very emotional topic in the Internet community.
Some community members feel the firewall concept is very
powerful because firewalls aggregate security functions in a
single place, simplifying management, installation and
configuration. Others feel that firewalls are damaging for the
same reason: they provide "a hard, crunchy outside with a soft
chewy center", i.e., firewalls foster a false sense of
security, leading to lax security within the firewall
perimeter. They observe that much of the "computer crime" in
corporate environments is perpetrated by insiders, immune to
the perimeter defense strategy. Firewall advocates counter
that firewalls are important as an additional safeguard; they
should not be regarded as a substitute for careful security
management within the perimeter. Firewall detractors are also
concerned about the difficulty of using firewalls, requiring
multiple logins and other out-of-band mechanisms, and their
interference with the usability and vitality of the Internet.
However, firewalls are a fact of life in the Internet today.
They have been constructed for pragmatic reasons by
organizations interested in a higher level of security than may
be possible without them. This section will try to outline
some of the advantages and disadvantages of firewalls, and some
instances where they are useful.
Consider a large organization of thousands of hosts. If every
host is allowed to communicate directly with the outside world,
attackers will attempt to penetrate the organization by finding
the weakest host in the organization, breaching its defenses,
and then using the resources of that host to extend the
penetration further within the organization. In some sense,
firewalls are not so much a solution to a security problem as
they are a reaction to a more basic software
engineering/administration problem: configuring a large number
of host systems for good security. If this more basic problem
could be solved, firewalls would generally be unnecessary.
It is interesting to consider the effect that implementing a
firewall has upon various individuals in the organization.
Consider first the effect upon an organization's most secure
host. This host basically receives little or no extra
protection, because its own perimeter defenses are as strong or
stronger than the firewall. In addition, the firewall will
probably reduce the connectivity available to this host, as
well as the reliability of the communications path to the
outside world, resulting in inconvenience to the user(s) of
this host. From this (most secure) user's point of view, the
firewall is a loss.
On the other hand, a host with poor security can "hide" behind
the firewall. In exchange for a more limited ability to
communicate with the outside world, this host can benefit from
the higher level of security provided by the firewall, which is
assumed to be based upon the best security available in the
entire organization. If this host only wants to communicate
with other hosts inside the organization, the outside
communications limitations imposed by the firewall may not even
be noticed. From this host's viewpoint, better security has
been gained at little or no cost.
Finally, consider the point of view of the organization as a
whole. A firewall allows the extension of the best security in
the organization across the whole organization. This is a
benefit (except in the case where all host perimeter defenses
in the organization are equal). Centralized access control
also becomes possible, which may be either a benefit or a cost,
depending upon the organization. The "secure" hosts within the
organization may perceive a loss, while the "unsecure" hosts
receive a benefit. The cost/benefit ratio to the organization
as a whole thus depends upon the relative numbers of "secure"
and "unsecure" hosts in the organization.
Consider some cases where firewalls do not make sense. An
individual can be thought of as an organization of one host.
The security of all the host(s) is thus (trivially) identical,
and by definition the best available to the organization. In
this case the choice of firewall is simple. Does this
individual wish to communicate with the outside or not? If
not, then the "perfect" firewall is implemented (by complete
disconnection). If yes, then the host perimeter will be the
same as the firewall perimeter, so a firewall becomes
unnecessary.
Another interesting case is an organization that consists of
individuals with few shared interests. This might be the case
of a service provider that sells public access to the network.
An unrelated community of subscribers should probably be
considered as individuals, rather than an organization.
Firewalls for the whole organization may make little sense in
this case.
To summarize, the benefit of a firewall depends upon the nature
of the organization it protects. A firewall can be used to
extend the best available protection within the organization
across the entire organization, and thus be of benefit to large
organizations with large numbers of poorly administered hosts.
A firewall may produce little or no perceived benefit, however,
to the individuals within an organization who have strong host
perimeters already.
3.2 Application-Layer Firewalls
An application-layer firewall can be represented by the following
diagram.
C <---> F <---> S
Here the requesting client C opens its transport connection to the
firewall F rather than directly to the desired server S. One
mechanism for redirecting C's request to F's IP address rather
than S's could be based on the DNS. When C attempts to resolve
S's name, its DNS lookup would return a "service redirection"
record (analogous to an MX record) for S. The service redirection
record would return the IP address of F.
C enters some authentication conversation to identify itself to F,
and specifies its intention to request a specific service from S.
F then decides if C is authorized to invoke this service. If C is
authorized, F initiates a transport layer connection to S and
begins the operation, passing requests and responses between C and
S.
A major advantage of this scenario over an IP-layer firewall is
that raw IP datagrams are never passed through the firewall.
Because the firewall operates at the application layer, it has the
opportunity to handle and verify all data passing through it, and
it may be more secure against illicit rendezvous attacks (see
below).
Application layer firewalls also have important disadvantages.
For full benefit, an application level firewall must be coded
specifically for each application. This severely limits the
deployment of new applications. The firewall also represents a
new point of failure; if it ceases to be reachable, the
application fails. Application layer firewalls also may affect
performance more than IP-layer firewalls, depending on specific
mechanisms in use.
3.3 IP-Layer Firewalls
Our model of an IP-layer firewall is a multi-ported IP router that
applies a set of rules to each incoming IP datagram, to decide
whether it will be forwarded. It is said to "filter" IP
datagrams, based on information available in the packet headers.
A firewall router generally has a set of filtering rules, each of
which specifies a "packet profile" and an "action". The packet
profile specifies values for particular header fields, e.g.,
source and destination IP address, protocol number, and other
suitable source and destination identifying information (for
instance, port numbers). The set of possible information that may
be used to match packets is called an "association". The exact
nature of an association is an open issue.
The high-speed datagram forwarding path in the firewall processes
every arriving packet against all the packet profiles of all
active rules, and when a profile matches, it applies the
corresponding action. Typical actions may include forwarding,
dropping, sending a failure response, or logging for exception
tracking. There may be a default rule for use when no other rule
matches, which would probably specify a drop action.
In addition to the packet profile, some firewalls may also use
some cryptographic information to authenticate the packet, as
described below in section 3.3.2.
3.3.1 Policy Control Level
This section presents a model for the control of a firewall
router, with some examples of specific mechanisms that might be
used.
1. A client C attempts to access a service S. (Client here
can mean either a person or a process - that also is an
issue to be resolved.)
2. The initiation of access to that service may result in an
attempt to cross one or more boundaries of protection via
firewall router(s).
3. The policy control level sets filters in the firewall
router(s), to permit or deny that attempt.
The policy control level consists of two distinct functions,
authentication and authorization. Authentication is the
function of verifying the claimed identity of a user. The
authentication function should be distributed across the
Internet, so that a user in one organization can be
authenticated to another organization. Once a user is
authenticated, it is then the job of the authorization service
local to the resource being requested to determine if that user
is authorized to access that resource. If authorization is
granted, the filter in the firewall can be updated to permit
that access.
As an aid to understanding the issues, we introduce a
particular detailed mechanism. We emphasize that this
mechanism is intended only as an illustrative example; actual
engineering of the mechanism will no doubt lead to many
changes. Our mechanism is illustrated by the following sketch.
Here a user wishes to connect from a computer C behind firewall
F1, to a server S behind firewall F2. A1 is a particular
authentication server and Z1 is a particular authorization
server.
C <-------> F1 <-------> F2 <-------> S
\ /
\_____ /
\ \ /
A1 Z1
C attempts to initiate its conversation by sending an initial
packet to S. C uses a normal DNS lookup to resolve S's name,
and uses normal IP routing mechanisms. C's packet reaches
firewall router F1, which rejects the packet because it does
not match any acceptable packet profile. F1 returns an
"Authentication Required" error indication to C, including a
list of authentication/authorization servers that F1 trusts.
This indication might be a new type of ICMP Destination
Unreachable packet, or some other mechanism for communicating
with C.
When C receives the error indication, authenticates itself with
A1, one of the authentication servers listed in the error
indication, after validating A1's identity. C then requests
authorization from server Z1 (using a ticket provided by A1),
informs Z1 of the application it wishes to perform, and
provides a profile for the packets it wishes to pass through
F1. Z1 then performs an authorization function to decide
whether to allow C to penetrate F1. If C is to be allowed, Z1
then informs the firewall F1 to allow packets matching the
packet profile to pass through the firewall F1.
After C's packets penetrate F1, they may again be rejected by a
second firewall F2. C could perform the same procedures with
authentication server A2 and authorization server Z2, which F2
trusts. This is illustrated by the following schematic diagram
of the sequence of events.
----------+--------+--------+------------+------------+----
| C | A1 | Z1 | F1 | F2 | S
----------+--------+--------+------------+------------+----
| Sends pkt| | | | |
| to S ----------------------->Intercept;| |
| | | | requires | |
| | | |authenticat'n |
| <------------------------------- | |
|Auth'cate | | | | |
| C to A1 ----> | | | |
| |Provide | | | |
| <------- ticket| | | |
| Request | | | | |
|authoriz'n| | | | |
| -------------------> Is C| | |
| | |allowed?| | |
| | | OK ---------> | |
|Resend | | | Set filter | |
| first pkt| | | | |
| to S -------------------------->(OK)------>Intercept;|
| | | | | requires |
| | | | |authenticat'n
| <------------------------------------------- |
| (Repeat | | | | |
|procedure | | | | |
with A2,Z2)| | | | |
| ... | | | | |
|Resend | | | | |
| first pkt| | | | |
| ------------------------------>(OK)--------(OK)------>
| | | | | |
-----------+--------+--------+------------+------------+----
Again, we emphasize that this is only intended as a partial
sketch of one possible mechanism. It omits some significant
issues, including the possibility of asymmetric routes (see
3.3.3 below), and the possibility that the profiles may be
different in the two directions between C and S.
We could imagine generalizing this to an arbitrary sequence of
firewalls. However, security requires that each of the
firewalls be able to verify that data packets actually come
from C. This packet authentication problem, which is discussed
in the next section, could be extremely difficult if the data
must traverse more than one or possibly two firewalls in
sequence.
A firewall router may require re-authentication because:
* it has been added to the path by a routing change, or
* it has timed out the profile entry, or
* it has been newly re-activated, perhaps after a crash that
lost its list of acceptable profiles.
If C contacts authentication and authorization servers that S
trusts, C may utilize tickets given it by these servers when
initiating its use of S, and avoid re-authenticating itself to
S.
Although the authentication server A1 and the authorization
server Z1 are conceptually separate, they may run on the same
computer or router or even be separate aspects of a single
program. The protocol that C speaks to an An, the protocol
that C speaks to a Zn, and the protocol that Zn speaks to Fn
are not specified in these notes. The authentication mechanism
used with An and the packet profile required by a firewall Fn
are considered matters of policy.
3.3.2 Source Authentication
We next consider how to protect against spoofing the IP source
address, i.e., injecting packets that are alleged from come
from C but do not. There are three classes of mechanisms to
prevent such spoofing of IP-level firewalls. The mechanisms
outlined here are also discussed in Section 4.3 below.
o Packet Profile Only
The lowest level of security consists of allowing the IP-
layer firewall to filter packets purely on the basis of
the packet profile. This is essentially the approach used
by filtering routers today, with the addition of (1)
authentication and authorization servers to control the
filtering profiles, and (2) the automatic "Authentication
Required" notification mechanism. This approach provides
almost no security; it does not prevent other computers
from spoofing packets that appear to be transmitted by C,
or from taking over C's transport level connection to S.
o Sealed Packets
In the second level of security, each packet is "sealed"
with a secure hash algorithm. An authentication server Ai
chooses a secret and shares it with the source host S and
also with the authorization server Zi, which shares the
secret with the firewall Fi. Every packet that C
transmits contains a hash value that depends upon both the
contents of the packet and the secret value. The firewall
Fi can compute the same hash function and verify that the
packet was originated by a computer that knew the shared
secret.
This approach does raise issues of how much C trusts Zi
and Fi. Since they know C's secret, Zi or Fi could spoof
C. If C does not trust all Z's and F's in its path, a
stronger mechanism (see below) is needed.
A more difficult problem arises in authenticating C's
packets when more than one firewall lies in the path.
Carrying a separate seal for each firewall that is
penetrated would be costly in terms of packet size. On
the other hand, in order to use a single seal, all the
firewalls would have to cooperate, and this might require
a much more complex mechanism than the one sketched in the
previous section. Morever, it may require mutual trust
among all of the authentication servers Ai and
authorization servers Zi; any of these servers could
undermine all the others. Another possibility to be
investigated is to use hop-by-hop rather than end-to-end
authentication of C's packets. That is, each firewall
would substitute into the packet the hash needed by the
next firewall.
Multi-firewall source authentication is a difficult
problem that needs more investigation.
o Packet Signatures
In the third level of security, each packet is "signed"
using a public/private key algorithm. C shares its public
key with Zn, which shares it with Fn. In this scenario, C
can safely use one pair of keys for all authorization
servers and firewalls. No authorization server or
firewall can spoof C because they cannot sign packets
correctly.
Although packet signing gives a much higher level of
security, it requires public key algorithms that are
patented and currently very expensive to compute; their
time must be added to that for the hash algorithm. Also,
signing the hash generally makes it larger.
3.3.3 Other Firewall Issues
o Performance
An Internet-layer firewall has the advantage of generality
and flexibility. However, filtering introduces a
potential performance problem. Performance may depend
upon the number and position of the packet fields used for
filtering, and upon the number of rules against which a
packet has to be matched.
Denial of service attacks require that the per-packet rule
matching and the drop path be able to keep up with the
interface speed.
o Multicasting
To allow multicast traffic to penetrate a firewall, the
rule that is needed should be supplied by the receiver
rather than the sender. However, this will not work with
the challenge mechanism outlined in Section 3.3.1, since
"Authentication Required" notifications would be sent to
the sender, not to the receiver(s).
Multicast conversations may use any of the three levels of
security described in the previous section, but all
firewalls will have to share the same secret with the
originator of the data stream. That secret would have to
be provided to the receivers through other channels and
then passed to the firewalls at the receivers' initiative
(in much the same way that resources are reserved at
receiver's initiative in RSVP).
o Asymmetric Routing
Given a client computer C utilizing a service from another
computer C through a firewall F: if the packets returning
from S to C take a different route than packets from C to
S, they may encounter another firewall F' which has not
been authorized to pass packets from S to C (unlike F,
which has been). F' will challenge S rather than C, but S
may not have credentials to authenticate itself with a
server trusted by F'.
Fortunately, this asymmetric routing situation is not a
problem for the common case of single homed administrative
domains, where any asymmetric routes converge at the
firewall.
o Illicit Rendezvous
None of these mechanisms prevent two users on opposite
sides of a firewall from rendezvousing with a custom
application written over a protocol that may have been
authorized to run through a firewall.
For example, if an organization has a policy that certain
information is sensitive and must not be allowed outside
its premises, a firewall will not be enough to enforce
this policy if users are able to attach sensitive
information to mail and send it outside to arbitrary
parties. Similarly, a firewall will not prevent all
problems with incoming data. If users import programs and
execute them, the programs may have Trojan horses which
disclose sensitive information or modify or delete
important data. Executable code comes in many, many
forms, including PostScript files, scripts for various
interpreters, and even return addresses for sendmail. A
firewall can detect some of these and scan for some forms
of potentially hazardous code, but it cannot stop users
from transforming things that look like "data" into
programs.
We consider these problems to be somewhat outside the
scope of the firewall router mechanism. It is a matter of
the policies implemented by the organization owning the
firewalls to address these issues.
o Transparency for Security Packets
For the mechanisms described above to operate, the
"Authentication Required" notification and the
authentication/authorization protocol that is used between
the client computer and the authentication and
authorization servers trusted by a firewall, must be
passed by all firewalls automatically. This might be on
the basis of the packet profiles involved in security.
Alternatively, firewall routers might serve as
application-layer firewalls for these types of
communications. They could then validate the data they
pass to avoid spoofing or illicit rendezvous.
3.3.4 Firewall-Friendly Applications
Firewall routers have problems with certain communication
patterns where requests are initiated by the server, including
callbacks and multiple connections (e.g., FTP). It was
suggested that it would be useful to have guidelines to
application designers to help them to build 'firewall-friendly
applications'. The following guidelines were suggested:
1) no inbound calls (the xterm problem),
2) fixed port numbers (no portmapper or tcpmux),
3) integral redirection is good (application gateways),
4) no redirection in the protocol,
5) 32 bit sequence numbers that are crypto-strong random #'s,
and
6) fixed length and number of header fields.
Type fields are good, but they may not be needed if there are
fixed port numbers.
3.3.5 Conclusions
Compared to an application-layer firewall, an IP-layer firewall
scheme could provide a number of benefits:
- No extra authentication is required for end hosts.
- A single authentication protocol can be used for all
intended applications.
- An IP-layer firewall causes less performance degradation.
- An IP-layer firewall may be able to crash and recover
state without disturbing open TCP connections.
- Routes can shift without disturbing open TCP connections.
- There is no single point of failure.
- It is independent of application.
However, there are substantial difficult design issues to be
solved, particularly in the areas of multiple firewalls,
assymmetric routes, multicasting, and performance.
4. SECURE QOS FORWARDING
When the Internet supports special qualities-of-service (QOS) for
particular packet flows, there will be a new set of security
problems. There will be a need to authenticate and authorize users
asking for those QOS values that are expensive in network resources,
and it will be necessary to prevent theft of these resources and
denial-of-service attacks by others. This section contains a
conceptual model for these problems, which we may call secure QOS
forwarding. The issues here differ from end-to-end security and
firewalls, because QOS forwarding security may need to be enforced at
every router along a path.
It was noted that this is not a new problem; it was stated and solved
in a theoretical way in a thesis by Radia Perlman.
4.1 The Requirement for Setup
Setup is an essential part of any QOS mechanism. However, it may
be argued that there are also good engineering reasons for setup
in any Internet-layer security mechanism, even without QOS
support. In the abstract, one could imagine a pure datagram model
in which each IP packet separately carried the necessary
authorizations for all the stages in the forwarding path.
Realistically, this is not practical, since the security
information may be both unacceptably large and computationally
demanding for inclusion in every packet. This seems to imply the
need for some form of state setup for security.
Thus, we presume a two stage process that moves somewhat away from
the pure datagram model. In the first stage, the setup stage,
some state is established in the routers (and other network
elements) that describes how a subsequent stream of packets is to
be treated. In the second stage, the classification stage, the
arriving packets are matched with the correct state information
and processed. The terminology in use today calls these different
state descriptions "classes", and the process of sorting
"classification".
Setup can take many forms. It could be dynamic, invoked across
the network by an application as described above. The setup
process could also be the manual configuration of a router by
means of a protocol such as SNMP or remote login. For example, a
network link, such as a link across the Atlantic, might be shared
by a number of users who purchase it jointly. They might
implement this sharing by configuring a router with
specifications, or filters, which describe the sorts of packets
that are permitted to use each share. Whether the setup is
dynamic or manual, short-lived or semi-permanent, it has the same
effect: it creates packet classes in the router and defines how
packets are to be classified as they arrive.
Much of the current research on extensions to IP for QOS, such as
realtime service, has assumed an explicit setup phase and a
classification stage. The setup stage is accomplished using
protocols such as RSVP or ST-II, which also specify how the
subsequent classification is to be done. Security at the setup
stage would thus simply be an extension to such a protocol. It
should be noted that there are alternative proposals for realtime
QOS, based on an implicit setup process.
4.2 Securing the Setup Process.
To secure the setup process, we require that a setup request be
accompanied by user credentials that provide a trustworthy
assurance that the requester is known and is authorized to make
the request in question. We refer to the credentials used in the
setup phase as the high-level identification (HLID).
A simple version of this authorization would be a password on the
management interface to a router (the limitations of such a
password scheme are well known and not the issue here). In the
case of setup requests made by individual applications, some
user-specific authorization must be assumed.
While there could be any number of ways to organize the HLIDs, the
objective of scaling suggests that a global framework for user
naming and authentication would be useful. The choice of naming
framework is discussed further in Section 5. Note that this
discussion, which concerns controlling access to network resources
and security devices, is distinct from end-to-end authentication
and access control; however, the same authentication
infrastructure could be used for both.
In general, while significant engineering effort will be required
to define a setup architecture for the Internet, there is no need
to develop new security techniques. However, for the security
aspects of the classification process, there are significant
problems related to performance and cost. We thus focus on that
aspect of the overall framework in more detail.
Above, we defined the high-level ID (HLID) as that set of
information presented as part of a setup request. There may also
be a "low-level ID" (LLID), sometimes called a "cookie", carried
in each packet to drive classification. In current proposals for
IP extensions for QOS, packets are classified based on existing
packet fields, e.g., source and destination addresses, ports, and
protocol type.
It is important to note that the LLID is distinct from the address
of the user, at least conceptually. By stressing this distinction
we make the point that the privileges of the user are not
determined by the address in use. If the user's address changes,
the privileges do not.
The LLID in a packet acts as a form of tag that is used by some or
all routers along a path to make decisions about the sort of QOS
that shall be granted to this packet. An LLID might refer to a
data stream between a single source-destination address pair, or
it might be more general and encompass a range of data streams.
There is no requirement that the LLID embody a syntax that permits
a router to discern the QOS parameters that it represents, but
there also is no prohibition against imposing such a structure.
We propose that an IP datagram contain one LLID, which can be used
at various stages of the network to map the packet to a class. We
reject the alternative that the packet should have a variable
number of LLIDs, each one for a different point in the net.
Again, this is not just a security comment, but it has security
implications.
The attributes of the LLID should be picked to match as broad a
range of requirements as possible.
* Its duration (discussed below) must match both the needs of
the security protocol, balancing robustness and efficiency,
and the needs of the application, which will have to deal
with renewal of the setup when the LLID expires. A useful
end-node facility would be a service to renew setup requests
automatically.
* The degree of trust must be high enough to meet the most
stringent requirement we can reasonably meet.
* The granularity of the LLID structure must permit packet
classification into classes fine-grained enough for any
resource selection in the network. We should therefore
expect that each separate stream of packets from an
application will have a distinct LLID. There will be little
opportunity for aggregating multiple streams under one LLID
or one authenticator.
4.3 Validating an LLID
At a minimum, it is necessary to validate the use of an LLID in
context, i.e., to ensure that it is being asserted in an
authorized fashion. Unauthorized use of an LLID could result in
theft of service or denial-of-service attacks, where packets not
emitted by an authorized sender are accorded the QOS treatment
reserved for that sender (or for a group of which the sender is a
member). Thus, use of an LLID should be authenticated by routers
that make QOS decisions based on that LLID. (Note that not all
routers may "pay attention" to the LLID.)
In principle, the validity of an LLID assertion needs to be
checked on every packet, though not necessarily at every router;
it may be possible to restrict the checks to security perimeters.
At those routers that must validate LLIDs, there is an obvious
concern over the performance impact. Therefore, a router may
adopt a less rigorous approach to LLID validation. For example, a
router may elect to sample a data stream and validate some, but
not all, packets. It may also elect to forward packets first and
perform selective validation as a background activity. In the
least stringent approach, a router might log selected packets and
validate them as part of an audit activity much later.
There are several candidate techniques for validating the use of
LLIDs. We have identified three basic techniques, which differ in
terms of computational performance, bandwidth overhead, and
effectiveness (resistance to various forms of attack).
* Digital Signatures
The first technique entails the use of public key
cryptography and digital signatures. The sender of each
packet signs the packet (header and payload) by computing a
one-way hash over the packet and transforming the hash value
using a private key associated with the LLID. The resulting
authenticator value is included in the packet header. The
binding between the public key and the LLID is established
through a connection setup procedure that might make use of
public keys that enjoy a much longer lifetime. Using public
key technology yields the advantage that any router can
validate a packet, but no router is entrusted with data that
would enable it to generate a packet with a valid
authenticator (i.e., which would be viewed as valid by other
routers.) This characteristic makes this technique ideal
from the standpoint of the "principle of least privilege."
Public key cryptosystems such as RSA have the advantage that
validation of a signature is much faster than signing, which
reduces the router processing burden. Nonetheless, this
approach is not likely to be feasible for anything other than
selective checking by routers, given current public key
algorithm performance.
* Sealing
The next technique is based on the use of the same type of
one-way hash function used for digital signatures, but it
does not require signing the hash value. Here the sender
computes a one-way hash with a secret quantity (essentially a
"key") appended to the packet. This process is an example of
what is sometimes referred to more generically as
cryptographic "sealing." The inclusion of this key at the
end of the hash computation results in a hash value that is
not predictable by any entity not possessing the key. The
resulting hash value is the authenticator and is included in
the packet header. A router validates a packet by
recomputing the hash value over the received packet with the
same secret quantity appended. If the transmitted hash value
matches the recomputed hash value, the packet is declared
valid. Unlike the signature technique, sealing implies that
all routers capable of verifying a seal are also capable of
generating (forging) a seal. Thus, this technique requires
that the sender trust the routers not to misuse the key.
This technique has been described in terms of a single secret
key shared between the sender and all the routers that need
to validate packets associated with an LLID. A related
alternative strategy uses the same authenticator technique,
but shares the secret key on a pairwise basis, e.g., between
the sender and the first router, between the first router and
the next, etc. This avoids the need to distribute the secret
key among a large group of routers, but it requires that the
setup mechanism enable Router A to convince his neighbor
(Router B) that Router A is authorized to represent traffic
on a specific LLID or set of LLIDs. This might best be done
by encapsulating the packet inside a wrapper that both ends
of the link can validate. Once this strategy is in place, it
may even be most efficient for routers to aggregate traffic
between them, providing authentication not on a per-LLID
basis, since the router pairs are prepared to "trust" one
another to accurately represent the data stream LLIDs.
For a unicast data stream, the use of pairwise keying between
routers does not represent a real change in the trust
required of the routers or of the setup mechanism, because of
the symmetric sharing of the secret key. However, for a
multicast connection, this pairwise keying approach is
superior in that it prevents a router at one point in a
multicast tree from being able to generate traffic that could
be inserted at another point in the tree. At worst, a router
can generate spurious, but authenticatable, traffic only for
routers "below" it in the multicast tree.
Note that the use of network management fault isolation
techniques, e.g., sampling router traffic statistics at
different points along a data stream, should permit post hoc
detection of packet forgery attacks mounted by rogue routers
along a data stream path. Use of this technique could
provide a deterrent to such activity by routers, further
arguing for the pairwise keying approach.
The sealing technique is faster than the digital signature
technique, because the incremental hash calculation
(including the appended secret quantity) is much faster than
the cryptographic transformation required to sign a hash.
The processing burden is symmetric here, i.e., the sender and
each router devote the same amount of processing power to
seal a packet and to verify the seal. Also, a sealed hash
may be smaller than a signed hash, even if the same function
is used in both cases. (This is because the modulus size of
the public key signature algorithm and any ancillary
parameters tend to increase the size of the signed hash
value.) Moreover, one could use a hash function with a
"wide" value and truncate that value, if necessary to reduce
overhead; this option is not available when the authenticator
is a signed hash value.
As a variant on this technique, one could imagine a
"clearinghouse" that would receive, from the sender, the
secret key used to generate and validate authenticators. A
router needing to validate a packet would send a copy of the
packet to the clearinghouse, which would check the packet and
indicate to the router whether it was a valid packet
associated with the LLID in question. Obviously, this
variant is viable only if the router is performing
infrequent, selective packet validation. However, it does
avoid the need to share the authenticator secret among all
the routers that must validate packets.
For both of these techniques, there is a residual
vulnerability to denial-of-service attacks based on replay of
valid packets during the lifetime of a data stream. Unless
packets carry sequence numbers and routers track a sequence
number window for each data stream, an (external) attacker
can copy valid packets and replay them. It may be easiest to
protect against this form of attack by aggregating all
traffic between a pair of routers into a single flow and
providing replay protection for the flow as a whole, rather
than on a per data stream basis.
* Temporary Passwords
The final technique explored in the workshop takes a very
different tack to packet validation. The preceding
techniques compute a function of the bits in a packet and
transform that value in a fashion that prevents an intruder
from generating packets with valid authenticators. The
ability to generate packets with valid authenticators for a
given LLID requires access to a secret value that is
available only to the sender, or to the sender and to routers
participating in a given data stream.
In contrast, this third technique calls for the authenticator
to be a short term, secret quantity that is carried in the
packet header, without benefit of further protection. In
essence, this technique incorporates a short term "password"
into each packet header. This approach, like its
predecessor, requires that all of the routers validating the
LLID be privy to this authenticator. Moreover, the
authenticator is visible to any other router or other
equipment along the path, and thus this technique is much
more vulnerable than the previous ones.
Here the same authenticator may be applied to all packets
with the same LLID, since the authenticator is not a function
of the packet it authenticates. In fact, this suggests that
it is feasible to use the LLID as the authenticator.
However, adopting this tack would not be consistent with the
two previous techniques, each of which requires an explicit,
separate authenticator, and so we recommend against this
optimization.
Nonetheless, the fact that the authenticator is independent
of the packet context makes it trivial to generate (forge)
apparently authentic packets if the authenticator is
intercepted from any legitimate packet. Also, if the
authenticator can be guessed, an attacker need not even
engage in passive wiretapping to defeat this scheme. This
latter observation suggests that the authenticator must be of
sufficient size to make guessing unlikely, and making the
LLID and the authenticator separate further supports this
requirement.
The major advantage of this approach is one of performance.
The authenticator can be validated very quickly through a
simple comparison. Consistent with the need to protect
against guessing attacks, the authenticator need not consume
a significant amount of space in the packet header.
The use of a sequence number visible to the routers is an
interesting technique to explore to make these somewhat
vulnerable methods more robust. If each stream (each source
of packets) numbers its packets, then an intruder attempting
to use the network resource must delete the legitimate
packets, which in many cases would be difficult. Otherwise,
the router being attacked would notice duplicate sequence
numbers and similar anomalies. The exact details of the
numbering would have to be worked out, since for the
legitimate stream packets might be lost, which would cause
holes in the sequence space.
We do not consider here the issues of collusion, in which a user
with a given LLID and authenticator deliberately shares this with
another unauthorized user. This possibility should be explored,
to see if there is a practical advantage to this act, and thus a
real threat.
4.4 Dynamics of Setup
o Duration of LLID's
A key question in the use of LLIDs is how long they remain
valid. At one extreme, they last only a very short time,
perhaps seconds. This limits the damage that can be done if
the authenticator for the LLID is stolen. At the other
extreme, LLIDs are semi-permanent, like credit card numbers.
The techniques proposed above for securing the LLID traded
strength for efficiency, under the assumption that the peril
was limited by the limited validity of the LLID.
The counterbalancing advantage of long-term or semi-permanent
LLIDs is that it becomes practical to use primitive setup
techniques, such as manual configuration of routers to
establish packet classes. This will be important in the
short run, since deployment of security and dynamic resource
allocation protocols may not exactly track in time.
We conclude that the correct short-term action is to design
LLIDs under the assumption that they are fairly short lived,
and to tolerate, in the short run, a longer period of
validity. This would imply that we will get an acceptable
long-term mechanism in place, which operationally will have a
lower level of security at first. As we get better tools for
automatic setup, we can shorten the duration of validity on a
individual basis, without replacing mechanism in the packet
forwarding path.
o Setup Latency
The tradition of the Internet is not to impose any setup
latency in the communication path between end nodes. This
supports the classic datagram model for quick transactions,
etc., and it is a feature that should be preserved.
For setup that is done "in advance", either through a
management interface or by an end-node in the background, the
issue of latency does not arise. The latency issue occurs
for dynamic reservations made in response to a specific
application request.
We observe that while latency is a key issue, it is not
materially influenced by security concerns. The designers of
resource reservation protocols such as RSVP and ST-II are
debating the latency of these protocols today, absent
security. Adding an authenticator to the request message
will increase the processing needed to validate the request,
and might even imply a message exchange with an
authentication service, but should not substantially change
the real time of the setup stage, which might already take
time on the order of a round-trip delay. But the design of
the high level authentication and authorization methods for
the setup protocol should understand that this process, while
not demanding at the level of the per-packet processing, is
still somewhat time-critical.
One way of dealing with an expensive setup process is to set
up the request provisionally and perform the validation in
the background. This would limit the damage from one bad
setup request to a short period of time. Note, however, that
the system is still vulnerable to an attack that uses a
sequence of setup requests, each of which allows unauthorized
usage for at least a short period of time.
Note also that a denial-of-service attack can be mounted by
flooding the setup process with invalid setup requests, all
of which need to be processed and rejected. This could
prevent a valid user from setting up any state. However,
denial-of-service attacks based upon flooding leave very
large "finger prints"; they should not normally be an
important threat. If it is a problem, it may be possible to
incorporate a mechanism at the level of setup processing that
is equivalent to "fair queueing", to limits the damage from a
flooding attack at the packet level.
4.5 Receiver-Initiated Setup
Recent work on a QOS extension for the Internet, embodied in the
RSVP protocol, uses the model that the receiver will reserve
resources. This scheme is consistent with the current IP
multicast paradigm, which requires the receiver to join the
multicast group. The receiver reserves the resources to insure
that the multicast traffic reaches the receiver with the desired
QOS. In this case, it is the credentials (the HLIDs) of the
receivers that will be presented to the setup phase.
Note that receiver initiation requires an explicit setup phase.
Suppose setup were implicit, driven by pre-existing fields in the
packet. Then there would be no way to associate a packet with a
particular receiver, since in multicast, the address of the
receiver never appears in the packet.
Further, it is impossible in this case to perform a setup "in
advance", unless the sender and the receiver are very tightly co-
ordinated; otherwise, the receiver will not know in advance what
LLID will be in the packet. It is certainly impossible, in this
case, for the receiver to set up "semi-permanent" reservations for
multicast traffic coming to it. This, again, is not a security
issue; the problem exists without adding security concerns, but
the security architecture must take it into account.
4.6 Other Issues
4.6.1 Encrypting Firewalls and Bypass
Our view of security, both end node and network protection,
includes the use of firewalls, which partition the network into
regions of more or less trust. This idea has something in
common with the encrypting-firewall model used in the
military/intelligence community: red (trusted) networks
partitioned from black (untrusted) networks. The very
significant difference is that, in the military model, the
partition uses an encryption unit that encodes as much as
possible of the packet for its trip across the black network to
another red network. That is, the purpose of the encryption
unit, among others, is to provide a very high degree of
protection against disclosure for data housed within the red
networks. In contrast, our version of a firewall is more to
protect the trusted (red) region of the network from outside
attacks. It is concerned both with what comes in and with what
goes out. It does permit communication between a node on the
trusted and nodes in the untrusted parts of the network.
We would like to be able to adapt our model of secure QOS to
the case of military-style encrypting firewalls. However, this
use of encryption raises a problem with our model of secure
resource management, discussed above, which was based on a
two-stage process of setup and classification. This model is
problematic because it requires information to pass from the
red region to the black region in the clear. This information
includes both the setup packets themselves, if setup is done
dynamically from the end node, and the classification fields
(the LLIDs) in the data packets. Obviously, this information
cannot be encrypted when leaving the red region of the network,
since it would then be meaningless to the black net, so that
the black network would be unable to make resource allocation
decisions based on it.
To make this sort of control scheme work, it is necessary for
the encryption device to be programmed to permit certain
packets and fields in packets to pass through the encryptor in
the clear. This bypass of the encryption is considered highly
undesirable. In a high security situation, the process
generating the bypassing information might be corrupted, with
the result that information that should be controlled is
removed from the secure network by hiding it in the bypassed
fields of the packets.
We concluded, however, that this bypass problem is not
insurmountable. The key idea, as in all cases of bypass, is to
limit, rather than wholly outlaw, the information passing in
the clear. To limit the information needed for bypass, one can
either perform the setup as a management function totally
within the black environment, or divide the process into two
stages. The first stage, again totally in the black context,
defines a limited number of setup situations. The second stage
involves sending from the red net a very small message that
selects one request to be instantiated from among the pre-
defined set.
Perhaps the more difficult issue is the LLID in the packet
header. If the LLID is an explicit field (as we have discussed
so far, but see below), it represents a new field in each
packet, with perhaps as many as 32 bits. Again, the solution
is to limit the way this field can be used. When the end-node
performs a setup, it will specify the value of the LLID to be
used. This fact can be observed by the red/black encryption
unit, which can then limit the components of this field to the
values currently in use. To further improve the situation, the
encryption unit might be able to aggregate a number of flows
onto one flow for the purpose of crossing the black net, which
would permit a further reduction in the number of distinct
LLIDs that must escape the red region.
The details of this proposal, including some important issues
such as the time duration of LLIDs in this case, must be
considered further. However, the initial conclusion that
bypass can be incorporated into a general resource control
framework is very encouraging, since it suggests that both
military and commercial forms of security can be built out of
the same building blocks.
4.6.2 The Principle of Consistent Privilege
A well understood principle of security is the principle of
least privilege, which states that a system is most robust when
it is structured to demand the least privilege from its
components.
A related rule we observe is the principle of consistent
privilege. This can be illustrated simply in the case of
denial of service, where it is particularly relevant. For a
particular route, no assumption of service can be justified
unless we trust the routers to deliver the packets. If a
router is corrupted and will not forward packets, the only
solution is to find another route not involving this router.
We do not concern ourselves here with protocols for finding new
routes in the presence of a corrupted router, since this topic
is properly part of another topic, securing the network
infrastructure. We only observe that either we will get
service from the router or we will not. If the router is
corrupted, it does not matter how it chooses to attack us.
Thus, as long as the router is part of a forwarding path (most
generally a multicast forwarding tree), we should not hesitate
to trust it in other ways, such as by giving it shared resource
keys or LLID verifiers.
This illustrates the principle of consistent privilege. This
principle is exploited in the scheme for hop-by-hop or pairwise
use of secrets to validate LLIDs in a multicast tree. If a
single key is issued for the whole tree, then the privilege is
not consistent. We only need to trust a router with respect to
the nodes "below" it in the tree. If it fails to forward
traffic, it can affect only those nodes. But if we give it the
group key, then it can generate bogus traffic and inject it
into the tree at any point, affecting traffic for other parts
of the tree. If, on the other hand, we use pairwise keys, then
a corrupt node can only generate bogus traffic with the key for
traffic it would directly receive, which is the part of the
tree it could damage anyway.
Another requirement we must place on the network concerns
routing. If a firewall is in place, we must trust the routing
architecture not to bypass that firewall. One way to
accomplish this is to eliminate any physical path between the
regions other than those that go through the firewall.
Operational experience will be required to see if this simple
physical limit is an acceptable constraint.
4.6.3 Implicit LLID's
We stress the importance of a strong conceptual distinction
between the addresses in a packet and the LLID which is used to
classify the packet. The conceptual distinction is important,
but under limited circumstances it may be possible to overload
some of the packet fields and create an LLID from the current
packet header. For example, current packet classifiers for
IPv4, which are not secure but which seem to work for
classifying the packets into service classes, use a number of
the packet fields together as a form of LLID: the source and
destination IP addresses and ports plus the protocol type.
This sort of "implicit" LLID must be short-lived, especially if
the host can change its IP address as it moves. But if the
LLID is established by some sort of dynamic setup protocol, it
should be possible reestablish the LLID as needed.
The current IPv4 header has no authenticator field to validate
the LLID. An authenticator field could be optionally carried
in an option; adding it gives robustness to network
reservations. Any of the schemes described above for creating
an authenticator could be used, except that if the simple
password-style authenticator is used, it must be an explicit
separate field, since the LLID cannot be picked randomly.
4.6.4 Security without Setup
As we describe this architecture, the setup phase is an
essential part of the sequence. This suggests that the current
Internet, which has no setup protocols, cannot be secured
against denial-of-service attacks. It is important to explore
the limits of this point. As we stressed above, setup can
occur in many ways. Routers today offer management options to
classify packets based on protocol types and other fields found
in the header, and to use this classification to create a few
fair queueing classes that can prevent one class from
overloading the net to the exclusion of the others.
There are two problem here. The first is that for a setup done
using a management interface, the secret that is shared among
the source and the routers to validate the LLID must remain
valid for a long time, and it must be manually configured. The
second problem is that the granularity of the categories may be
coarse. However, it has been proposed, in a thesis by Radia
Perlman, that a router might create a separate fair queueing
class implicitly for each source address. This approach, which
uses the addresses as an implicit LLID, must have some form of
authenticator for robustness. But if the LLID can be trusted,
this scheme provides classification of traffic based only on an
implicit setup operation. The granularity of classification is
not sufficient to provide any QOS distinction. The only
objective is to prevent the traffic from one source from
flooding the net to the exclusion of another.
4.6.5 Validating Addresses
We make a claim here that if the LLID and the addresses in the
packet are conceptually distinct, and if there is a suitable
means to validate the LLID, then there is no reason to validate
the addresses. For example, a packet constructed with a false
source address does not seem to represent any security problem,
if its LLID can be validated.
An exception to this might possibly lie in communication with
mobile hosts, but it will require a complete model of threats
and requirements in the mobile environment to be sure.
However, we make the claim, as a starting point for discussion,
that if LLIDs are distinguished from addresses, many of the
security concerns with mobility are mitigated and perhaps
removed. This point should be validated by more detailed
consideration of the mobility problem.
4.6 Conclusions
a) It is important to conceptually separate a LLID (Low-Level
IDentifier) carried in a packet from addresses in the packet.
b) There will be a single LLID carried in each packet. Although
this might imply some additional state in the routers than if
multiple LLIDs were used, using only one LLID choice is more
scalable.
c) Hop-by-hop LLID authentication mechanisms might provide a
highly scalable approach that limits the distribution of
secrets. However, the robustness limitations must be
investigated thoroughly.
d) Statistical sampling or after-the-fact detection mechanisms
may be employed by routers to address performance concerns.
5. AN AUTHENTICATION SERVICE
The purpose of an authentication service is simply to verify names,
or more precisely to verify the origin of "messages". It differs
from the authorization service, which determines what services are
available to an authenticated name. We expect that authentication
will be an Internet-wide service, while authorization will be
specific to the resources to which access is being authorized.
This "identification" function can be used in several contexts, for
example:
* One-time passwords: "it is really <huitema@inria.fr> that is
responding to this challenge".
* Access to a firewall: "it is really <huitema@inria.fr> that is
trying to send data to host-A at port-a".
There are many Internet objects that we may want to name, e.g.,:
domain names: sophia.inria.fr
machine names: jupiter.inria.fr
service names: www.sophia.inria.fr
(in fact, a data base)
users: huitema@sophia.inria.fr
processes: p112.huitema@sophia.inria.fr
p112.sophia.inria.fr
universal resource locators:
http//www.sophia.inria.fr:222/tmp/foobar
One could be tempted to believe that the authentication service will
only be concerned with naming humans, as only humans are
"responsible"; a process obtains some access rights because it is
acting on behalf of a person. However, this is too reductive and
potentially misleading. We may have to authenticate "machines" or
hardware components. For example:
* When a machine boots it needs to access resources for
configuring itself, but it is not yet "used" by a person; there
is no user.
* On a "distributed processor", component CPUs may need to
authenticate each other.
Machines do differ from users; machines cannot keep their "secrets"
in the same way that people do. However, there is a big value in
having a simple and extensible name space.
5.1 Names and Credentials
We make the hypothesis that the authorization services will
generally use "access control lists" (ACLs), i.e., some definition
of a set of authorized users. A compact way to represent such a
set would be to allow "wildcard" authorizations, e.g., "anybody at
<Bellcore.com>", or "any machine at <INRIA.FR>". The
authentication service should be designed to facilitate the
realization of the authorization service and should support
"wildcards".
However, wildcards are not general enough. Assuming that we have
a hierarchical name space, a wildcarded entry is limited to the
naming hierarchy. For example, a name like
<huitema@sophia.inria.fr> could be matched by the wildcard
<*@sophia.inria.fr> or <*.inria.fr> or <*.fr>. This is useful as
long as one stays at INRIA, but does not solve the generic
problem. Suppose that an IETF file server at CNRI is to be
accessible by all IAB members: its ACL will explicitly list the
members by name.
The classic approach to naming, as exemplified in the X.500 model,
is to consider that people have "distinguished names". Once one
has discovered such a name through some "white pages" service, can
use it as an access key in a global directory service.
An individual may acquire authorizations from a variety of
sources. Using a pure, identity-based access control system, the
user would have to acquire multiple identities (i.e.,
distinguished names), corresponding to the roles in which she is
authorized to access different services. We discuss this approach
in the next section.
An alternative approach is for the user to have a very small
number of identities, and to have the grantors of authorizations
issue (signed) credentials granting permissions to the user,
linked to her ID. These additional signed credentials are known
as "capabilities". The user can then establish her identity
through a generic identity credential, e.g., an X.509 certificate,
and can establish authorization by presenting capabilities as
required. This is somewhat analogous to a person acquiring credit
cards linked to the name on a driver's license, and presenting the
appropriate credit card, plus the license for picture verification
of identity.
5.2 Identity-Based Authorization
Let's open the wallet of an average person: we find several
"credit cards" in it. We all have many "credit cards", e.g.,
company cards, credit cards, airline frequent flyers memberships,
driver licenses. Each of these cards is in fact a token asserting
the existence of a relation: the bank certifies that checks
presented by the bearer will be paid, the traffic authorities
certifies that the bearer has learned how to drive, etc. This is
an example of identity-based authorization, in which an individual
is given different names corresponding to different relations
entered into by that individual.
If we imagine that the name space is based upon DNS (domain)
names, then for example, the person mentioned above could be
authenticated with the names:
customer@my-big-bank.com
customer@frequent-flyer.airline.com
The model we used here is that "the name is an association". This
is consistent with name verification procedures, in which that one
builds a "chain of trust" between the user and the "resource
agent". By following a particular path in the trust graph, one
can both establish the trust and show that the user belongs to an
"authorized group".
The existence of "multiple names" for a person may or may not
imply the existence of an "equivalence" relation. It may be
useful to know that <huitema@sophia.inria.fr> and
<huitema@iab.isoc.org> are two names for the same person, but
there are many cases where the user does not want to make all his
tokens visible.
5.3 Choosing Credentials
Let's consider again the example of Christian Huitema accessing a
file at CNRI. He will have to interact with INRIA's outgoing
firewall and with CNRI's incoming controls. Regardless of whether
authorization depends upon capabilities or upon multiple
association names, a different credential may be needed in each
firewall on the path. For example, assuming multiple names are
used, he will use an INRIA name, <huitema@sophia.inria.fr>, to be
authorized by INRIA to use network resources, and he will use an
IAB name, <huitema@iab.isoc.org>, to access the file server. Thus
comes an obvious problem: how does he choose the credential
appropriate to a particular firewall? More precisely, how does
the computer program that manages the connection discover that it
should use one credential in response to INRIA's firewall
challenge and another in response to CNRI's request?
There are many possible answers. The program could simply pass
all the user's credentials and let the remote machine pick one.
This works, but poses some efficiency problems: passing all
possible names is bulky, looking through many names is long.
Advertising many names is also very undesirable for privacy and
security reasons: one does not want remote servers to collect
statistics on all the credentials that a particular user may have.
Another possibility is to let the agent that requests an
authorization pass the set of credentials that it is willing to
accept, e.g., "I am ready to serve CNRI employees and IAB
members". This poses the same privacy and security problems as
the previous solutions, although to a lesser degree. In fact, the
problem of choosing a name is the same as the generic "trust path"
model. The name to choose is merely a path in the authentication
graph, and network specialists are expected to know how to find
paths in graphs.
In the short term, it is probably possible to use a "default name"
or "principal name", at least for local transactions, and to count
on the user to "guess" the credential that is required by remote
services. To leave the local environment we need only the local
credentials; to contact a remote server we need only the
destination credentials. So we need one or maybe two credentials,
which may be derived from the destination. It will be very often
the case that the generic credential is enough; then wildcards;
then "FTP provided" tokens.
6. OTHER ISSUES
6.1 Privacy and Authentication of Multicast Groups
Multicast applications are becoming an increasingly important part
of Internet communications. Packet voice, video and shared
whiteboard can be powerful productivity tools for users. For
these applications to have maximum value to their users, a variety
of security services will be required.
Existing techniques are directly applicable to providing privacy
for a private teleconference. If each member of the conference
shares a single key for a symmetric encryption algorithm (such as
DES), existing point-to-point security techniques can be extended
to protect communication within the group from outsiders.
However, slight modifications to existing techniques are required
to accommodate the multicast environment. Each packet will
require independent cryptographic processing to ensure that
packets from multiple sources can be independently decrypted by
the numerous receivers, particularly in the presence of lost
packets. N-party authentication and key management will be
required to establish the shared key among the proper group
members. This can be done by extending existing two-party key
management techniques pairwise. For example, the conference
manager may provide the key to each member following individual
authentication; for example, this could be implemented trivially
using PEM technology. The overhead experienced by each host
computer in the conference will be similar to that of existing
point-to-point encryption applications, This overhead is be low
enough that, today, software encryption can offer adequate
performance to secure whiteboard and voice traffic, while hardware
encryption is adequate for video.
The nature of multicast communication adds an additional
requirement. Existing multicast conferences provide gradual
degradation in quality as the packet loss rate increases. To be
acceptable, authentication protocols must tolerate lost packets.
Techniques to accomplish this efficiently need to be developed.
One initial sketch is outlined below. Engineering work will be
required to validate the practicality of this approach.
The use of symmetric encryption provides the members of the
conference with effective protection from outsiders. However,
because all members of the conference share a single key, it does
not provide a means of authenticating individual conference
members. In principle, existing techniques, based on one-way hash
functions coupled with digital signatures based on asymmetric
encryption algorithms, can provide individual authentication.
One-way hash functions such as MD5 are comparable in cost to
symmetric encryption. However, digital signatures are
considerably more costly, both in computation and in communication
size. The degree of overhead depends on the quality of
authentication required.
In summary, realtime authentication at the granularity of group
membership is easy and cheap, but individual authentication is
costly in time and space. Over time, the costs of both
communications and processing are expected to decline. It is
possible that this will help make authentication at the level of
individual conference participants. There are two conflicting
trends: (1) increasing CPU speeds to provide symmetric
encryption, and (2) increasing communication data rates. If both
technologies increase proportionally, there will be no net gain,
at least if the grain size is measured in terms of bits, rather
than as a period in seconds.
The group felt that the correct approach to end-to-end controls is
the use of encryption, as discussed above. The alternative is to
control the ability of a user to join a multicast group as a
listener, or as a speaker. However, we are not comfortable with
the level of assurance that we can offer if we attempt to ensure
end-to-end semantics using these means. Any passive penetration
of the network, i.e., any wire-tap, can compromise the privacy of
the transmitted information. We must acknowledge, however, that
problems with deployment of encryption code and hardware, and
especially problems of export controls, will create a pressure to
use the tools described in Section 4 to implement a form of end-
to-end control. Such a decision would raise no new issues in
security technology. The shared key now used for encrypting the
data could instead be used as the basis for authenticating a
multicast group join request. This would require modification of
the multicast packet format, but nothing more. Our concern is not
the technical difficulty of this approach, but the level of
assurance we can offer the user.
6.2 Secure Plug-and-Play a Must
Plug-and-play is the ability to plug a new device into a network
and have it obtain the information it needs to communicate with
other devices, without requiring any new configuration
information. Secure plug-and-play is an important Internet
requirement, and a central architectural issue is whether it can
be made to scale well.
For plug-and-play operation, a new machine that is "plugged" into
the network needs to:
(1) Obtain an locator so it can communicate with other devices
(2) Register or obtain a name to be identified by (e.g., machine
name)
(3) Discover services available on the network (e.g., printers,
routers, file servers, etc.)
(4) Discover other systems on the network so it can communicate
with them.
In some environments, no security mechanisms are required because
physical security and local knowledge of the users are sufficient
protection. At the other end of the spectrum is a large network
with many groups of users, different types of outside connections,
and levels of administrative control. In such environments,
similar plug-and-play capabilities are needed, but the new device
must be "authenticated" before it can perform these functions. In
each step in the discovery process the new device must
authenticate itself prior to learning about services.
The steps might be:
- Obtain a HLID from a smart card, smart disk, or similar
device.
- Authenticate itself with the first plug-and-play server using
its HLID, to register a name and to find the location of
other services.
- Discover services available on the network (e.g., printers,
routers, file servers, etc.) based on its HLID.
- Discover other systems on the network so it can communicate
with them.
The problem of taking a system out of the box and initially
configuring it is similar to the problem of a mobile or portable
machine that a human wants to connect to a local network
temporarily in order to receive services on that network. How can
the local network authenticate the human (and therefore the
human's machine) and know which services this visiting machine is
permitted to use?
The human must be endowed with a high level identifier (HLID)
which acts as his/her passport and can be verified by the local
network. This high level identifier must be globally unique and
registered/assigned by some recognized authority.
When the human plugs the machine onto a local net, the machine
identifies itself to the net with the human's high level
identifier. If local net has a policy of permitting anyone to
plug and play on its network, it will ignore the HLID and assign
an address (locator), permitting the visitor unrestricted access
and privileges. More likely, the local net will authenticate the
HLID prior to granting the visitor an address or any privileges.
At this point, the HLID has only authenticated the visitor to the
local network; the issue of which services or resources the
visitor is entitled to use has not been addressed. It is
desirable to develop a low-overhead approach to granting
authentications to new users. This will help in the case of
visitors to a site, as well as new users joining a facility.
6.3 A Short-Term Confidentiality Mechanism
Authentication has customarily been achieved using passwords. In
the absence of active attacks, the greatest threat to computer
system security may be the ease with which passwords can be
"snooped" by the promiscuous monitoring of shared-media networks.
There are known security techniques for achieving authentication
without exposing passwords to interception, for example the
techniques implemented in the well-known Kerberos system.
However, authentication systems such as Kerberos currently operate
only in isolation within organizational boundaries. Developing
and deploying a global authentication infrastructure is an
important objective, but it will take some years. Another useful
approach in the short term is the use of a challenge-response user
authentication scheme (e.g., S/Key).
One of the groups explored another interim approach to guarding
passwords: introducing a readily-used confidentiality mechanism
based on an encrypted TCP connection. This would operate at the
IP level to encrypt the IP payload, including the TCP header, to
allow the nature as well of the contents of the communication to
be kept private. It could be implemented to provide either
"strict" protection (the connection fails if the other side cannot
decrypt your data stream) or "loose" protection (falling back to
non-private TCP if decryption fails).
Loose protection would allow interoperability with older hosts in
a seamless (non-user-intrusive) manner.
One-time keys may be exchanged during the SYN handshake that
starts the TCP connection. Using one-time keys avoids a need for
infrastructure support and does not require trust between the
organizations on the two ends of the connection. Tieing the key
exchange to the SYN handshake will avoid the possibility of having
the connection fully open without knowing the state of encryption
on both ends of the connection. Although it may still be
theoretically possible to intercept the SYN exchange and subvert
the connection by an active "man-in-the-middle" attack, in
practice such attacks on TCP connections are quite difficult
unless the routing protocols have been subverted.
The keys could be exchanged using a new option that specifies the
key exchange protocol, the data encryption algorithm, and the key
to be used to decrypt the connection. It could be possible to
include multiple options in the same SYN segment, specifying
different encryption models; the far end would then need to
acknowledge the option that it is willing to use. In this case,
the lack of an acknowledgement would imply disinterest in
decrypting the datastream. If a loose privacy policy were in
force, the connection could continue even without an
acknowledgment. The policy, "strict" or "loose", would be set by
either the user or the default configuration for the machine.
One must however observe that a TCP option can carry only a
limited amount of data. Efficient protection against crypto-
analysis of the Diffie-Hellmann scheme may require the use of a
very long modulus, e.g., 1024 bits, which cannot be carried in the
40 bytes available for TCP options. One would thus have either to
define an "extended option" format or to implement encryption in a
separate protocol layered between TCP and IP, perhaps using a
version of "IP security". The detailed engineering of such a
solution would have to be studied by a working group.
A TCP connection encryption mechanism such as that just outlined
requires no application changes, although it does require kernel
changes. It has important drawbacks, including failure to provide
privacy for privacy for UDP, and the great likelihood of export
control restrictions. If Diffie-Hellman were used, there would
also be patent issues.
7. CONCLUSIONS
As a practical matter, security must be added to the Internet
incrementally. For example, a scheme that requires, as a
precondition for any improvement, changes to application code, the
DNS, routers and firewalls all at once will be very hard to deploy.
One of the reasons the workshop explored schemes that are local to
the IP layer is that we surmise that they might be easier to deploy
in practice.
There are two competing observations that must shape planning for
Internet security. One is the well known expression: "the best is
the enemy of the good." The other is the observation that the
attacks are getting better.
Finally, it should noted that the principle of least privilege, which
was mentioned above, may be in contradiction to the principle of
least cost.
7.1 Suggested Short-Term Actions
The general recommendation for short-term Internet security policy
was that the IETF should make a list of desirable short-term
actions and then reach out to work with other organizations to
carry them out. Other organizations include regionals, which may
be in a good position to provide site security counseling services
to their customers, vendors and other providers, and other
societies. We should also give input to the US government to
influence their posture on security in the direction desired by
the community.
A suggested preliminary list of short-term actions was developed.
o Perform external diagnostic security probes
Organizations should be encouraged to use CRACK and other
tools to check the robustness of their own passwords. It
would also be useful to run a variety of security probes from
outside. Since this is a very sensitive issue, some care
needs to be taken to get the proper auspices for such
probing.
Useful probe tools include:
ISS: Klaus (GA)
SATAN: Farmer Venema
ICEPICK: NRL
o Determine Security-Risk Publication Channels
What channels should be used for disseminating information of
security risks?
o Encourage use of one-time passwords.
Available packages: S/Key, SecurID, Enigma, Digital Pathways.
o Develop and publish guidelines for protocol developers, for
security-friendliness and firewall-friendliness.
o Control topology to isolate threats
o Set privacy policy:
* Always
* As much as possible
o Bring Site Security Handbook up to date
o Support use of Kerberos
The subject of the "Clipper chip" came up several times, but there
was not sufficient discussion of this very complex issue for this
grouip to reach a recommendation. It has been observed that there
are a number of quite differing viewpoints about Clipper.
o Some people accept the government's Clipper proposal,
including key escrow by the US government and the
requirement that encryption be in hardware.
o Some people don't mind key escrow by the government in
principle, but the object to the hardware requirement.
o Some people don't mind key escrow in principle, but
don't want the government to hold the keys. They would
be comfortable with having the organization which owns
the data hold the keys.
o Some people don't want key escrow at all.
o Some people don't mind the hardware or the key escrow,
but they don't think this will be acceptable to other
countries and thus will not work internationally.
This report takes no position on any of these viewpoints.
7.2 Suggested Medium-Term Actions
These actions require some protocol design or modification;
however, they use existing security technology and require no
research.
o Authentication Protocol
There is a problem of the choice of technology. Public key
technology is generally deemed superior, but it is patented
and can also induce relatively long computations. Symmetric
key technology (Needham-Schroeder algorithm, as used in
Kerberos) has some technical drawbacks but it is not
patented. A system based on symmetric keys and used only for
authentication would be freely exportable without being
subject to patents.
o Push Kerberos
Engineering is needed on Kerberos to allow it to interoperate
with mechanisms that use public key cryptography.
o Push PEM/RIPEM/PGP...
o Develop an authenticated DNS
o Develop a key management mechanism
o Set up a certificate server infrastructure
Possible server mechanisms include the DNS, Finger, SNMP,
Email, Web, and FTP.
o Engineer authentication for the Web
7.3 Suggested Long-Term Actions
In this category, we have situations where a threat has been
identified and solutions are imaginable, but closure has not been
reached on the principles.
o Executable Apps
o Router sabotage counter-measures
o Prevent Byzantine routing.
o Proxy Computing
o Decomposition of computers
o Are there "good" viruses?
APPENDIX A -- Workshop Organization
The following list of attendees indicates also the breakout group to
which they were assigned.
Breakout Groups
Group I.1 Leader:
1 Christian Huitema, INRIA (IAB)
1 Steve Bellovin, AT&T
1 Bob Braden, ISI (IAB)
1 John Curran, NEARNET
1 Phill Gross, ANS (IETF/IAB)
1 Stev Knowles, FTP Software (Internet AD)
1 Barry Leiner, USRA (IAB)
1 Paul Mockapetris, ISI
1 Yakov Rekhter, IBM (IAB)
1 Dave Sincoskie, Bellcore (IAB)
Group I.2 Leader:
2 Steve Crocker, TIS (Security AD)
2 Jon Crowcroft
2 Steve Deering, PARC
2 Paul Francis, NTT
2 Van Jacobson, LBL
2 Phil Karn, Qualcomm
2 Allison Mankin, NRL (Transport AD, IPng AD)
2 Radia Perlman, Novell
2 John Romkey, ELF (IAB)
2 Mike StJohns, ARPA (IAB)
Group I.3 Leader:
3 Dave Clark, MIT
3 Deborah Estrin, USC
3 Elise Gerich, Merit (IAB)
3 Steve Kent, BBN (IAB)
3 Tony Lauck, DEC (IAB)
3 Tony Li, CISCO
3 Bob Hinden, Sun (IESG->IAB liaison, Routing AD)
3 Jun Murai, WIDE (IAB)
3 Scott Shenker, PARC
3 Abel Weinrib, Bellcore
The following were able to attend only the third day, due to a
conflicting ISOC Board of Trustees meeting:
Scott Bradner, Harvard (IPng AD)
Jon Postel, ISI (IAB)
The workshop agenda was as follows.
Tues Feb 8
9:00 - 10:30 Plenary
Discuss facilities, meeting goals, agenda, organization.
Establish some minimal common understandings. Assign
scenarios to Breakout I groups.
10:30 - 13:00 Breakout I meetings
Each breakout group examine one or more scenarios and
formulate a list of design questions. Lunch available on
11th floor.
13:00 - 15:00 Plenary
Report, discuss. Collate and shorten list of design
issues. Organize Breakout II groups to work on these
issues.
15:00 - 17:30 Breakout IIa meetings
Work on design issues.
Wed Feb 9
9:00 - 10:00 Plenary
Report, discuss.
10:00 - 13:30 Breakout IIb meetings
More work on design questions, develop list of
requirements.
13:30 - 14:30 Plenary
Report, discuss.
15:30 - 17:30 Breakout III groups
Thurs Feb 10
9:00 - 9:30 Plenary
9:30 - 11:00 Breakout Groups (wrapup)
11:00 - 12:00 Plenary
Discuss possible short-term security recommendations
13:00 - 14:00 Plenary -- Discuss short-term security issues
14:00 - 14:30 Plenary -- Presentation by Steve Bellovin
14:30 - 16:00 Plenary -- Long- and Medium-term
Recommendations
The following scenarios were used as a starting point for
discussions. It distinguished security-S (security as a service to
the end systems) from security-M, security as a mechanism to support
other services. The workshop was intended to be primarily concerned
with interactions among the following different *services*:
o Security-S
o Routing
o Multi-destination delivery (mcast-S)
o Realtime Packet scheduling (realtime)
o Mobility
o Accounting
(and maybe large-scale?)
These categories were then applied to the following scenarios:
S1. Support a private teleconference among mobile hosts connected to
the Internet. [Security-S, mcast-S, realtime, mobility]
S2. The group in S1 is 1/3 the Internet, i.e., there are VERY severe
scaling problems. [Security-S, mcast-S, realtime, mobility,
large-scale]
S3. Charge for communication to support a video teleconference.
[Accounting, realtime, mcast-S]
S4. I am travelling with my laptop. I tune in to radio channel IP-
RADIO, pick-up the beacon and start using it. Who gets the
bill? Why do they believe this is me? Is "me" a piece of
hardware (IP address) or a certified user (PEM certificate)?
[Mobility, accounting (, realtime, mcast-S)]
S5. A Politically Important Person will mcast an Internet
presentation, without danger of interruptions from the audience.
S6. The travel industry wants to use Internet to deliver tickets to
customer premises directly in a secure way, but the customer has
only dial-up capability. [Security-S, mobility]
S7. I am traveling with my laptop and this friendly host is running
the autoconfiguration protocol. I immediately get an address as
"mac1.friendly.host.com". (What is the difference between my
laptop and a bona fide autoconfigured local station?)
[Security-S, mobility]
S8. Multiple people are connected to a subnetwork providing mobility
(e.g., cellular, packet radio). The subnetwork is connected to
multiple places in the "fixed" backbone. How can routing be done
efficiently? [Routing, mobility]
The following scenarios that were suggested do not fit into the
primary thrust of the workshop, generally because they are single-
issue topics. Most of them are pure security topics and are
concerned with the security perimeter. The last two do not fit into
our classification system at all.
S9. XYZ corporation has two major branches on opposite ends of the
world, and they want to communicate securely over the Internet,
with each branch having IP-level connectivity to the other (not
through application gateways).
S10. I am visiting XYZ corporation, with my laptop. I want to
connect it to their LAN to read my email remotely over the
Internet. Even though I am inside their corporate firewall,
they want to be protect their machines from me.
S11. XYZ corporation is trying to use the Internet to support both
private and public networking. It wants to provide full
connectivity internally between all of its resources, and to
provide public access to certain resources (analogous of
anonymous ftp servers)
S12. The travel industry wants to use Internet to deliver tickets to
customer premises directly in a secure way.
S13. Some hacker is deliberately subverting routing protocols,
including mobile and multicast routing. Design counter
measures.
S14. Part of the Internet is running IPv4 and part is running IPng
(i.e. the Internet is in transition). How can we assure
continued secure operation through such a transition?
S15. A corporation uses ATM to connect a number of its sites. It also
uses Internet. It wants to make use of the ATM as its primary
carrier, but also wants to utilize other networking technologies
as appropriate (e.g., mobile radio). It wants to support all
media (data, voice, video).
Security Considerations
This memo is entirely concerned with security issues.
Authors' Addresses
Bob Braden [Editor]
USC Information Sciences Institute
4676 Admiralty Way
Marina del Rey, CA 90292-6695
Phone: (310) 822-1511
EMail: Braden@ISI.EDU
David Clark
MIT Laboratory for Computer Science
545 Technology Square
Cambridge, MA 02139-1986
Phone: 617-253-6003
EMail: ddc@lcs.mit.edu
Steve Crocker
Trusted Information Systems, Inc.
3060 Washington Road (Rte 97)
Glenwood, MD 21738
Phone: (301) 854-6889
EMail: crocker@tis.com
Christian Huitema
INRIA, Sophia-Antipolis
2004 Route des Lucioles
BP 109
F-06561 Valbonne Cedex
France
Phone: +33 93 65 77 15
EMail: Christian.Huitema@MIRSA.INRIA.FR