[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [RRG] Tunnel fragmentation/reassembly for RRG map-and-encaps architectures




Noel,

A mapping cache is a fine and tractable object IF AND ONLY IF the
working set at that location is tractable.

Definitely a key architectual observation, although I'd rephrase it more explicitly (although this is no doubt what you meant with your abbreviated
version) as:

"A mapping cache is a fine and tractable object IF AND ONLY IF the working set at that location is enough smaller than the full map that the savings are significantly larger than the overhead (complexity, etc) of having a
  cache."


I disagree here, a bit. It's possible to make the full map tractable and in that case, the cache is still tractable. It would provide no savings over the full map at that point. There are other architectural drivers that might still make it the preferred approach.


So there are two questions. The first is 'is this the case here'?


It is in some cases. Certainly in the SOHO kind of setting, the working set is likely to be trivially small. However, (and Dino is the one who pointed this out) there are a number of large content providers that have content that has extremely broad distribution, are exceedingly latency sensitive, and thus would have a working set that's effectively the full map.


The second
is, 'if not, is it practical to keep the full map'? Because if the answers
are 'no' and 'no', it's back to the drawing board....


That very much goes to the granularity of the map. If you go to a per-host map, then I have some concerns. It would make things very expensive. However, the number of places that would have this level of expense is very, very small. Architecturally, it might still fly.



So what about the first question? I think for a lot of ITR's (and maybe almost all), this could be true. (Although this is just a guess, and could be way wrong. For instance, spam and port-scanning seem to be from everywhere, to everywhere, and that will blow the caches. Although since those are all incoming traffic, if we design the system so that incoming requests carry, as an optimization, the reverse mapping, we'd be OK, in terms of load on the
resolution system.)

Anyway, for 'normal' users, for smaller sites, even when you integrate across
a number of users, they are communicating with only a fraction of the
endpoints in the Internet anyway.

For another (and I think this is a more important factor), as the Internet grows and the Web has a lot more content in other languages, you'll see a natural splitting up of the user population into communities of interest; English-speakers aren't going to be looking at a lot of sites in Mandarin,
etc, etc.

One thing that could throw a monkey wrench into this is if we see a lot of
cross-group hosting; i.e. web servers which contain content in many
languages, so that even if the user communities are disjoint, they may be
interacting with a single shared substrate of servers.

So except for the Googles of the world, maybe caching will work?


Yes, I have to agree. In fact, if you can make the system carry the reverse mappings (do they need to be secured?) you might even make the Googles work.


The obvious counter-point is going to be 'well, why doesn't caching work in the routers, then'? I think that's because (especially in core routers) the traffic aggregation as you go up the transmission hierarchy (think of it as
the same as going from surface roads up to high-speed limited-access
highways) naturally brings up a bigger set of sources and destinations. E.g. in the US there's a lot of 'through' traffic (q.v. all the razz-ma- tazz in the US about eavesdropping on transit traffic from a non-US-source to a
non-US-destination).


That's correct. At the edge of the network, the working set will be far smaller because you have a much smaller cross section of sources and destinations. In the transit case, you'll simply die.

Note that we also saw this with netflow.  ;-) ;-(


a) I'm in favor of caches in some places and b) I'm against them in
others.

More detail? Remember, there are no 'core ITR's'...



Again, it's all about the working set. Normal enterprises should be cacheable. Google's might not be. Comcasts would be cacheable, but you have to put the cache in the PE box, not the CE because the end user can't manage it.

No, what I saw was a total mess of mechanisms, with no coherency.

Is coherency in the eye of the beholder? Systems with radically different
operating points are going to look radically different.

As long as each operates well, they don't interfere with one another, and there is a good algorithm/whatever for deciding which one to use in a given
situation, is it a problem if they are very dissimilar?


Something that looks like three different systems still looks like three different systems. What we're talking here is a smooth continuum of solutions with a common infrastructure. There was no dial that you could set for "light", "normal", and "heavy duty".

Tony


--
to unsubscribe send a message to rrg-request@psg.com with the
word 'unsubscribe' in a single line as the message text body.
archive: <http://psg.com/lists/rrg/> & ftp://psg.com/pub/lists/rrg