I will give you a couple of examples, I think DJ may be able to provide some others. I'll start from the bottom up with self-regulatory traditions that came from the engineers who built the Internet and the primary one that really affects us is the robots.txt protocol, which I mentioned earlier, which is the way -- there's actually technically two varieties. You can use robots.txt or something called metatags, which is text in the source code of a page that's not visible to the user. Used following a standardised protocol that every webmaster can follow the same way, that every search engine can understand, for webmasters to give instructions saying, "Don't index me", or they can vary a little, they can say, "Index me but don't show a snippet", or things like that, so that's a sort of a foundational example in the world we operate.
An example that comes closer to the kinds of things we think of as regulation is the work done by groups like the IWF, the Internet Watch Foundation in the UK, and some comparable groups like the BPJM in Germany and Nikmeg(?) in the US. The IWF is primarily private, I think almost all of its funding comes from corporate members, and it creates standards for creating and disseminating lists of urls with child abuse content, very abhorrent content, so that we can get those lists disseminated quickly to every who might need to act to them. It's a pretty effective example of a self-regulatory body that came together through industry agreement.
DJ, I think you --