[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
RE: [syndication] site-wide metadata [was: RFC: myPublicFeeds.opml]
Thanks Rick, wish I'd said that ;-)
> -----Original Message-----
> From: Rick Bradley [mailto:roundeye@roundeye.net]
> Sent: 15 October 2003 23:01
> To: syndication@yahoogroups.com
> Subject: Re: [syndication] site-wide metadata [was: RFC:
> myPublicFeeds.opml]
>
>
> * Dave Winer (dave@userland.com) [031015 14:14]:
> > 2. The train left the station on this with robots.txt. The
> world survived.
> > Sorry, I noticed. I wasn't supposed to, I guess, but I did.
>
> The robots.txt convention is not scalable, addresses a different
> population, and was a punt at a time when most of the future use cases
> we're now addressing were unimaginable. It's like pointing to chariots
> as a best practice when proposing a way to increase fuel efficiency in a
> new car.
>
> > 3. Instead of impeding growth, it would be great if the W3C did
> things to
> > foster growth,
>
> Last I checked the W3C brainstorms a lot and issues recommendations.
> One can follow said recommendations or choose not to. I don't see how
> the W3C is impeding much of anything in that manner. In fact the amount
> of positive growth on the Internet is staggering. Methinks you're
> confused.
>
> > like declare one top-level name to be reserved, and then establish a
> > web app to reserve names within that folder. It would take a weekend
> > to write and deploy. I'll help.
>
> I don't understand how issuing a recommendation that serves little but
> to further your misguided discoverability scheme is "foster[ing]
> growth". Noone is stopping you from writing and deploying anything --
> in fact I'm firmly convinced it's impossible to deter you from doing
> whatever you please, protestations of impedance notwithstanding. A
> number of us are concerned that your connections with Userland
> (disclaimers aside) and their history of adopting strongly-pushed
> weakly-considered designs from time to time, will result in widespread
> adoption of another bad idea.
>
> > 4.Further, only optional functionality should depend on these
> names, nothing
> > mission critical (as previous applications have done).
>
> Robots.txt, and favicon.ico represent optional non-mission-critical
> functionality. They are also conventions imposed by lack of design, and
> as designs will not scale as prototypes for discoverability services.
>
> It is one thing to look at an early punt (robots.txt) and say "hey we
> survived that", or at a monopolist's feature extension (favicon.ico)
> with no concern for the impact of their poor design on the Internet and
> say "see, there's another example"; but in this age, where we should not
> plead ignorance, shoving a single file in the *server's* DocumentRoot is
> poor planning, poorly scalable, and makes no consideration for future
> use cases. In fact it makes little consideration for current use cases.
>
> Some people will push a bad idea relentlessly because it's expedient for
> whatever quick hack they're trying to push at the moment. For my small
> part, as things stand with discoverability:
>
> - Yes, I could implement some /foo.opml file dynamically, even going so
> far as to serve up different content based upon agent, referer, host,
> and source information, and maybe be able to then partition
> user/category/application-specific feeds, but it would be working
> *against* the designed discoverability system, not with it.
>
> - But I won't, because it's a broken design being pushed for no clear
> good reason. I'll either use the format and just <link> it or I'll
> support a different standard altogther (there's almost certainly
> going to come an RDF standard if this OPML+/foo.opml monstrosity
> escapes the womb).
>
> - And I'll go a step further and probably devnull requests from IPs who
> ever go-a-fishing for such a file -- their software's poorly designed
> and is no concern of mine, let the users complain upstream to their
> boneheaded vendor.
>
>
> I can sum up my opposition to the discoverability scheme under question
> in two points:
>
> 1 - Don't go fishing.
>
> Robots.txt is /tolerated/ because Google is a good thing.
>
> Favicon.ico is widely regarded as an annoyance by everyone who
> understands what it represents, even by some who play the
> favicon.ico game ("what, one icon for the whole domain???").
>
> With each default file that propagates into the consensus standard
> stream there is more incentive for someone else to not think
> through the consequences and add yet another. Eventually everyone
> will assume that the Right Way to standardize discovery is to look
> for some file in / (or /w3c or /default or all the way up the URL
> path to the root). Pretty soon you've got so many possible things
> that could be on your server that you're getting more hits for
> discovery than you are for documents -- AND THERE'S NO WAY TO OPT
> OUT.
>
> The only way to stop this descent into discoverability hell is to
> stop it before it really starts.
>
> 2 - There are unrecognized use-cases which will break.
>
> Many of the stakeholders in web language standards are stakeholders
> precisely because they're well-entrenched in the current paradigm.
> The current paradigm knows websites in the 1-server 1-site model or
> the 1-user 1-site model. The current paradigm knows "blogs" where
> there is a "page" that has a linear list of content items. In both
> of these paradigms the notion of a single file per domain may be
> stretchable, but it's already at the stretching point as people are
> breaking the website and blog mold. (The bulk of current interfaces
> are also hierarchical at the URL level.)
>
> From a technical perspective there is very little that is
> interesting about a website these days. Similarly, there is very
> little technically interesting about blogs these days. In fact,
> blogs are boring, wikis are boring, websites are boring (and all
> have been for at least a couple of years now). They're boring like
> light bulbs and microwave ovens -- useful and once novel, but now
> merely utilitarian.
>
> Designing for the current predominant use cases is (a) designing
> for the boring, and (b) designing for quick obsolescence. We have
> to make sure not to break things, but we should be designing
> something that will at least be useful next year.
>
> A couple of things that don't fit with the /foo.opml
> discoverability scheme but which are (or have been) interesting:
>
> - graph-based content architectures: The notion of a "page" (or a
> site) is, frankly, too limiting (and too unimaginative), of
> linear content lists is simplisme, of having some limiting
> document specify layout is silly. View content as nodes in
> graphs, with edges representing certain relationships between
> reusable content.
>
> 4 years ago I worked on a system where content could be reused
> across a set of Internet domains (e.g., you could transparently
> access objects visible on foobar.com on barfoo.com, presuming
> they were hosted on the same server -- syndication between
> servers would be straightforward but was never developed), and
> arbitrarily repurposed: select a graph from the repository and a
> starting node, choose a set of rendering widgets, and view the
> rendered content defined by a graph traversal.
>
> You can apply a REST interface to such a system but the notion
> of a nested hierarchy of URLs didn't particularly fit, and the
> thought of having a single index of feeds (which are just
> renderings of graphs from certain starting points) at the / of
> every participating domain is a non-starter.
>
> The way one might build this now has changed dramatically: XML
> databases, XML fragments for nodes, RDF for edges, XSLT for
> templates, REST and/or web services for manipulating the
> repository, RSS et al. for syndication, etc. But the
> applications are still compelling, and the incongruity with this
> discoverability scheme still significant.
>
> - Convergent applications: I feel the "place" where I store
> information for my use should be the same "place" where I store
> information for public consumption, which should be the same
> "place" where I access applications that work with my data. I
> have a web-based jukebox system whose data I am now adding to an
> XML repository, along with other unrelated data I am bringing
> online and across-online (if you will). The new player
> interface (a player client downloads a playlist to play or
> stream from its location) will just use RSS feeds to list songs
> to be played. Those feeds would be listed where appropriate,
> but wouldn't be listed with feeds related to (say) recent
> publications, or recently read news related to Linux. Feeds
> come and go as players come and go. Public feeds may be
> usefully discoverable, while private feeds would not be.
>
> It's doubtful that the interfaces to these various applications
> will be organized hierarchically (some of the accesses may be
> simply through XPath+XSLT), so /foo.opml-type discoverability is
> going to be a waste of time.
>
> These may not be the most concise use-cases, but they represent real
> directions different from the overdone blog+website world that we
> appear to be designing for (or in the case of /foo.opml the static
> webpage and tacky animated-GIF world we'd be designing back to).
>
> > That's about all I want to say on this. I've noticed how much
> effort one has
> > to put in to innovate here. This mail list is broken. Sorry, I
> forgot. ;->
>
> Meditation on the causes of purported brokenness is suggested.
>
> Rick
> --
> http://www.rickbradley.com MUPRN: 747
> | it, and that is the
> random email haiku | key to passing things into
> | long term memory.
>
>
>
> Your use of Yahoo! Groups is subject to http://docs.yahoo.com/info/terms/
>
>