Embedded Markup Considered Harmful
A second-generation religion shifts emphasis, because people's concerns have changed--perhaps with new lands to conquer, less worry about sin. The priests adjust the previous fundamental idea to grapple with the new situation. But in this second generation, priests must also show fidelity to the terminology of the earlier generation, framing their new concerns amongst the old ideas wherever possible. Everyone is stuck with the concepts already elucidated.
SGML is a first-generation religion. Its founding idea was to represent nameless fonts and abstracted text blocks at one remove from complete specification, so that the fonts and text blocks could be reformatted by changing a short list of definitions.
This idea was then expanded to fill the universe, becoming a technique for the sequential, hierarchical representation of any data, with embedded tags representing Beginnings and Ends. Great emphasis was put on formal correctness, defining a strict standard for compliance. Originally intended to create order in type-font selection, SGML has been extended and extended to fill the universe, becoming a reference language of sequential attributes and now hypertext links and graphics (HTML). Its believers think SGML can represent anything at all--at least, anything they approve of.
But now we see a change. The second generation of the SGML faith is the HTML religion, whose intention and outlook are entirely different, but which preaches in the robes of the old. A new land has been conquered--the Web. There is great prosperity, as in the time of Solomon, so that sin--formal correctness--is not a worry.
There is no one reason this approach is wrong; I believe it is wrong in almost every respect. But I must be honest and acknowledge my objection as a serious paradigm conflict, or (if you will) religious conflict. In paradigm conflict and religious conflict, there can be no hope of doctrinal victory; the best we can seek is for both sides to understand each other fully and cordially.
SGML's advocates expect, or wish to enforce, a universal linear representation of hierarchial structure.
I believe that if this is a factual claim of appropriateness, it is a delusion; if it is an enforcement, it is an intolerable imposition which drastically curtails the representation of non-hierarchical media structure.
I will turn to general problems of the embedded method. I have three extremely different objections to embedded markup. The first is simple; the second is complicated to explain; and the third challenges the claim of generality.
If material is to be edited, it generally must be frequently counted to perform the edit operations. Tags throw off the counts. This means that while text is being reworked, some other representation must be maintained, or complex tricks invoked to maintain counts.[2] This seems quite wrong.
Network electronic publishing offers a unique special-case solution to the copyright problem that has not been generally recognized. I call it transpublishing. Let me explain.
In paper publishing, there are two copyright realms: a fortified zone of copyrighted material, defended by its owners and requiring prior negotiation by publishers for quotation and re-use; and an unfortified zone, the open sea of public domain, where anything may be quoted freely--but whose materials tend to be outdated and less desirable for re-use.
Transpublishing makes possible a new realm between these two, where everything may be treated as boilerplate (as with public-domain material), but where publishers relinquish none of their rights and receive revenue exactly proportional to use.
Two different parties have legitimate concerns. Original rightsholders are concerned for their territory of copyrighted material, as defined by law, so that they may maintain and benefit from their hard-won assets. But the public (everybody else, as well as rightsholders in their time off) would like to re-use and republish these materials in different ways.
What if a system could exist which would satisfy all parties--copyright holders and those who would like to quote and republish? What if materials could be quoted without restriction, or size limit, by anyone, without red tape or negotiation--but all publishers would continue to furnish the downloaded copies, and would be exactly rewarded, being paid for each copy?
Transpublication is a unique arrangement--only possible online--which can achieve this win-win solution.[3]
Naturally the original rightsholder must give permission for this in advance ("transcopyright"). [3]
Transpublishing turns all participating materials into virtual clip art, freely to be recomposited into new online contexts. Its advantages are special. It provides a bridge to the original (a great benefit to understanding the written intent of the author, and possibly the author's reputation).
Furthermore, with a suitable micropayment system,[4] transpublishing should provide also a means by which the publisher is paid for each manifestation[5] thus quoted.
Second--and it has taken a long time to get to this point[6]--the quoting author may legitimately want to change fonts and and markup.
This is done all the time in scholarly writing and serious journalism, with phrases like "emphasis mine." It needs to be possible in transpublishing to change emphasis and other attributes by nullifying the original markup. Of course, re-emphasizing through markup is an editorial modification, subject to judgment calls and issues of academic etiquette. But the inquiring reader can always follow the bridge of transclusion to see the original as formatted by the author.
There are two markup solutions to make transpublishing work with SGML and HTML.
The tags can be like those of SGML, but they are not embedded in the text itself. They are in parallel streams which reference positions in the text data stream. Thus each tag is preceded by a count showing how far the tag is after the previous tag. (This incremental counting, rather than stating each tag's distance from the beginning, is to facilitate editing.)
This method has several advantages:
Thus I believe we should call it "the Parallel Representation of SGML," and make it an optional part of the SGML standard.
This is smarmier at the implementation level, losing the benefit of clean counting and requiring a more complex editing apparatus. Otherwise it has the advantages of parallel markup: pluralism, editability, and transclusion with variation.
Note that this is tag override, not overload, since no symbol is being redefined.
I believe that embedded structure, enforcing sequence and hierarchy, limits the kinds of structure that can be expressed. The question we must ask is: What is the real structure of a thing or a document? (And does it reasonably fit the allowed spectrum of variation within the notational system?)
You can always force structures into other structures and claim that they're undamaged; another way to say this is that if you smash things up it is easier to make them fit. Enforcing sequence and hierarchy simply restricts the possibilities.
Like a TV dinner, embedded markup nominally contains everything you could want. "What else could you possibly want?" means "It's not on the menu."
This issue creates a very different focus from that of the markup community: the task I see is not merely to represent frozen objects tossed over the transom by an author or management, or format static structures for printout or screen, but to maintain a continuing evolutionary base of material and to track the changes in it.
To find the support functions really needed for creative organization by authors and editors, we must understand the exact representation and presentation of human thought, and be able to track the continuities of structure and change.
This means we must find a stable means of representing structure very different from the sequential and hierarchial--a representation of structure which recognizes the most anarchic and overlapping relations; and the location of identical and corresponding materials in different versions; which recognizes and maintains constancies of structure and data across successive versions, even as addresses of these materials become unpredictably fragmented by editing.
Thus deep version management--knowing locations of shared materials to the byte level--is a vital problem to solve in the design of editing systems. And the same location management is necessary on a much broader scale to support transpublishing.
Embedded markup cannot represent this at all, and merely adds obstacles (impeded data structure) to solving these rich addressing problems.
I would propose a three-layer model:[8]
Few understand the true nature of hypertext and its relation to thought, let alone the vast interconnection of ideas, and the way that most expressions of ideas sever and misrepresent them. Today's popular but trivially-structured Web hypertext has excused people from seeing the real hypertext issues, or being able to create and publish deep complexes of thought.
We greatly need a general structure to represent all forms of interconnection and structure, and changes in both content and structure; and to visualize and re-use variants and alternatives, comparing them in context in order to understand and choose.
Mapping these serious concerns to an SGML-HTML template is not a minor inconvenience but an impossible violation of the problem.
Of course, people always try to fit information into a familiar mold, even when that structure has shown itself inhospitable, unshaped to that information. C. Northcote Parkinson has pointed out [4] that the fullest flowering of a paradigm, at least as seen by its participants--all gaps closed and issues unseen, the people no longer aware that there are any unsatisfied problems--may indicate that the paradigm is near its end.
Mr. Nelson has written several books, the most recent being The Future of Information (1997), as well as numerous articles, lectures, and presentations. He is best known for discovering the hypertext concept and for coining various words which have become popular, such as "hypertext," "hypermedia," "cybercrud," "softcopy," "electronic visualization," "dildonics," "technoid," "docuverse," and "transclusion."
He received a B.A. in Philosophy from Swarthmore College in 1959 and an M.A. in Social Relations from Harvard in 1963.
XML.com Copyright © 1998-2006 O'Reilly Media, Inc.