September 28th, 2008 ontoligent
In a recent message to the Humanist listserv, Willard McCarty directs our attention to these remarks from Peter Sloterdijk’s essay, “The Operable Man: On the Ethical State of Gene Technology” (emphasis mine)”
In the current state of the world, the single most striking feature of intellectual and technological history that is that technological culture is producing a new state of language and writing. This new state has hardly anything in common anymore with traditional interpretations of language and writing by religion, metaphysics and humanism. The old House of Being turns out to be something wherein a residence in the sense of dwelling and of the bringing close of the distant is hardly possible any longer. Speaking and writing in the age of digital codes and genetic transcriptions no longer make any kind of familiar sense; the typefaces of technology are developing apart from transmission, and no longer evoke homeliness or the effects of befriending the external. On the contrary, they increase the scope of the external and that which can never be assimilated. The province of language is shrinking, while the sector of straight-forward text is growing. Heidegger, in his letter “On Humanism,” expressed these problems in an old-fashioned, yet factually correct manner, when he called homelessness the outstanding ontological feature of man’s contemporary modus essendi.
Homelessness is coming to be the destiny of the world. Hence it is necessary to think that destiny in terms of the history of Being … Technology is in its essence a destiny within the history of Being … As a form of truth technology is grounded in the history of metaphysics.”
This makes a great deal of sense to me. The idea of homelessness (alienation, really) describes the combined effects of the non-linearity of the database as symbolic form (Manovich) on cognition and the translocatedness introduced by life in the cloud on sociality. The two forces reinforce each other: as the web/cloud becomes more pervasive and invasive by the multiplication and miniaturization of innumerable personal devices, the form of the database is increasingly privileged as the dominant medium of knowledge formation and acquisition. Indeed, we, as members of the cloud (Wesch’s machine), actually become members of an actual, massive database — the social graph that connects us all in the web, and which Nick Carr rightly calls one big brain.
The role of the biological is also worth noting: in the form of genetics, biological images have served since the beginning of the cybernetic era (1950s) the role of naturalizing core symbol, legitimating the category of Information by both expressing and reinforcing the idea that knowledge consists essentially in strings of bits (like a ticker tape), with entropy serving as the core codec for how organisms communicate. This relationship has only gotten stronger with the rise of the web and the success of biological science and nanotechnology. (For more on the ideological connection between biology and information, see Kay’s Cybernetics, Information, Life: The Emergence of Scriptural Representations of Heredity for a critical genealogy, and Meyer and Davis’ It’s Alive! The Coming Convergence of Information, Biology, and Business for an unabashed attempt to monetize it.)
Ultimately, this technologically-driven Heideggerian homelessness raises questions about the validity of notions about community and sociality raised by Web 2.0 enthusiasts, especially those in the instructional technology sector. Is virtual community really community? Does it share, for example, any of the traits described by Tönnies’ notion of community (gemeinschaft)? Or is it community in the sense of Benedict Anderson’s “imagined community”? Not that the two are entirely antithetical — Anderson’s notion of national community is an interesting synthesis of the community and society set up by Tönnies. But if virtual community is the ideological inheritor of the imagined national community, what form of social and political organization does it relfect and reinforce? What is the habitat for this new habitus? Whatever it is, it has a lot to do with Google, it’s off-shore cities of data, and globalization.
Posted in theory | 5 Comments »
September 9th, 2008 ontoligent
Proposal
The Semantic Web (SW), also known as Web 3.0, promises to transform the World Wide Web from a large, indexed collection of hyperlinked documents into a vast knowledge base of documents, maps, and other media forms that can be linked, remixed, and mashed up by intelligent agents for a variety of purposes. Your mission is to create a prototypical content development system — by modding a blog or wiki, for example — that would allow for the mass creation of mashable semantic content. Such a tool, if simple enough to use, could make a viral contribution to the emergence of the SW.
Background
The key feature of the SW (and the reason it’s called semantic in the first place) is the addition of a layer of markup to the web’s documents that describes, in both a machine and human readable way, the “meaning” of what documents contain. For example, instead of simply italicizing and capitalizing a string of text in a document to signify that it is a book title, one would wrap the string in a title tag, like so: “<title>Price and Prejudice</title>.” In addition, these tags would be part of a larger framework, or ontology, that would provide the opportunity for machines to disambiguate further the meanings of the strings in documents.
With a critical mass of properly marked up documents, it is easy to write programs that can combine documents to produce other documents or media forms that summarize or visualize the content of the source documents. In the language of relational databases, SW markup allows programmers to develop agents that can automatically perform join-like queries in a document collection, producing interesting and useful reports. (Object-oriented database developers may recognize the concept of traversal here.) The key is a shared set of tags and identifiers among the documents.
The problem, of course, is how to generate a critical mass of marked up documents. Traditional SW proponents envision a new regime of mark-up in which all documents are written in according to a newer, more complex dialect of XML to replace (x)HTML. This is known as the “bottom-up” approach to the SW. We know that’s never going to happen. If anything, Web 2.0 technologies have taught us that critical masses of content, in which network effects are possible, require very low thresholds of participation. The reason blogs and wikis are so incredibly successful is that they are incredibly easy to use.
Another approach to the SW is the “top-down” approach. In this view, machines do a lot of the work to interpret what humans already know how to read. In effect, this view puts its faith in artificial intelligence and the model of Google in being to tell what an italicized, capitalized string of text is in a document. (Projects like CiteSeer have shown that this is possible with traditional citation styles in academic documents.) The problem with this approach is that, beyond some simple things like text references, it is very hard to write programs that can read text they way humans do. We will get there at some point (probably sooner than we imagine), but we aren’t there now.
In between these views is a method that adopts the wisdom of Web 2.0 applications — the use of microformats. Instead of trying to reinvent either the way content producers write, or the way search engines work, the microformats method piggy-backs on existing mark-up and search practices and adds the 20% of effort that may make the 80% difference (or maybe it’s 1 to 99). For example, RDFa is a standard that allows users to add attributes to existing XHMTL documents that provides semantic content. And there are already in use other standards for adding semantic content to documents: hCalendar, FOAF, XFN, hCard, etc.
But these formats are still not easy enough to use. Ideally, one would have a user interface that makes it easy to add these attributes to elements as one writes them, something as easy or easier than adding a trackback or a set of tags to a blog post. Imagine being able to block off an arbitrary segment of text, have a dialog box appear asking what kind of content this is, and then adding a few simple attributes, perhaps with some AJAX-fed data fields to smooth the process. Imagine also being able to add any number of microformats to a document, and getting them from a public repository.
Efforts to produce such a tool would not be in vain. We know that there is already a productive mutual calibration that goes on between content providers and search engines. For example, there is a whole subindustry in SEO — search engine optimization — in which content providers have demonstrated their willingness to adapt content to the ways of search engines, such as Google. And it works the other way, too: all the buzz about Web 3.0 has led companies like Yahoo! to incorporate semantic web principles into their search engines, leading to a new field of semantic search optimization. The industry seems ripe for a spark to catalyze the various developments in the field of the semantic web — see Calais, Twine, and dbPedia for some examples.
The content development tool suggested by this proposal could be that spark. Such a spark could ignite a dynamic in which search engines will begin to seek out, privilege, and select for documents with semantic content. The pressure will then be on for content providers — and I mean anyone who produces web content, not just big companies — to shape up their content for the search engines, and the forces of distribution will pull production in its considerable draft.
A Practical Note
As for which content development system to use, three come to mind as both standard and extensible: MediaWiki, WordPress, and Drupal.
Posted in proposal | No Comments »