www.nettime.org
Nettime mailing list archives

Re: <nettime> Goodbye Classic ?
t byfield on Wed, 7 Nov 2007 12:24:33 +0100 (CET)


[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]

Re: <nettime> Goodbye Classic ?


fc-nettime {AT} plaintext.cc (Tue 11/06/07 at 05:58 PM +0100):

> You can work around it with emulators like SheepSaver and Basilisk II.

True, but it's also worth noting that this kind of solution is part of 
the broad trend of -- I'll use a phrase that floated by somewhere I've
forgotten -- devolving risk to the consumer. I hope I don't need to add 
that 'risk' and 'consumer' are very loaded terms; it might make more 
sense to speak of devolving additional roles and responsibilities to 
individuals. Either way, the point is that requisite levels of technical 
know-how are becoming an increasingly burdensome aspect of many kinds of
cultural activity and exchange. It doesn't take much technical experience
to check out a book from a library, make a photocopy, keep an eye on the 
movie listings of repertory movie houses, record a cassette tape, or bend 
a hanger into a makeshift TV antenna. And indeed the CD-ROM as a storage 
and transport medium has proven to be astoundingly durable; the first one 
I saw was 1980 or 1981, when the word on the street was expressed in terms
of the faults of vinyl -- so, for example, it was claimed that you could
run over a CD with a car and it'd still play. I never actually saw an 
'industry rep' make that claim, but I assume that the person who owned 
that insanely expensive CD playback device heard it from an interested 
party -- back in that now-inconceivable world when it would take a full-
fledged [and -funded] technical specialist or investigative journalist to 
dredge up more accurate facts. That might seem digressive, but it's not:
the volume and velocity of ambient technical 'knowledge' needed to do 
anything more than merely 'consume' anything other than the 'media' put
in front of us is really breathtaking compared to just a few years ago.
What's even more amazing is the ease with which this knowledge transfer
takes place. My daughter, who's just over two, can: (1) reliably disable 
the keypad lock on my mobile phone using a technqiue I haven't figured
out; (2) plug a USB cable into a USB plug (as opposed to Firewire, say);
(3) unreliably but persistently manipulate a remote control to maneuver
through chapters on a DVD and songs on a Mac; and so on. No one taught 
her any of this; on the contrary, most of these activities are ultimately
little more than dressed-up versions of the kinds of cognitive puzzles 
that are typical fatures of toys made for kids that age. There are lots
of ways to interpret this; one importan one, I think, is to acknowledge
that the flipside of what's often spoken of in terms of cultural loss is
also a testament to human ingenuity. You needn't look much further than,
say, the Dead Media archives to see how *little* has been lost. Or, if
you'd like, you can look back as far as the mythopoetic burning of the 
Library of Alexandria. Would it be nice to have all the books? Sure. But
let's not delude ourselves: if we *did* have them, the result would be 
parochial bickering among specialists in various fields of classical 
studies. The loss en masse has a much broader value, as a symbol in all
its luminous generality. The same applies to more recent references like
_Fahrenheit 451_: this or that camp may invoke it to promote the idea 
that its world is being destroyed, but, really, does anyone propose any
*specific* work will be lost? Maybe, but not nearly as often as people
invoke it as a ~metaphor for a more discurive loss.

But why should it be the responsilibility of the 'consumer' to conjure 
up ways to preserve particular media objects? Why shouldn't it equally
be the responsibility of those who made the works in question to 'update'
their works? Maybe, as you suggest and based on the immanent lessons of
the problem, into what they think will be more neutral, dirable forms?

I've bumped into this in a few forms recently: Hypercard stacks for kids;
William Forsythe's _Improvisation Technologies_ CD-ROM; and an unfinished
project of my own, a 'fictional phone book' that consisted of a mass of 
my own amateur programming in high-level languages to extract data from 
the first consumer-level databases of national addressing info. In the 
case of the Hypercard stacks, there's a micro-fandom scene that's given
pointers about how to port the works (albeit in Japanese, which isn't 
my forte); I wouldn't think of suggesting to the authors that they bore
a responsibility to port the works to a newer (and equally ephemeral)
format. In the case of Forsythe's work, I think there's a strong case 
to be made than ZKM, which played an important role in publishing the
work, to play a similar role in preserving it. And in the case of my 
own, unfinished project, the 'context' has changed so dramatically that
it would make sense to finish it not to complete it but to *reconstruct*
it as a token from the different world of, oh, 1993. Which brings us to
your next point:

> But, at the risk of sounding like a zealot, this is the typical
> example of how proprietary software platforms, and dependency on them
> [which includes dependency on binary compatibility], will always bite
> you in the end. The problem wouldn't exist if those media works had
> been written as Web CGIs in Perl outputting HTML 2.0, or generating
> JPEGs, for example.

You do sound like a zealot, which is fine -- no judgment of any kind is
needed or appropriate. But it's important to recognize that the kinds 
of standards you mention (Perl, JPEG, etc) are, as you suggest, more 
resilient because they're open -- but only 'more' resilient. Perl is 
not Latin, and JPEGs are not the codex; or maybe they are -- the result
is the same, only the time scale changes. Adhering to these standards,
whosen 'openness' is defined and evaluated in large part against the 
backdrop of dominant proprietary standards, would grant these works at
most a few more decades of compatibility. You could point at something
like I/O/D's Webstalker and lament the fact that it was written as a 
Director Projector -- but that was fact was an explicit aspect of its 
creation in its context and, equally important, one reason it was such
a success in that context. If it had been written in an open standard,
it would have been inaccessible to most of the people who appreciated 
it -- who weren't about to invest in some baroque hack like Tenon's 
MachTen (or better, A/UX). 

Back in the day, as they say, some SoHo artist whose name I've forgotten
made what then seemed like a horribly cynical piece, called "Talent," I
think -- a dozen or so headshots of the art stars of that art season. I
have no idea what the artist who made it thought, but it aged quickly 
and, I think, well -- and though I haven't seen it in close to twenty 
years, I'd venture to say that it's probably one of the finest pieces 
of art that came out of that scene. The particulars about who went on 
to become blue-chip or has-been are less important than the -- to use
the phrase of the day -- 'generative' quality of the distinctions as they
unfold over time. We don't know which standards will be accepted or 
applied in the future; and while it's certainly worth considering which
ones will survive, it's also worth considering how paying too much 
attention to the imagined future historicity of what we make can remove
us from making things here and now. 

Cheers,
T





#  distributed via <nettime>: no commercial use without permission
#  <nettime>  is a moderated mailing list for net criticism,
#  collaborative text filtering and cultural politics of the nets
#  more info: http://mail.kein.org/mailman/listinfo/nettime-l
#  archive: http://www.nettime.org contact: nettime {AT} kein.org