I sat on my tall stool behind the counter in my parents’ music store, looking past my open history textbook to the dirty snow and paper trash blowing down the street in the darkening afternoon. A lone figure shuffled down the opposite sidewalk, past the jewelry store, and stopped on the corner in front of the drug store at the stoplight, his helmeted head cast down, waiting for the traffic light to turn. I scanned a few more paragraphs in my textbook until he entered, heralded by a chorus of automated door chimes and blown in by a gust of frozen air.
Lately, I’ve been waxing romantic about traffic accidents. It has something to do with all the news about driverless cars, also known as autonomous vehicles, or AVs. Since 2015, when Tesla released its Model S, which could park on its own and drive solo on highways, car and tech companies have been hotly competing to achieve the next breakthrough. Now I hear that, by 2020, Google will release a car that has no steering wheel or pedals for accelerating and braking. This prospect sounds terrifying — until you consider that 90 percent of traffic blunders are attributed to human error. With AVs, the techies proclaim, such error will go the way of the Dodo. We are entering an era of Utopian travel, “the accident-free society.” More… “The Accident Free Society”
Jen DeGregorio's writing has appeared in The Baltimore Review, The Collagist, PANK, Perigee (Apogee online), The Rumpus, Third Coast, Spoon River Poetry Review, Women's Studies Quarterly, Yes Poetry and elsewhere. She has taught writing to undergraduates at colleges in New Jersey and New York and is currently a PhD student in English at Binghamton University, State University of New York.
The future isn’t what it used to be. We need new futures.
Science fiction traditionally has had the task of providing us with alternative visions of the future. For the most part, it has done a terrible job. The main reason for its failure is that it assumes global uniformity.
In optimistic visions of the future, there is a liberal and democratic world government, or perhaps an interplanetary federation. In dystopias, there is a single global tyranny. In post-apocalyptic novels and movies set in the aftermath of a nuclear war, nuclear bombs seem to off gone off everywhere in the world, even in places remote from the homelands and allies of the major combatants. More… “The Future of the Future”
Hello, everyone. My presentation today is about the harm that PowerPoint presentations are doing to the way we think and speak. To illustrate the danger, this warning is in the form of a PowerPoint presentation.
Next slide, please.
For nearly two millennia, from Isocrates and Cicero to the 19th century, the art of rhetoric was at the center of the Western tradition of liberal education. The liberally educated citizen was taught to reason logically and to express thoughts in a way calculated to inform and, when necessary, to motivate an audience. More… “PowerPoint Makes Us Stupid”
Can advanced technology make us worse off? I recently had occasion to ponder the question. I was standing in line at a movie theater, when the computer that had replaced the cash register froze up. None of the staff on duty at the theater could fix it. No sales could be made and no tickets issued. I left, with a new appreciation for the old-fashioned mechanical cash register.
This was not the first time I have wondered about what might be called technological regress. Having had bad experiences with inaccessible computer files at work and at home, I now keep multiple paper copies of all important publications and information, in formats which would have been familiar to Gutenberg.
It’s an ambiguous pileasure, but a pleasure nonetheless, to look back at the photographs I took decades ago when I was in my teens and early 20s. Those images make me realize there are moments that seemed so consequential to me, that felt so charged with happiness and meaning, that I felt compelled to make a record of them. They also remind me that I was once young and dewy and that I am not so any longer.
These photographs have a random, desultory quality to them. Each one is like the proverbial insect that got stuck in amber. If things had been different on some particular day 150 million years ago, if a different unsuspecting fly or beetle had blundered its way to its sticky doom, then the chunk of cretaceous petrified tree resin that I hold in my hand would display a different mummified arthropod. It’s the same with all these photos I have of people I didn’t know, or barely knew — friends, or friends of friends, or other young people who frequented the same bars and restaurants long ago.
Most of the photos were taken with a 35mm Olympus camera my father gave me as a birthday present in 1981. When I was 13, I didn’t think to ask for or even to want a fancy camera. I was more interested in getting my own stereo so that I could listen to Blondie 45s in my room. It seemed to me that my father did a perfectly good job of taking family photos with his Nikon camera and I even had an Instamatic that I could resort to in extremis. Still, he wanted me to have a really good camera; he had always loved taking pictures and thought that I might too one day.
I brought the Olympus with me to stave off boredom during the obligatory attendance at my brother’s soccer games, which is the reason why I now have charming photos of the boys I went to high school with wearing their sweaty, grass-stained uniforms. In college, I got in the habit of lugging the camera around on weekends when I left my suburban campus for Boston: on exciting, unpredictable nights when I didn’t quite know where I might go and what might happen. The camera, an encumbrance on the dance floor, provided an excuse to sit out the dancing since the evening’s photographer had the unspoken right to drink rum and cokes on the sidelines while waiting for a picturesque excuse to snap the shutter.
I think the key thing was the literal and figurative weightiness of the camera. I took it out for special events or when I wanted to make ordinary events special. It recorded things, but its presence changed the atmosphere and the images being taken. Thirty years ago, cameras weren’t ubiquitous. Even if the subjects were encountered at random, making photographic images required some commitment. It strikes me that my photos taken back in the 1980’s and 1990’s have a good deal in common with carefully posed daguerreotypes — perhaps more than they do with the throwaway digital images I have on my smartphone today.
For most of history, of course, there were no cameras, no way for ordinary people to record the visual with any measure of verisimilitude. Even after the 1839 introduction of the first camera for professional photographers, the Giroux Daguerreotype, it took another 50 years for George Eastman to bring to market the easy-to-operate, amateur-friendly Kodak. Then, right at the turn of the century, he came out with the small, portable, one-dollar Brownie camera — which made the technology available to just about everyone. Portability is relative, however. I took my childhood Instamatic with me on the junior high field trip from Albany, New York to the Boston area, where I struck a sassy pose in front of Concord’s “rude bridge that arched the flood,” along with pre-teen girls named Andrea, Pam, and Lara. That camera and the low-quality snapshots it tended to produce now seem as archaic as Emerson’s anachronistic poetic diction.
Today I am a librarian at a university and have plenty of opportunities to observe college students. They are, as we all are now, a little trigger-happy when it comes to the cameras on their smartphones. Still, they’ve been taught to be nervous about exposure — and I’m not referring to the amount of light hitting a camera’s sensors. Incriminating photos, posted thoughtlessly or even maliciously, can ruin a reputation or repel a potential employer. They can even go viral. I try to imagine how stultifying it must be to have to worry about the omnipresence of camera phones and Facebook, the fear of being captured chugging beer at a keg party or puffing on a joint in a friend’s dorm room. I’ve added this low-grade but constant anxiety to the list of reasons why I don’t envy today’s college students, along with the high cost of tuition and the dismal job market for recent graduates.
If you’re worried about the potentially compromising future uses of your photographs, it’s easy to forget to question whether those images will actually have a persistent afterlife. My sense is that the longevity of today’s photographic record is in question; digital images are less stable than those printed on high-quality photographic paper. Decades from now, today’s college students might not have a photographic record of their undergraduate years at all.
There was data loss before the internet of course. To use literature as an example, Sappho, whose work is now only available in tantalizingly fragmentary form, is one well-known example. Her poems were recorded on papyrus, a writing material from Egypt that was used in the ancient world as an alternative to more expensive (and, as it happens, more durable) parchment. Archivists and fans of Thomas Pynchon will recognize the characteristic disintegration of papyrus over the centuries as an example of “inherent vice” (that is, the way in which a material’s own chemical makeup leads to its self-destruction).
There’s also data loss due purely to human neglect. In the 17th century, Ben Johnson bemoaned the loss of a painting in his poem “Picture Left in Scotland.” In it, he regrets having entrusted a portrait of himself to a woman who didn’t love him enough to take good care of or appreciate his gift. He expresses regret too that he is no longer young. He seems to have hated the sight of his middle-aged self in the portrait, and hated even more the fact that the woman he was enamored with clearly didn’t care for it either. “My hundred of gray hairs,/Told seven and forty years,” he wrote, adding ruefully: “she cannot embrace/ My mountain belly and my rocky face.”
I feel his pain, having just turned 47 myself.
The carelessness with which the painting was forgotten wounded him, but it wasn’t tragic; the reader senses that aging Ben was probably better off without the heartless little minx. The loss of the current college generation’s photos — images of poignantly young people who will inevitably grow apart from one another — won’t be a tragedy, either, perhaps; as I say, ordinary people have only had access to photography for a bit more than a century, and it’s possible today’s college students will not miss what their distant ancestors never had: a photographic record of the past. Still, I think it’s hard for one generation to accept less than what the previous one had, and I think it is painful to lose what could have been saved. What’s more, I believe there will be a collective data loss that affects millions of people and alters future generations’ ability to know us through our creations and artifacts.
Many commentators have noted the Faustian bargain we’ve made with the likes of Facebook, YouTube, and Instagram: in exchange for a purportedly free service, we have handed over our personal data to corporations. Furthermore, the so-called “culture of free” has eroded the ability of musicians, photographers, and writers (among others) to make a living from their work. Less examined is the inability of free services to preserve our data. The aforementioned internet-based services make money selling data, not preserving it, and obviously corporations exist to make money. This means that they will preserve our images stored on their servers only for as long as it is cost-effective to do so — and not a moment longer. There have been hacks of websites — Target’s for example — that have led to data leaks. But as we wring our hands over the potential for identity theft, we should remember that any hacker who can steal information could, of course, also erase it. (This happened to Wired writer Mat Honan.) Furthermore, time is the ultimate thief of digital information. The slow-motion erosion of papyrus over the centuries is nothing compared to the rapidity of data loss in a digital environment. Files crash, and new iterations of hardware and software can make old data inaccessible.
My own smartphone is filled to capacity with images that I’ve failed to back up, even though I know well how easy it is to lose them. Internet culture has come to favor dissemination over preservation, and it has encouraged users, even vigilant ones like me, do the same. It’s easy for me to post images that can be viewed by millions, while properly saving my images takes effort and even some money. In spite of my hundreds of gray hairs and my 47 years, I, short of time and cash, fecklessly procrastinate rather than immediately create satisfactory backups of everything I own and want to save. Every time I look at my old college photographs I feel grateful to have grown up in a time when careless handling of images didn’t lead to catastrophe — viral images whose spread is impossible to halt or lost photographic treasures that even an expert may not be able to retrieve from oblivion. •
There’s a Steve Jobs movie coming out — yes, another. Here’s a review. And in case you missed them, here’s a list of (most of) the artistic depictions of the iCon.
Fans of JK Rowling rejoice! Humanity is getting closer to a scalable invisibility cloak, according to a study published in the journal Science.
In the age of the iPhone photographer, Wolfgang Tillman is leveraging technology to help him stand out from the crowd. His latest show, “PCR”, is full of photographs that “would not have been possible ten years ago.” •
Exhibition labels – those little placards on the walls beside the thing part of the art – are no longer optional to conscientious looking. In the imagined past, we might have taken in a landscape and breathed, “Isn’t that lovely? Just look at those brushstrokes!” But the priorities of contemporary art have foreclosed on this option. We might even go so far as to define contemporary art as that species of aesthetic work (as opposed to modern, folk, etc.) for which the label is as important as the specimen. Otherwise we risk missing the conceptual forest for the material trees.
Since the 1990s one of the more dominant strains of new work has been a technologically-oriented conceptualism that — if it doesn’t dispose of humanity altogether: the bodily, the gestural, the domestic – finds its importance only in the way those bodies have extended themselves through technology, a technology well-removed from the state of nature, the nuts-and-bolts of existence, anything a patron could begin to brim sentimental over.
Such art is often interested in incorporating materials not typically found in an art gallery, at least not 25 years ago: open source models, industrial units, server platforms, new plastics. At its best, such work lays bare the hidden cables that stitch our new world together, make us suddenly hyper-conscious of the pattern of which we form a part, an invisible part. Exhibition labels are doubly important in such work: without the ideas behind the work, which almost always require spelling-out, we have nothing. More… “Immaterial Worlds”
John Cotter’s first novel Under the Small Lights appeared in 2010 from Miami University Press. A founding editor at the review site Open Letters Monthly, John’s published critical work in Sculpture, Bookforum, and The The Poetry Foundation. Say hi at John [at] JohnCotter [dot] net.
Your heart is racing and your breath feels constricted. You’ve started to sweat around your hairline. You’re not sure if you should call 911, lay down, open a window, or…put down the book.
Have you ever read a book and found yourself at a loss as to what to think and feel?
Yeah. Me neither. But for those who have, consider Sensory Fiction, a wearable device and augmented book now in prototype at MIT. It straps on, and most of its brains ride right between the shoulder blades; it mostly looks like a techno-savvy baby carrier.
As the protagonist’s emotional or physical state changes, so does the reader’s, via ambient light, slight vibrations, and, get this: localized temperature fluctuations and constricting airbags that actually change the reader’s heart rate. The emotional response I’m getting right now, without wearing the device, is: fear. The device has airbags?
Ten thousand is a likable number. It’s rotund and cheerful, both aspirational and accessible. Ten thousand is like the serious but fun kid who sat near you in high school chemistry — not as goofy as 3,850 in the back row, or as aloof and vain as 100,000 up front.
Ten thousand is also how many steps a day you’re told to take if you want to maintain good health. It’s been prescribed like a multivitamin, or those eight, eight-ounce glasses of water we were once instructed to consume every day to stay hydrated and rosy.
And it’s definitely a thing. The American Heart Association recommends 10,000 steps every day. The Kaiser Permanente health group administers its own program, “designed to help you gradually increase your physical activity level and work toward a goal of walking 10,000 steps each day.” If you buy a FitBit, one of the new generation… More…