[Paleopsych] Technology Review: The Infinite Library

Premise Checker checker at panix.com
Fri Apr 15 20:23:47 UTC 2005


The Infinite Library
http://www.technologyreview.com/articles/05/05/issue/feature_library.asp?p=0

    By Wade Roush May 2005

    The Bodleian Library at the University of
    Oxford in England is the only place you are likely to find an Ethernet
    port that looks like a book. Built into the ancient bookcases
    dominating the oldest wing of the 402-year-old library, the brown
    plastic ports share shelf space with handwritten catalogues of the
    university's medieval manuscripts and other materials. Some of the
    volumes are still chained to the shelves, a 17th-century innovation
    designed to discourage borrowing. But thanks to the Ethernet ports and
    the university's effort to digitize irreplaceable books like the
    catalogues--which often contain the only clue to locating an obscure
    book or manuscript elsewhere in the vast library--users of the
    Bodleian don't even need to take the books off the shelves. They can
    simply plug in their laptops, connect to the Internet, and view the
    pertinent pages online. In fact, anyone with a Web browser can read
    the catalogues, a privilege once restricted to those fortunate enough
    to be teaching or studying at Oxford.

    The digitization of the world's enormous store of library books--an
    effort dating to the early 1990s in the United Kingdom, the United
    States, and elsewhere--has been a slow, expensive, and underfunded
    process. But last December librarians received a pleasant shock.
    Search-engine giant Google announced ambitious plans to expand its
    "Google Print" service by converting the full text of millions of
    library books into searchable Web pages. At the time of the
    announcement, Google had already signed up five partners, including
    the libraries at Oxford, Harvard, Stanford, and the University of
    Michigan, along with the New York Public Library. More are sure to
    follow.

    Most librarians and archivists are ecstatic about the announcement,
    saying it will likely be remembered as the moment in history when
    society finally got serious about making knowledge ubiquitous.
    Brewster Kahle, founder of a nonprofit digital library known as the
    Internet Archive, calls Google's move "huge....It legitimizes the
    whole idea of doing large-volume digitization."

    But some of the same people, including Kahle, believe Google's efforts
    and others like it will force libraries and librarians to reëxamine
    their core principles--including their commitment to spreading
    knowledge freely. Letting a for-profit organization like Google
    mediate access to library books, after all, could either open up
    long-hidden reserves of human wisdom or constitute the first step
    toward the privatization of the world's literary heritage. "You'd
    think that if libraries are serious about providing access to
    high-quality material, the idea of somebody digitizing that stuff very
    quickly--well, what's not to like?" says Abby Smith, director of
    programs for the Council on Library and Information Resources, a
    Washington, DC, nonprofit that helps libraries manage digital
    transformation. "But some librarians are very concerned about the
    terms of access and are very concerned that a commercial entity will
    have control over materials that libraries have collected."

    They're also concerned about the book business itself. Publishers and
    authors count on strict copyright laws to prevent copying and reuse of
    their intellectual property until after they've recouped their
    investments. But libraries, which allow many readers to use the same
    book, have always enjoyed something of an exemption from copyright
    law. Now the mass digitization of library books threatens to make
    their content just as portable--or piracy prone, depending on one's
    point of view--as digital music. And that directly involves libraries
    in the clash between big media companies and those who would like all
    information to be free--or at least as cheap as possible.

    Whatever happens, transforming millions more books into bits is sure
    to change the habits of library patrons. What, then, will become of
    libraries themselves? Once the knowledge now trapped on the printed
    page moves onto the Web, where people can retrieve it from their
    homes, offices, and dorm rooms, libraries could turn into lonely
    caverns inhabited mainly by preservationists. Checking out a library
    book could become as anachronistic as using a pay phone, visiting a
    travel agent to book a flight, or sending a handwritten letter by
    post.

    Surprisingly, however, most backers of library digitization expect
    exactly the opposite effect. They point out that libraries in the
    United States are gaining users, despite the advent of the Web, and
    that libraries are being constructed or renovated at an unprecedented
    rate (architect Rem Koolhaas's Seattle Central Library, for example,
    is the new jewel of that city's downtown). And they predict that
    21st-century citizens will head to their local libraries in even
    greater numbers, whether to use their free Internet terminals, consult
    reference specialists, or find physical copies of copyrighted books.
    (Under the Google model, only snippets from these books will be
    viewable on the Web, unless their authors and publishers agree
    otherwise.) And considering that the flood of new digital material
    will make the job of classifying, cataloguing, and guiding readers to
    the right texts even more demanding, librarians could become busier
    than ever.

    "I chafe at the presumption that once you digitize, there is nothing
    left to do," says Donald Waters, a former director of the Digital
    Library Federation who now oversees the Andrew W. Mellon Foundation's
    extensive philanthropic investments in projects to enhance scholarly
    communication. "There is an enormous amount to do, and digitizing is
    just scratching the surface."

    Digitization itself, of course, is no small challenge. Scanning the
    pages of brittle old books at high speed without damaging them is a
    problem that's still being addressed, as is the question of how to
    store and preserve their content once it's in digital form. The Google
    initiative has also amplified a long-standing debate among librarians,
    authors, publishers, and technologists over how to guarantee the
    fullest possible access to digitized books, including those still
    under copyright (which, in the United States, means everything
    published after January 1, 1923). The stakes are high, both for Google
    and for the library community--and the technologies and business
    agreements being framed now could determine how people use libraries
    for decades to come.

    "Industry has resources to invest that we don't have anymore and never
    will have," points out Gary Strong, university librarian at the
    University of California, Los Angeles, which has its own aggressive
    digitization programs. "And they've come to libraries because we have
    massive repositories of information. So we're natural partners in this
    venture, and we all bring different skills to the table. But we're
    redefining the table itself. Now that we're defining new channels of
    access, how do we make sure all this information is usable?"

    Breaching the Walls
    Even for authorized users, access to the Bodleian Library's seven
    million volumes is anything but instant. If you are an Oxford
    undergraduate in need of a book, you first send an electronic request
    to a worker in the library's underground stacks. (Before 2000 or so,
    you would have handed a written request slip to a librarian, who would
    have relayed it to the stacks via a 1940s-era network of pneumatic
    tubes.) The worker locates the book in a warren of movable shelves (a
    space-saving innovation conceived in 1898 by former British prime
    minister William Gladstone) and places it in a plastic bin. An
    ingenious system of conveyor belts and elevators, also built in the
    1940s, carries the bin back to any of seven reading rooms, where it is
    unpacked, and the book is handed over to you.

    The process can take anywhere from 30 minutes to several hours. But
    once you finally have the book, don't even think about taking it back
    to your dorm room for further study. The Bodleian is a noncirculating
    legal deposit library, meaning that it is entitled to a free copy of
    every book published in the United Kingdom and the Republic of
    Ireland, and it guards those copies jealously. The library takes in
    tens of thousands of books every year, but the legend is that no book
    has ever left its walls.

    But a digital book needn't be loaned out to be shared. And Oxford's
    various libraries have already created digital images of many of their
    greatest treasures, from ninth-century illuminated Latin manuscripts
    to 19th-century children's alphabet books. Most of these images can be
    examined at high resolution on the Web. The only catch is that
    scholars have to know what they're looking for in advance, since very
    few of the digital pages are searchable. Optical character recognition
    (OCR) technology cannot yet interpret handwritten script, so exposing
    the content of these books to today's search engines requires typing
    their texts into separate files linked to the original images. A
    three-person team at Oxford, in collaboration with librarians at the
    University of Michigan and 70 other universities, is doing just that
    for a large collection of early English books, but the entire effort
    produces searchable text for only 200 books per month. At that rate,
    making a million books searchable would take more than 400 years.

    That's where Google's resources will make a difference. Susan
    Wojcicki, a product manager at Google's Mountain View, CA, campus and
    leader of the Google Print project, puts it bluntly: "At Google we're
    good at doing things at scale."

    Google has already copied and indexed some eight billion Web pages,
    which lends credibility to its claim that it can digitize a big chunk
    of the 60 million volumes (counting duplicates) held by Harvard,
    Oxford, Stanford, the University of Michigan, and the New York Public
    Library in a matter of years. It will be a complex task, but one that
    is in some ways familiar for the company. "It's not just feeding the
    books into some kind of digitization machine, but then actually taking
    the digital files, moving those files around, storing them,
    compressing them, OCR-ing them, indexing them, and serving them up,"
    points out Wojcicki. "At that point it becomes similar to all of
    Google's other businesses, where we're managing large amounts of
    data." But the entire project, Wojcicki admits, hinges on those
    digitization machines: a fleet of proprietary robotic cameras, still
    under development, that will turn the digitization of printed books
    into a true assembly-line process and, in theory, lower the cost to
    about $10 per book, compared to a minimum of $30 per book today.

    Neither Google nor its partner libraries have announced exactly how
    the process will work. But John Wilkin, associate university librarian
    at the University of Michigan, says it will go something like this:
    "We put a whole shelfful of books onto a cart, keeping the order
    intact. We check them out by waving them under a bar code reader.
    Overnight, software takes all the bar codes, extracts machine-readable
    records from the university's electronic catalogue, and sends the
    records to Google, so they can match them with the books. Then we move
    the cart into Google's operations room."

    This room will contain multiple workstations so that several books can
    be digitized in parallel. Google is designing the machines to minimize
    the impact on books, according to Wilkin. "They scan the books in
    order and return the cart to us," he continues. "We check them back in
    and mark the records to show they've been scanned. Finally, the
    digital files are shipped in a raw format to a Google data center and
    processed to produce something you could use."

    The Book Web
    Exactly how readers will be able to use the material, however, is
    still a bit foggy. Google will give each participating library a copy
    of the books it has digitized while keeping another for itself.
    Initially, Google will use its copy to augment its existing Google
    Print program, which mixes relevant snippets from recently published
    books into the usual results returned by its Web search tool. A user
    who clicks on a Google Print result is presented with an image of the
    book page containing his or her keyword, along with links to the sites
    of retailers selling the print version of the book and keyword-related
    ads sold to the highest bidders through Google's AdSense program.

    Does it bother librarians that Moby-Dick might be served up alongside
    an ad for the latest Moby CD? "To say we haven't worried about it
    would be wrong," says Wilkin. "But Google has a `good citizen'
    profile. The way they use AdSense doesn't trouble me. And if suddenly
    access were controlled, and there was a cost to view the materials, we
    could still offer them for free ourselves, or at least the
    out-of-copyright materials."

    In fact, Google may put the entire texts of these public-domain
    materials online itself. In the future, Google could even use those
    materials to create a kind of literary equivalent of the Web, says
    Wojcicki. "Imagine taking the whole Harvard library and saying, `Tell
    me about every book that has this specific person in it.' That in
    itself would be very powerful for scholars. But then you could start
    to see linkages between books"--that is, which books cite other books,
    and in what contexts, in the same way that websites refer to other
    sites through hyperlinks. "Just imagine the power that that would
    bring!"

    (Wojcicki's example shows how history can, indeed, come full circle.
    Google founders Larry Page and Sergey Brin developed BackRub, the
    predecessor to the Google search engine, while working on an early
    library digitization project at Stanford that was funded in part by
    the National Science Foundation's Digital Libraries Initiative. And
    PageRank, Google's core search algorithm, which orders sites in search
    results based on the number of other sites that link to them, is
    simply a computer scientist's version of citation analysis, long used
    to rate the influence of articles in scholarly print journals.)

    The Michigan library, says Wilkin, may do whatever it likes with the
    digital scans of its own holdings--as long as it doesn't share them
    with companies that could use them to compete with Google. Such
    limitations may prove uncomfortable, but most librarians say they can
    live with them, considering that their holdings wouldn't be digitized
    at all without Google's help.

    Closed Doors?
    But others are more cautious about the leap Google's partner libraries
    are taking. Brewster Kahle, who is often described as an inspiring
    visionary and sometimes as an impractical idealist, founded the
    nonprofit Internet Archive in 1996 under the motto "universal access
    to human knowledge." Since then, the archive has preserved more than a
    petabyte's worth of Web pages (a petabyte is a million gigabytes),
    along with 60,000 digital texts, 21,000 live concert recordings, and
    24,000 video files, from feature films to news broadcasts. It's all
    free for the taking at www.archive.org, and as you might guess, Kahle
    argues that all digital library materials should be as freely and
    openly accessible as physical library materials are now.

    That's not such a radical idea; free and open access is exactly what
    public libraries, as storehouses of printed books and periodicals,
    have traditionally provided. But the very fact that digital files are
    so much easier to share than physical books (which scares publishers
    just as MP3 file sharing scares record companies) could lead to limits
    on redistribution that prevent libraries from giving patrons as much
    access to their digital collections as they would like. "Google has
    brought us to a tipping point that could define how access to the
    world's literature may proceed," Kahle says.

    In Kahle's view, every previous digitization effort has followed one
    of three paths; with a bit of oratorical flourish, he calls them Door
    One, Door Two, and Door Three. (Kahle acknowledges up front that his
    picture is simplified, and that these aren't necessarily the only
    paths open to libraries today.)

    Door One, says Kahle, is epitomized by Corbis, an image-licensing firm
    owned by Microsoft founder Bill Gates. Since the early 1990s, Corbis
    has acquired rights to digital reproductions of works from the
    National Gallery of London, the State Hermitage Museum in St.
    Petersburg, Russia, the Philadelphia Museum of Art, and more than 15
    other museums. In some cases, it's now impossible to use these images
    without paying Corbis. "This organization got its start by digitizing
    what was in the public domain and essentially putting it under private
    control," says Kahle. "The same thing could happen with digital
    literature. In fact, it's the default case."

    Behind Door Two, parallel public and private databases coexist
    peacefully. Here Kahle cites the Human Genome Project, which
    culminated in two versions of the DNA sequence of the human genome--a
    free version produced by government-funded scientists and a private
    version produced by Rockville, MD-based Celera Genomics and used by
    pharmaceutical companies to identify new drug candidates. The model
    has worked well in genomics, and Google seems to be setting out on a
    similar path, as it keeps one copy of each library's collection for
    itself and gives away the other. Kahle worries, however, that the
    restrictions Google imposes on libraries will prevent them from
    working with other companies or organizations to disseminate digital
    texts. Libraries might be barred, for example, from contributing
    material to projects such as the Internet Archive's Bookmobile, a van
    with satellite Internet access that can download and print any of
    20,000 public-domain books.

    Door Three, Kahle's favorite, hinges on new
    partnerships in which private companies offer commercial access to
    digital books while public entities, such as libraries, are allowed to
    provide free access for research and scholarship. Here his main
    example is the Internet Archive's collaboration with Alexa, a company
    founded by Kahle himself in 1996 and sold to Amazon in 1999. Alexa
    ranks websites according to the traffic they attract, and its servers,
    like Google's, constantly crawl the Internet, making copies of each
    page they find. But after six months, Alexa donates those copies to
    the Internet Archive, which preserves them for noncommercial use.
    "Jeff [Bezos, Amazon's CEO] was okay with the idea that there are some
    things you can exploit for commercial purposes for a certain amount of
    time, and then you play the open game," says Kahle. "Libraries and
    publishing have always existed in the physical world without damaging
    each other; in fact they support each other. What we would like to see
    is this tradition not die with this digital transformation."

    So which alternative comes closest to Google's plans? Google is no
    Corbis, says Wojcicki, but is nonetheless limited in what it can
    share. "Door One was never our intention, nor is it even practical,"
    she says. "And we can't do Door Three, because we're not the rights
    holders for much of this material. So Door Two is probably where we're
    headed. We're trying to be as open as possible, but we need to hold to
    our agreements with different parties."

    Precisely to avoid questions about copyright, Oxford librarians have
    decided that only 19th- and early 20th-century books will be handed
    over to Google for digitization. "Some of the other libraries,
    including Harvard, have agreed to have some in-copyright material
    digitized," says Ronald Milne, acting director of the Bodleian
    Library. "They are quite brave in taking it on. But we didn't
    particularly want to go there, because it's such a hassle, and we
    didn't want to get on the wrong side of the book laws."

    At the same time, though, the American Library Association is one of
    the loudest advocates of proposed legislation to reinforce the "fair
    use" provisions of federal copyright law, which entitle the public to
    republish portions of copyrighted works for purposes of commentary or
    criticism. And two of Google's partner universities--Harvard and
    Stanford--are also supporters of the Chilling Effects Clearinghouse, a
    website that monitors allegations of copyright infringement brought
    against webmasters, bloggers, and other online publishers under the
    controversial Digital Millennium Copyright Act (DMCA) of 1998. Mass
    digitization may eventually force a redefinition of fair use, some
    librarians believe. The more public-domain literature that appears on
    the Web through Google Print, the greater the likelihood that citizens
    will demand an equitable but low-cost way to view the much larger mass
    of copyrighted books. "I think this will be another piece of good
    pressure, another factor in the whole debate over the DMCA," says
    Wilkin.

    The Mixing Chamber
    If you're over 30, today's libraries are probably nothing like the
    ones you remember from childhood. Enter any major library today and
    you'll find an armory of computers and a platoon of specialists, from
    the reference librarians who are expert at accessing online resources,
    to the acquisitions officers who decide which books, CDs, DVDs, and
    subscriptions to purchase, to the computer geeks who keep the
    building's network running.

    Digitization and the growing power of the Internet are making all of
    these people's jobs more complex. Acquisitions experts, for example,
    can no longer just rely on the traditional quality filter imposed by
    the publishing industry; they must evaluate a much larger mass of
    material, from newly digitized print books to the millions of Web
    pages, blogs, and news sites that are born digital. "On the Internet,
    publishing is a promiscuous activity," observes Abby Smith of the
    Council on Library Information and Resources. "Libraries are confused
    and challenged about how to collect and select from that material."

    Then there are the problems of cataloguing and preserving digital
    holdings. Without the proper "metadata" attached--author, publisher,
    date, and all the other information that once appeared in libraries'
    physical card catalogues--a digital book is as good as lost. Yet
    creating this metadata can be laborious, and no international standard
    has emerged to govern which kinds of data should be recorded. And
    considering the limited life span of each new data format or
    electronic storage medium (have you used a floppy disk lately?),
    keeping digital materials alive for future generations will,
    ironically, be much more costly and complicated than simply leaving a
    paper book on a library shelf.

    But even if every book is reduced to a few megabytes of 1s and 0s
    residing on some placeless Web server, libraries themselves will
    probably endure. "There is no one in the field of librarianship who
    thinks the library is disappearing as a physical space," says Smith.
    Seattle's exuberant new Central Library, for example, is built around
    a four-story spiral ramp that enables an unprecedented immediacy of
    access to its physical book collection. But at the same time, the
    library provides 400 public-use computers (compared to 75 in the
    library that previously occupied the site), buildingwide Wi-Fi access,
    and a high-tech "mixing chamber" where an interdisciplinary reference
    team uses an array of print and electronic resources to answer
    patrons' questions. More than 1.5 million people visited the new
    library in 2004--almost three times the entire population of Seattle.

    "The real question for libraries is, what's the `value proposition'
    they offer in a digital future?" says Smith. "I think it will be what
    it has always been: their ability to scan a large universe of
    knowledge out there, choose a subset of that, and gather it for
    description and cataloguing so people can find reliable and authentic
    information easily." The only difference: librarians will have a much
    bigger universe to navigate.

    Stephen Griffin, the former director of the National Science
    Foundation's Digital Libraries Initiative (a Clinton-era project that
    funds a variety of university computer-science studies on managing
    electronic collections), takes a slightly different view. Ask him how
    he thinks libraries will function in 2020 or 2050--once Google or its
    successors have finished digitizing the world's printed knowledge--and
    he answers from the reader's point of view. "The question is, how will
    people feel when they walk into libraries," he says. "I hope they feel
    the same--that this is a very welcoming place that is going to help
    them to find information that they need. As we bring more technology
    in, the notion of libraries as places for books may change a bit. But
    I hope people will always find them a comfortable place for thinking."


More information about the paleopsych mailing list