Category Archives: search engines / OPACs

campuscast

Egads…it looks like the web 2.0 / social networking movement is really picking up steam in academia.  There seems to be a new online lecture-cast platform that’s unveiled every week.  Inevitably it’s an encouraging movement, as colleges and universities are realizing they have the ability to capture lightning in a bottle as it strikes on campus.  Librarians are cashing in as well, as increasingly we are the ones who are doing the recording, organizing and disseminating of such content.

The movement toward increased lecture-cast only makes perfect sense.  A special event happens on campus, and with the permission to record and eventually post said lecture, an institution not only contributes to the overall body of knowledge, but also promotes itself and faculty in the process.

Notable platforms

  • YouTube – EDU – With brilliant examples like the periodic table of videos, more and more schools are uploading content, forcing YouTube to organize a separate space for the professional geeks.
  • FORA.tv – Focuses more on the prestigious speakers than the institutions. Important indeed.
  • Academic Earth – currently featuring lectures from Berkeley, Harvard, MIT, Princeton, Stanford, and Yale.
  • iTunes U – Virtual beheamoth that it is, iTunes U is a subset of the iTunes Store contaning massive amounts of lectures, supplementary course info, and stuff of a general academic nature.  Typically free, of course.

media specialists and college librarians the same?

Here’s a recent article from the NYT, talking about information literacy among elementary school students, and the work it takes for media specialists to break through to their “patrons”.  It’s amazing how the perceptions of information literacy and web habits among children mirror those of college students.

It’s an interesting article that details the gamut of issues that librarians are facing, including:

  • Budget cuts
  • Librarians on the front lines battling info illiteracy
  • Dealing with outdated collections and limited funds
  • actually making a difference

Here’s the scary part:

Even teachers find that they learn from Ms. Rosalia. “I was aware that not everything on the Internet is believable,” said Joanna Messina, who began taking her fifth-grade classes to the library this year. “But I wouldn’t go as far as to evaluate the whole site or look at the authors.”

And:

During a lunch period earlier this month, Gagik Sargsyan, 13, slunk into the library and opened a laptop to research a social studies paper on the 1930s and 1940s.

“Have you looked at any books?” Ms. Rosalia asked.

A look of horror came over Gagik’s face. “No,” he said.

Not that surprising, really. But it’s self-evident to regard the stagnation of info seeking behavior among students of all levels, the OPACs, catalogs, databases that aren’t primarily utilized.  It does seem that students are taught little more than to fill in bubbles when not surfing the web. 

the incidental opac

conceptLibrarians, I’ve come to understand, facilitate things.  Just like those late-night, seedy, ever anonymous entrepreneurs on streetcorners and in beer gardens possessing the ability to procure certain items on short notice for other unnamed yet interested parties, librarians too, embrace their responsibility of passing on their coveted contraband of information or that of retrieving such information.

And considering information retrieval, I’m incessantly perplexed with the utter obliviousness users have toward their library catalog.  It’s as if users take pride, relishing a certain sense of entitlement in their lack of curiosity toward navigating library resources.  Hence, the librarian is forced to find new ways to shuffle these students like cattle through the  slaughterhouse of information literacy or competency.

I’m not all that surprised that we now are induced to a vomit-inducing display of flashing lights and multimedia just to get students’ attention.  Should users actually spent five minutes exploring their OPAC (or listening to their librarians), they might actually learn how supremely practical subject headings can be.

Take for example, aquabrowser, a different kind of OPAC designed to display relationships based on searching terms.  My local public library uses it, along with the option of using a more traditional OPAC.  Aquabrowser uses a visual diagram of one’s search terms, highlighting possible misspellings, relationships, translations or thesaurus terms for one’s search.

I personally like it, however I feel it’s designed for the user who has no idea what they’re looking for, wherein I posit the hypothesis that those users are for the most part uncommon. Traditional OPACs will get the user to their items just as fast if not faster assuming they know what they’re looking for.

Users want to know if their materials are already checked out before they want to know what you have.  Therefore, the fact that you have an OPACs is incidental and it will be used primarily when one’s primary request has become unavailable.

Egads, you may be thinking…what is my point anyways?  Having OPACs that visually diagram your search, all supplemental and wondrous as they may be, may not necessarily be more useful than the standard OPACs, though less “dynamic” in the Web 2.0 sense.

Users, particularly college-level users mind you, aren’t familiar with their collections, and thus their OPACs.  I suppose that’s part of what makes us librarians freaks…we willingly, involuntarily befriend our collection regardless of whether a copy of Mall Cop has already been ordered and is on its way. Getting users to use the catalog for its own sake is herculean.

resisting google: not so futile

Not too long ago I mused upon the idea of how some search engine companies are trying to provide more  human interaction when one has an online reference question, by either doing the searching or providing suggestions on how to perform the search.  This quasi virtual reference seems to be catching on, and librarians are suddenly becoming more recognized for the credibility they provide in their reference work.

This sentiment is the impetus for a new project that aims to compete with likes of the great goog, Reference Extract.  The project, an ever-increasing collaboration of libraries, aims to differ from Google in the credibility taken from the shrewd linkages that librarians provide in applying sound information literacy principles. Said better than myself:

Users will enter a search term and get results weighted towards sites most often referred to by librarians at institutions such as the Library of Congress, the University of Washington, the State of Maryland, and over 1,400 libraries worldwide.

The issue of credibility is interesting when compared to the measure of relevancy and popularity Google bases its index on.  The issue of credibility is more fully explained:

In essence linkages between web pages by anyone is replaced by citations to web pages by highly trained librarians in their daily work of answering the questions of scholars, policy makers and the general population. Instead of page rank, the team refers to this as “reference weighting.”

That is to say, it is no great leap to believe that working one-on-one with a librarian would yield highly credible results, but it also appears that gathering the sites librarians point to across these one-on-one interactions and making them searchable continues to yield highly credible results. Further since the librarians answer question on very wide range of topics, their answers can be applied to a general purpose search engine.

I find it clever that the organizers of RefEx measured their index by using the custom search engine provided by Google…beating it at its own game perhaps.

It is important to note that by using the Google Custom Search Engine service the exact same technology was used to search and rank the results, the only thing that varied was that one was an open web search, and one was limited to only those pointed to by reference librarians. So, even outside of the library website context the credibility of librarians is retained.

We may index less pages, but the ones we point to are more informationally literate. One question to walk away from with this: does less material indexed = more reliable?  Philosophically speaking, words like popular, relevant, and usefulness will cause debate; academically speaking, this justifies the librarian’s attempt to wean those frothing, zombie-like patrons away from The Google and more toward our subscribed databases, online resources and guides.  And with RefEx, Google’s helping us do it.

life in googlevision

The great and powerful Goog has now acquired the archived photos from LIFE magazine, and it’s publicly available on each of your interwebs:

The collection includes the entire works of Life photographers Alfred Eisenstaedt, Gjon Mili and Nina Leen. Also available are: the Zapruder film of the Kennedy assassination; Dahlstrom glass plates of New York from the 1880′s; and Hugo Jaeger Nazi-era Germany 1937-1944.

Dawn Bridges, a spokeswoman for TimeInc, the archives in their entirety would be available in the first quarter of next year. She said it was would not just be historical. “We will be adding new things. There will be thousands of new pictures from DC for the inauguration on January 20,” she said.

What’s cool is that according to the article, 97% of the photos (10 million) have never before been seen. Here’s Google’s portal for accessing the photos. A prominent issue now to consider is whether the photos are in the public domain. Obviously, the older ones might just be, but what about the ones less than the 70 or so years it takes for fair use? Pretty groovy for browsing, though.

search engine overload…or overlord?

Seems like search engines have been springing up all over the place.  Soon enough there will be needed search engines to search search engines (oh wait…we already have those). In any case, the emergence of new breed of mechasearchers has me intrigued whether or not Google might be spreading itself a bit too thin with all their gizmos in development.  I’m curious about the avenues that these particular developers are taking so that they just might be the one to slay the great Goog.  Three current avenues are particularly intriguing.

Preserve what little humanity we have left with ChaCha

ChaCha is a company that is building on the idea that it is not so much the technology that is delivering your indexed content as it is the humanoids manipulating the technology.

Thus Spake Zara-chacha:

ChaCha is conversational, fun, and easy to use. Simply ask your question like you are talking to a smart friend and ChaCha’s advanced technology instantly routes it to the most knowledgeable person on that topic in our guide community. Your answer is then returned to your phone as a text message within a few minutes.

Not that it’s necessary to use a live guide as their search engine works perfectly fine, but hooking a live one can be helpful especially if you’re not near a pulsing box of pixellation and you have your phone with you.  Texting your searches seems like all the rage, but mind you, standard rates may apply.

Make it sound as human as possible with Powerset

Taming the beast is the aim of Powerset, the beast being the search technology that cannot understand our queries.  So like ChaCha, there is nothing wrong with us, but that blasted speech sytnax that computers simply can’t understand.  Powerset writes it out for us:

Powerset’s goal is to change the way people interact with technology by enabling computers to understand our language. While this is a difficult challenge, we believe that now is the right time to begin the journey. Powerset is first applying its natural language processing to search, aiming to improve the way we find information by unlocking the meaning encoded in ordinary human language.

So with the intent of not having to resort to technical, complicated search strings, Powerset wants our search results directly related to the flow of our informal speech patterns.  In its infancy, Powerset currently indexes only articles submitted to Wikipedia, though containing several viewing options, references, and citations one would expect from a typical wikipedia entry.

Index early, index often with Cuil

And then there’s Cuil. Apparently created by defectors from the great Goog, these two have started their own search engine, and though like Shaquille O’Neal running a not-so-fast break, it’s definitely gaining momentum. So much so that it boasts possessing the world’s biggest index:

The Internet has grown exponentially in the last fifteen years but search engines have not kept up—until now. Cuil searches more pages on the Web than anyone else—three times as many as Google and ten times as many as Microsoft.

Rather than rely on superficial popularity metrics, Cuil searches for and ranks pages based on their content and relevance. When we find a page with your keywords, we stay on that page and analyze the rest of its content, its concepts, their inter-relationships and the page’s coherency.

Then we offer you helpful choices and suggestions until you find the page you want and that you know is out there. We believe that analyzing the Web rather than our users is a more useful approach, so we don’t collect data about you and your habits, lest we are tempted to peek. With Cuil, your search history is always private.

Very interesting claim as well that Cuil has no interest whatsoever with collecting user data or the habits thereof and indexing by popularity.  In any case, Cuil certainly intends to raise the stakes.

Three different philosophies, three different search engines.