At IBM, That Google Thing Is So Yesterday
By JAMES FALLOWS
Published: December 26, 2004
UDDENLY, the computer world is interesting again. The last three months of 2004 brought more innovation, faster, than users have seen in years. The recent flow of products and services differs from those of previous hotly competitive eras in two ways. The most attractive offerings are free, and they are concentrated in the newly sexy field of “search.”
Google, current heavyweight among systems for searching the Internet, has not let up from its pattern of introducing features and products every few weeks. Apart from its celebrated plan to index the contents of several university libraries, Google has recently released “beta” (trial) versions of Google Scholar, which returns abstracts of academic papers and shows how often they are cited by other scholars, and Google Suggest, a weirdly intriguing feature that tries to guess the object of your search after you have typed only a letter or two. Give it “po” and it will show shortcuts to poetry, Pokémon, post office, and other popular searches. (If you stop after “p” it will suggest “Paris Hilton.”) In practice, this is more useful than it sounds.
Microsoft, heavyweight of the rest of computerdom, has scrambled to catch up with search innovations from Google and others. On Dec. 10, a company official made a shocking disclosure. For years Microsoft had emphasized the importance of “WinFS,” a fundamentally new file system that would make it much easier for users to search and manage information on their own computers. Last summer, the company said that WinFS would not be ready in time for inclusion with its next version of Windows, called Longhorn. The latest news was that WinFS would not be ready even for the release after that, which pushed its likely delivery at least five years into the future. This seemed to put Microsoft entirely out of the running in desktop search. But within three days, it had released a beta version of its new desktop search utility, which it had previously said would not be available for months.
Meanwhile, a flurry of mergers, announcements and deals from smaller players produced a dazzling variety of new search possibilities. Early this month Yahoo said it would use the excellent indexing program X1 as the basis for its own desktop search system, which it would distribute free to its users. The search company Autonomy, which has specialized in indexing corporate data, also got into the new competition, as did Ask Jeeves, EarthLink, and smaller companies like dTSearch, Copernic, Accoona and many others.
I have most of these systems running all at once on my computer, and if they don’t melt it down or blow it up I will report later on how each works. But today’s subject is the virtually unpublicized search strategy of another industry heavyweight: I.B.M.
Last week I visited the Thomas J. Watson Research Center in Hawthorne, 20 miles north of New York, to hear six I.B.M. researchers describe their company’s concept of “the future of search.” Concepts and demos are different from products being shipped and sold, so it is unfair to compare what I.B.M. is promising with what others are doing now. Still, the promise seems great.
Two weeks before our meeting, I.B.M. released OmniFind, the first program to take advantage of its new strategy for solving search problems. This approach, which it calls unstructured information management architecture, or UIMA, will, according to I.B.M., lead to a third generation in the ability to retrieve computerized data. The first generation, according to this scheme, is simple keyword match – finding all documents that contain a certain name or address. This is all most desktop search systems can do – or need to do, because you’re mainly looking for an e-mail message or memorandum you already know is there. The next generation is the Web-based search now best performed by Google, which uses keywords and many other indicators to match a query to a list of sites.
I.B.M. says that its tools will make possible a further search approach, that of “discovery systems” that will extract the underlying meaning from stored material no matter how it is structured (databases, e-mail files, audio recordings, pictures or video files) or even what language it is in. The specific means for doing so involve steps that will raise suspicions among many computer veterans. These include “natural language processing,” computerized translation of foreign languages and other efforts that have broken the hearts of artificial-intelligence researchers through the years. But the combination of ever-faster computers and ever-evolving programming allowed the systems I saw to succeed at tasks that have beaten their predecessors
Please continue to read here:
Please login or Register to submit your answer