Web Search Engines: Part 1 and 2
I never realized how much space search engines needed to perform searches. Granted, I realize how many web sites they search, but the sheer amount of space needed is impressive, especially the 100 terabyte figure the articles mentions.
I'm a bit confused about how the hashing function works. How many computers are used in a major search engine? For example, how many computers does Google use for its search functions?
Even though the article was supposed to be about searching in common terms, but I still don't really understand web search engines. I don't think that understanding all the nuts and bolts behind web search engines is important, though. As long as I understand how to search using that web search engine, I'll be successful.
Current developments and future trends for the OAI Protocol for Metadata Harvesting
Why does the e-print archives community care so much about the Open Archive Initiative? What sort of needs did the OAI meet?
I thought the idea of the Sheet Music Consortium was really neat. My husband plays two instruments, so he's always looking for new music. Their idea of creating a service provider that would allow users to search for sheet music is really useful.
White Paper: The Deep Web: Surfacing Hidden Value
Like the first reading, this reading also contained information about how much data can be found online. It's amazing how many deep websites exist and how much information they contain.
I found this article to be slightly more understandable than the first article. However, it's very clearly an advertisement. I know the preface mentions the advertisementy nature of the of the article, but it should be better noticed.
I was also very interested in seeing what websites were part of the deep web. I was surprised that Amazon, UPS, eBay and MP3.com were on the list. I expected most of the websites on the list to be government websites. Those ones were a surprise.
No comments:
Post a Comment