Somewhere out there is a universe that looks exactly like this one, and appears to run exactly like this one. Except something’s a little different. A little off. It’s as if search engines took a left turn instead of a right turn, back in the early 2000s. Instead of using only using meta descriptions and possibly body text from web pages for descriptive text, or snippets, for those pages in search results, they learned a new trick. Imagine that the content surrounding anchor text in a link to a page was collected and evaluated based upon a quality score, and that this associated and usually descriptive text was used to generate snippets instead?
My thought on the possibility is that often anchor text doesn’t do the best job of describing a page, and often links to a page are from a third party who might not have the same interest in writing text that might make a good snippet for a page. But, Google filed a patent for such an approach back in 2003. And it was granted this week – so they pursued what was described within the patent for over a decade as well. The patent does mention that headings on pages might also be used as potential snippets for pages, and provide the following example: “Computers > Algorithms > Compression”. But that’s a small part of the patent. They don’t limit it to anchor text that a site might provide itself, like in breadcrumb trail navigation for a page.
There’s also a part to this approach that recognizes that many pages have more than one link to them, so a choice would need to be made as to the best “snippet” to show.
My neighbor has run over my last two phonebooks, and rendered them virtually unusable. We share the same driveway, and it appears that running over phonebooks, and then backing up to make sure they are really dead has officially become a custom in Virginia, or at least in my neighborhood. It’s OK though, since I can’t remember the last time I’ve used a phone book. I may have a couple of times earlier this century, but I’m not sure. I definitely haven’t used one in in the past couple of years (my neighbor keeps killing them).
On the Fourth of July, Apple published a patent application that describes Routes based on User Ratings and Real Time Accident Reporting. Both Apple and Google have been using GPS information to monitor and report upon gridlock and traffic speed estimates, but imagine both providing a richer and fuller social experience involving the world around them. Imagine being able to choose different routes and see social annotations added to different options on those routes. Here’s a screenshot from Apple’s patent filing:
Google was granted an updated version of a patent this week that looks at how the search engine might use directories in URL structures to help it better understand the categories on a Web site, and to categorize new pages and directories that might be added to a site. The patent tells us that this might enable the search engine to add supplemental information to pages, such as advertisements that fall within the categories displayed upon the site.
Some other patents I’ve written about in the past shows that the search engines might be doing more with categories than just deciding upon which ads to show on a page
Imagine that you have a site about car parts, and you decided to organize the pages of the site first by car make, so the main categories on your site are different brands, and your second level of directories is organized by car models. You might then have sub-sub-categories that are organized by different systems within cars, such as “electrical,” “transmission,” “cooling,” “suspension,” and so on. URLs for a couple of your pages might look like:
In the very near future, you may be able to perform searches at Google without bothering to type or speak a query. Instead, you might be able to just shake your phone, or hold down a button for a certain amount of time, and tell your phone something like “search now”. Known as parameterless searches, this type of search can depend upon the context within which the search is performed.
For instance, imagine being driven to work at 50mph, and you shake your phone. It tells you that there’s congestion ahead, and offers an alternative route.Or it shows you a map with color-coded traffic information for different streets nearby according to traffic conditions. Or, you may have an appointment with a client made by email and included on your calendar, and you want to find and check the email to make sure that you have the right phone number. It could show the number and offer to make the call on your behalf. If you regularly take a train at around 8:00 am on weekday mornings, shaking your phone at 7:50 am might trigger a realtime schedule for the rails.
Context information for a parameterless search could include things such as:
Will Google Plus morph into a social version of eBay, where people can post things for sale, write reviews of their belongings and show them off to their circles?
A few days ago, the Google File System Blog reported upon a new application from Google, which it referred to by the name Google Mine (as in, “it’s mine,” rather than as in Google mining more information about your life). The post also mentions a version for Android on Google’s internal Play store. The post tells us that the app allows you to be pretty active about posting your belongings.
As you can see, the service lets you enter a lot of information about your objects. For example, you can change the status of an object to “lent”, “given away”, “got it back”, “lost it”, “had in the past”. You can post videos about the object, write reviews, add it to a wishlist and maybe others can buy it using Google Shopping. You can also check popular items and the items others have shared.
Last week, Google was granted a number of patents exploring different aspects of how documents on the Web might be ranked in part based upon topics identified for those documents and the expertise and/or authority of authors involved in the creation of the documents. The process also describes how Google might use different methods to determine the authority of multiple authors who may have worked to create the documents.
This sounds very similar to some statements that Google’s Matt Cutts made at the start of May in a video about What should we expect in the next few months in terms of SEO for Google?
I’ve written about Google Bombs in the past, and how a bio page featuring President George Bush ranked highly on a search for “Miserable Failure” as a result of a Google Bomb, in a post from 2011 titled How a Search Engine Might Fight Googlebombing.
In a post from earlier today, Nemek Nowaczyk wrote a post Google Bombing the Knowledge Graph: Who’s a Liar? He noticed that on a search for “liar” (in Polish), Poland’s Prime Minister Donald Tusk appears in the knowledge base results for the search. Nemek sent me an email with a link to his post. Within seconds, I was typing one of the better known English language Google bomb phrases into a Google search, with a guess as to what I would see there.
Ok, so the top knowledge base result on a search for “miserable failure” wasn’t George Bush. But a smiling George Bush was close enough to be a “see results about” disambiguation knowledge panel result.
The great thing about HTML is that it’s so flexible and offers so many ways to do things. The worst thing about HTML is that it’s so flexible and offers so many ways to do things. I’ve looked at a lot of websites and I still see people doing things new ways.
An issue that’s often common to many websites is when a page on a site can be found at more than one URL. This might be done by a site owner for a number of reasons, and in a number of ways. It might be an issue related to a content management system that’s being used as well.
A patent application published by Google explores how the search engine might recognize when it finds a URL through a web crawl and another URL through a feed, such as a product feed, with both URLs referring to the same page, but those URLs are structured differently.
This seems like potentially a lot of work to me, and the patent filing has me shaking my head that Google might use resources to figure out duplicated content on a site, even if it potentially might enable the search engine to understand URLs and associated products and other information that it might identify better.