The great thing about HTML is that it’s so flexible and offers so many ways to do things. The worst thing about HTML is that it’s so flexible and offers so many ways to do things. I’ve looked at a lot of websites and I still see people doing things new ways.
An issue that’s often common to many websites is when a page on a site can be found at more than one URL. This might be done by a site owner for a number of reasons, and in a number of ways. It might be an issue related to a content management system that’s being used as well.
A patent application published by Google explores how the search engine might recognize when it finds a URL through a web crawl and another URL through a feed, such as a product feed, with both URLs referring to the same page, but those URLs are structured differently.
This seems like potentially a lot of work to me, and the patent filing has me shaking my head that Google might use resources to figure out duplicated content on a site, even if it potentially might enable the search engine to understand URLs and associated products and other information that it might identify better.
Google was granted a patent this week that describes how web sites might be given quality ratings, based upon a model that looks at human ratings for a sample set of sites, and web site signals from those sites.
The patent tells us that the advantage of such an approach would be to:
Provide greater user satisfaction with search engines
Return sites having a higher quality rating than a certain threshold
Ranking sites appearing in search results based upon quality
Identifying quality sites without having a human view the site first
This patent was originally filed in 2008, and the use of quality signals sound similar to what Google has shared with us regarding the Panda Update. It’s more of a search quality “improvement” than a web spam penalty.
The patent uses blogs as a type of site that it can be applied to within its claims and description section. One of the inventors, Christopher C. Pennock was a Senior Software Engineer on Google Blog Search, according to an early 2009 SMX Session with him which discusses ranking signals in Blog Search.
On May 1st, Google’s Head of Webspam Matt Cutts published a video in his series of Google Webmaster Help videos, answering the question, “What’s the latest SEO misconception that you would like to put to rest?”
For some reason, Matt decided to focus upon patents, with a video about people possibly placing too much faith in what is uncovered in patents related to search engines. To a degree, I agree with his response, but I was reached out to by a number of people who saw the video as something aimed specifically at me, since I write about search related patents so often. I felt that I had no choice but to respond. Here’s the video from Matt:
A transformation was triggered at Google with their announcement of the Knowledge Graph in the Official Google Blog post, Introducing the Knowledge Graph: things, not strings. That transformation was one less concerned with matching keywords, and more concerned with matching concepts, understanding entities, and bringing knowledge about entities to searchers in knowledge panels next to search results.
Google published a patent application last week that describes the knowledge panels that appear next to search results as part of the new knowledge graph. Here’s the video that accompanied the post (note the reference to a “panel” in the presentation):
Google acquired the company Wavii for a little more than $ 30 Million in April. There was some speculation that Wavii was an effort to match Yahoo’s purchase of Summly, which summarizes news from the Web.
A Wavii app did do just that – acquired and summarized news from the Web. When Wavii emerged from stealth mode, it was touted as a personalized news aggregator based upon topics rather than keywords. The app closed down with Google’s acquisition of the company, and instead of providing news aggregation services, it appears that the technology will help fuel Google Now, Google’s Knowledge Base, and Google Glass, according to the TechCrunch article linked above.
One of the more interesting discussions about Google Glass I’ve seen recently was in a forum where one of the participants was describing his own homemade version of Google Glass, which he named “Flass” (if someone at Google happens to be reading this, you should send him a pair of Google Glass, just because.) What was really interesting was that he was using a MyVu display in his clone.
I call it interesting because Google seems to have acquired a number of the patents from The MicroOptical Corporation, which was the predecessor to MyVu. MyVu no longer appears to be in business, and according to LinkedIn, the Founder and CEO and CTO of MyVue is now the Director of Operations at Google X. Here’s a view of one of the pairs of glasses created by MyVue (MicroOptical):
In an ideal world, your site architecture should be set up so that search engine crawlers are only able to visit each page of your site at one web address, and no more. You may be laughing, but when Google sends you the “I give up, your site has too many URLs” message in Google Webmaster Tools, you won’t be then. Seriously.
Keep Colors and Sizes Together
If you create multiple product pages where the only thing different is offering the product in red or green or blue, or small or medium or large, you are creating too many pages. True when you decide to let “email a friend” pages get indexed, and “Add to my wishlist,” and “Compare Products” and other pages that Google doesn’t want in its index either.