AC Hoteles des-baneada

Confirmado. 10 días después del baneo de AC Hoteles, Google ha rehabilitado sus páginas en el buscador. Aunque 10 días de penalización en un sector (hotelero) en el que los clicks se pagan a más de 2 euros no parece demasiado… De hecho, tengo mi teoría de que el hecho de que la historia del baneo de AC Hoteles haya corrido como la pólvora por internet haya hecho más por el posicionamiento de AC Hoteles que las malas prácticas de la empresa de SEO que la cadena hotelera contrató. Algo que ya vi en el famoso caso del baneo de BMW… ¿Y si fuera todo una estrategia de link-baiting con el propósito de ganar links a mansalva? Aunque arriesgada (pocos clientes aceptarían, por el perjuicio causado a la imagen de marca), sería ciertamente efectiva. Por muchos nofollow que los bloggers pongan…

AC Hoteles baneada de Google

La web de AC Hoteles ha sido baneada por Google por emplear técnicas “black hat” como cloaking y noscript. BlogOnTravel explica relativamente bien por qué.

HotelBlog explica muy bien las causas aquí.

En Noviembre se baneó al ISP Acens (des-baneada dos semanas después)

La empresa de SEO responsable de la penalización es muy conocida (éticamente es feo decir su nombre), y es miembro de Sempo, así que queda claro que ser miembro de Sempo no es garantía de que no se usen técnicas black hat. De hecho, por lo que yo sé Sempo sólo te exije que pagues una cuota para permitirte poner su sello en tu web, por tanto no es ningún sello de calidad aunque en muchas webs de empresas de marketing online lo pongan con esa intención.

21 Great SEO Tips From Google’s Matt Cutts

This is a compilation of stuff Matt Cutts has said historically, minus some of the more recent stuff here, here, and here.  I decided I'd dig backwards and document some of the older stuff.  I dated it accordingly.  Here it is:

Matt recommends using dashes over underscores to delimit words in urls. 2005.
Google does not algorithmically penalize for dashes in the url despite the fact that some have raised it as a possible heuristic for spam detection. I think WordPress pretty much precludes this anyway. 2005.

Google takes action on individual instances of spam when they find it, but they focus on creating better algorithmic solutions. He states that he would not recommend using sneaky JavaScript redirects because your sites may get nailed in the near future. 2005.

Google updates their index data — including backlinks and PageRank all the time.  However, they export and publish new backlinks and PageRank data approximately every three months.  New backlinks and PageRank are meaningless — it is not an update.  The information is likely already factored in for awhile before you see it. 2005.

Hiding text using similar colors and background colors can actually be worse than using the same colors. Using “#EEEEEE” instead of solid white on a solid white background can look worse — as if you’re trying to hide it.  I suspect this is a heuristic for detecting hidden text. 2005.

If you sell links, Matt says you should use link condoms.  Otherwise your reputation may fall.  I assume this means they will devalue your outbound links. 2006.

Googlebot can only crawl the free portions that non-subscribed users can access since it does not log in.  Therefore, be sure to excerpt material in the free version that that offers value. Vanessa Fox. 2006.

If you must use Flash, you must also make an HTML version available as well. Block the Flash version from the crawlers with a robots.txt file. Vanessa Fox. 2006.

Use user-friendly URLs like "african-elephants.html," and not "343432ffsdfsdfdfasffgddddd.html." Don't overdo it either — african-elephants-and-their-habitats-etc-etc-etc-etc.html. Vanessa Fox. 2006.

Assign unique, descriptive <title> tag and headings to every page. Vanessa Fox. 2006.

Minimize the number of redirects upon hitting a URL. Vanessa Fox. 2006.

Minimize the number of URL parameters — 1-2 parameters if possible. Vanessa Fox. 2006.

Don’t use a parameter named “id=” in a URL for anything other than a session ID. Otherwise, it may not be included in the index. Vanessa Fox. 2006.

Earned-links are earned and given by choice.  Google does consider buying text links for PageRank purposes to be outside our quality guidelines. Vanessa Fox. 2006.

Blackhat SEOs may be leery of using Google for analytics, but regular site owners should be reassured. Vanessa Fox. 2006.

Google’s is against selling/buying links, and Matt indicates they are good at spotting them — both algorithmically and manually. Sites that sell links can lose their trust in search engines. 2006.

Google is focusing on detecting spam in other languages in 2006 — Italian, Spanish, Chinese, etc. 2006.

External (domainA -> domainB) 302 redirects are largely treated as 301s now. 2006.

Canonicalization is the process of picking the best url when there are several choices, and it usually refers to home pages — vs. vs.  Since all these urls are different, a web server could return completely different content for all the urls above. When Google “canonicalizes” a url, it tries to pick the best one and elimintes the others. To help Google, link to resources on a site consistently, and use 301 redirects to enforce it. 2006.

Do not use the URL removal tool to remove if you are worried about URL canonicalization and have both and in the Google index. This will remove the entire site. 2006.

Search engines can perform canonicalization for things like keeping or removing trailing slashes, upper vs. lower case, or removing session IDs from bulletin board or other software. 2006.

404s (Gone, but may reappear) are treated the same as 410s (Gone, but will not reappear).  Most web masters use 404s as 410s anyway. 2006.

Following these tips will probably help you rank better, so long as you actually trust Matt Cutts.  I'd assert it's wise to approach some of his advice with skepticism, since I'm sure Google gives him guidelines as to what he can actually say, but following the advice above won't get you into any trouble.

These icons link to social bookmarking sites where readers can share and discover new web pages.
  • digg
  • Furl
  • Reddit

Hilltop: la primera base de Google

Anoche leí el documento “Hilltop: A Search Engine based on Expert Documents”, el documento del que presuntamente se derivó la idea del PageRank del algoritmo primario de Google. El documento es obra de Krishna Bharat y George A. Mihaila y describía el funcionamiento de Hilltop, un motor de búsqueda basado en la evaluación de cuán experto es un documento.

El fragmento que yo creo que es más importante es este:

“We believe a page is an authority on the query topic if and only if some of the best experts on the query topic point to it. Of course in practice some expert pages may be experts on a broader or related topic. If so, only a subset of the hyperlinks on the expert page may be relevant.”

“In such cases the links being considered have to be carefully chosen to ensure that their qualifying text matches the query. By combining relevant out-links from many experts on the query topic we can find the pages that are most highly regarded by the community of pages related to the query topic. This is the basis of the high relevance that our algorithm delivers.”

Esto es la piedra angular de la evaluación de sitios por “autoridad”. Es decir, igual que ocurre en investigación científica, en que la autoridad de una obra científica viene determinada por el número de veces que ha sido citada en otras publicaciones científicas, el PageRank evalua las páginas por el número de links que han ganado en otras páginas.