Search Engine Optimization articles.

Obama and the White House Robots.txt

It has been talking a lot about how Barack Obama has used the Internet to publicize his candidacy and for mobilizing voters. It has also been commented on many blogs the ambitious technology plan Obama for America (you can read here).

But one of the things that has caught more attention and few people have noticed: the change that has suffered the Robots.txt the website of the White House, very much in line with what Obama preaches.

What is Robots.txt?

It is a text file containing instructions on pages and not visitable visitable by Robots, a web page. That is, it indicates which parts of the website should not be scanned by robots.

Normally, it is content that appears on the website, but only want it to be accessible to people who surf the web, you do not want it indexed content appears in search engines. It is also used when a content manager creates duplicate content and thus penalized by search engines.

This file is created following the instructions can be found here: RobotsAnd all the robots that follow the "Robots Exclusion Protocol"Undertake to heed these instructions.

If a website has created this text file, robots understand that they can index it (although to be sought from that page robots.txt robots generate a 404 and therefore, it is recommended that a blank page is created and FTP upload is named Robots.txt to thus generated 404 on the page will be real and can be released by the webmaster).

Let's return to the White House Robots.txt

Until a few days ago, when explained in class what a file Robots.txt and what is the "Protocol Robots Exclusion"I put several examples to illustrate the different types we can create robots.txt to instruct robots indexers:

  • A blank page robots.txt
  • Robots.txt A page with more or less "normal" instructions
  • A robots.txt page completely exaggerated and misplaced.

Well ... Obama has me "saboteado"Examples and it loaded my example of malpractice in a matter of Robots.txt: The webmaster of the new website of the White House has created a new Robots.txt well done, clear and concise.

The webmaster of George Bush Jr., had created a robots.txt with thousands and thousands of pages with forbidden access to the robots. Or say fits that ... there was nothing interesting in that content (once he had dedicated me to go read what they did not want it to be indexed ... pictures of the first lady, speeches, etc ...). But it showed well that the White House had a somewhat archaic what is internet and publishing content concept.

The new websmaster, in this sense, is shown to have much clearer what should be the website of an institution like the White House.

It is worth ... but how was this Robots.txt?

Fortunately, in the slides of my classes always I include screenshots of what I explain, not fail me the internet or where class have no connection ... (how sad to have to always think about this possibility).

So under these lines (at end of post) include the image I have filed and that now becomes history ... (Look at the bar scroll the pantallazo... It is the one that shows the magnitude of the listing)

The current robots.txt page can see it by clicking here: Robots.txt of Casablanca with Obama .

For more information about how to create a Robots.txt or what it is, you will find here: Robots.txt and in the Free Course Search Engine Optimization our website: Search Engine Optimization Course

Robots.txt of Casablanca

Does Google's Golden Triangle is still valid? The answer is yes

Who else who has less ever see a heat map that runs through the network, Eyetracking done with technology and showing what is called "The Golden Triangle Google”: The golden zone in which you have to appear if you want to be something on the web… the space where all the eyes of Google users converge. I myself, I have used it a great number of times in classes and conferences, to graphically show the importance of appearing in the top positions of Google, be it in the form of Adword, Whether in the form of natural positioning (SEO).Eyetracking - Triangulo de Oro de Google

But this heat map has its years, and every time I use it I wonder whether it will be really valid now that Google displays results including maps, books, definitions, etc. (The image I show along these lines corresponds to the classical heat map to which I am referring).

Hence one of the things I wanted to test in testing we have been doing during the months of December and January to draw up a study of Search Marketing-oriented tourism, using technology Eyetracking, was whether the Golden Triangle was still valid , or otherwise, now he appeared another geometric figure. This is the study: Eyetracking Search Marketing

Well, after performing the tests (54 individuals by testing 4 different tasks, altogether 216 heat maps results pages Google), we conclude that it remains valid, especially when very specific searches are performed and offer results found in the top positions, although the lateral Adwords are now most watched shows the classical triangle.Eyetracking Hotel Pulitzer Barcelona PCT

Let's look at some of these heat maps:

This I show below is a heat map to search "Hotel Pulitzer Barcelona"(Without quotes) adding the maps of 30 users tested, taken at random and untargeted. The task was to find the URL official Hotel Pulitzer (he was both Google Maps and the first result SEO). I show the map is a map of percentages fixing eyes (the same type the "classic" only that our EyeTracker different).

We can see how the triangle is formed, though somewhat blurred because of the map.

If this same heat map we show by absolute values ​​of fixating, then Eyetracking Hotel Pulitzer Barcelona AbsolutoIf we see the triangle, and we also note that some users have looked at the Adwords side, which is barely visible on the map by percentages. (See attached image)

A curious observation about the side Adwords: If this same search took, heat maps users under 25 will see something very different: not one has looked at the side Adwords and the triangle appears clean but somewhat disfigured, and focused on higher Adwords and natural results.

It seems that those under 25 are very clear that the lateral Adwords are advertising and therefore do not deserve your attention ... (this topic is important and worth analyzing further because it refers to the theories of communication Postmodern well that both the map and the subject I discuss in another post).

The complete study can be downloaded here:

The complete study can be downloaded here: Eyetracking Search Marketing

This study may also be interested in other analyzes gaze on whether users see or not see advertising. Eyetracking_Media_Espana

Google Maps vs SEO: SEO keeps winning

During the months of December and January we have been conducting a battery of tests to develop a study Search Marketing-oriented tourism. The tests for the study were performed using Eyetracking technology and a personal interview after the test is completed, to know where users tested looked and looked there and why not elsewhere.

In total, 54 people aged between 16 and 68 years were tested.

The results have been very interesting and will publish on this blog as we move forward in analyzing the data.

The first of the curious things we've noticed is that While Google Maps appear is important, more important still running first for natural positioning. Especially in specific searches. In general searches such as "hotel in Barcelona"There is no doubt about the importance of appearing on Google Maps. In this post we will analyze the specific searches.

Let's see why we have come to this conclusion:

One of the tasks to be carried out by users was to find the website of the Hotel Pulitzer in Barcelona between all the results that appear on Google by searching for "Hotel Pulitzer Barcelona"(Without the quotes). Apareceían 234,000 results.
Eyetracking Hotel Pulitzer Barcelona
Looking at the heat map image I show along these lines, we can see that the official website of the hotel appears in both Google Maps and in the first place by natural positioning. As well:

  • A 51.85% of users in clicaron first result of natural position.
  • 40.74% of users clicaron on the web that appears on Google Maps.
  • 7.41% of users (4) were wrong and clicaron on Adwords or other natural results, convinced that clicaban the hotel website.

When we asked why they had clicked clicked where the answers were as follows:

  • Those who had clicked on SEO results indicated that rely more of these results. In fact, 22% had not even looked at the area of ​​Google Maps and gone directly to the SEO area. When asked why they did not trust, most indicated that they thought it was instead advertising and knew the results that follow the map are not advertising.
  • Those who clicaron on the map indicated that clicaron on the map that appeared in the middle of the screen and saw that it was directly and the hotel page.

Most people indicated that the URL of the hotel was what I told them it was the hotel's official website (obviously they called the "hotel management" not the URL, I am writing URL to facilitate the writing of this post). 25% also indicated that the title of the result indicates "Official Website" had reaffirmed that they were right and that was the official website.

  • 4 people wrong and clicaron on other outcomes, they were convinced that they were right and had given the correct web (until they saw the web to which arrived after clicking, of course, at which time they realized their mistake). In most cases it was the URL that misled them and made them believe they were on the correct result. URLs that were the type clicaron "Hotel-Pulitzer-Barcelona.h.rez.com"Hence some people confused. Although the titles of the ads also helped confuse "Hotel Pulitzer Barcelona"As the title of a Adword is able to confuse anyone if not take into account other parameters.

As for Adwords in general, almost everyone looked at some point 3 advertisements are highlighted in the upper zone (only 25% not looked), regarding the lateral Adwords, were not so lucky, and only 22% watched them.

In another post again we discuss the issue SEO vs. Google Maps, but using an example in which the results it delivers the page, in both cases are different (the Avanzo ... SEO wins again).

Later note: The study can download it here: Eyetracking Search Media .

Presentation of the Guide to Electronic Commerce in Igualada

Today we have carried out the presentation of the Electronic Commerce Guide at the headquarters of the Chamber of Commerce of Barcelona in Igualada.

The presentation was made by the Director of that office. To accompany the event, Alex de Anta gave a talk on the key success factors in implementing an Electronic Commerce, and I gave a talk explaining a part of the Guide.

As time was limited (1 hour and a half), I only explained the 9 basic techniques to attract visitors to a website, and 6 basic techniques to convert them to business contact. These techniques only have explained above, unfortunately, because it would have needed a minimum of 10 hours to explain this part as a monographic course.

The guide itself can read the same information, but more extensive and complete. I have also indicated participants in the web GeaMarketing and this blog will find additional information and other presentations that can help them expand any of the items discussed in the paper.

I forgot to comment on companies that have attended the presentation of the Guide, the Chamber of Commerce subsidizes 50% of the cost of conducting a small study on the Strategic Plan for the Implementation of an Electronic Commerce, for any company of Barcelona area you want to start selling over the Internet. The total cost has to bear the company is 750 euros and can be ordered through the delegations of the House of Barcelona.

The next presentation will be on Thursday November 27 in Mollet del Valles. At 9h30 'am, at the headquarters of the Barcelona Chamber of Commerce held in this population.

See you there.

I Roundtable web browsers: marketing and search engine optimization

On Tuesday 28 October, at 18:30, will be held in the Auditorium of the building Ramblas Universitat Pompeu Fabra (Ramblas 32, Barcelona), presenting a new edition of the Master in Search Engine Optimization and Digital Marketing and as a panel discussion on web Search.

I have the pleasure of sharing roundtable Fernando Macia from Human Level Communications, Who will talk about SEO and Christopher Rovira Research Group UPF DigiDoc to discuss finders training and research. I will discuss SEM (Search Engine Marketing) and explain "6 techniques that will help us optimize our search engine marketing budget".

Each of us will present his vision on search engines and share related experiences market intelligence, content strategy and brand positioning.

At the end of the presentations, a discussion in which attendees can ask their questions will open.

More information about the event:

See you!

More information about the event: Roundtable Search Engine Optimization

This is the presentation with which I illustrated my talk:

We talk to each other.

Google updates the PageRank value of your bar

On April 30 Google updated the PageRank showing us your toolbar. This is the second update of 2008. The next update is expected in late September.

What is PageRank and its importance in the sorting algorithm results?

PageRank is the algorithm that Google uses to give a numerical value to the importance of a web page. This value is used as part of the algorithm that sets the order in which search results are displayed on Google.

The PageRank is named in honor of its creator Larry Page. PageRank does not mean "ranking of pages".

The purpose of PageRank is to assign a numerical value to web pages according to the number of times the recommended and other pages according to PageRank having these pages. That is, it establishes the importance of that website.

Over the years Google has been modified this algorithm in order to neutralize some traps made by webmasters who wish to modify their position fraudulently. In this latest update it seems to have also modified the algorithm as numerous websites have lowered their PageRank ... but fortunately, many others have gone up, including some of our customers.

How do the updates ?: difference between the actual PageRank and PageRank Google bar work

The PageRank shows the Google bar only updated once every 4 months or so.

The bar shows a PageRank base 10 on a logarithmic scale. That is, it is easy to climb from 0 to 1 or 2 to 3, but instead is very difficult to climb from 5 to 6, and even more up 6 to 7. But this is not the actual PageRank of our website, but value that Google assigned the last time you updated the PageRank bar.

During 2007 there were 3 updates PageRank in the Google bar: in January, April and late September. Forecasts for this year are the same: we waited 3 updates.

The PageRank that Google uses for its calculations is much more accurate and uses a much larger scale, we do not know how, and Google is a complete secrecy in this regard, although it seems that is based 100. Their internal servers update it daily.

When is the next update?

If all goes well, we should wait for the month of September. So all actions to increase the PageRank we perform from now, will not be reflected in the Google bar until this month.

This does not mean that before September, our actions are useless. Nothing is further from reality. Recall that Google works with a PageRank in real time.

In my bar I do not see Google PageRank, what I have to do?

By default, the Google bar does not include this information, but from bar options can include PageRank, and thus, while browsing the network will know the PageRank of the pages you visit. This will help you know what websites should try to include your links to increase your PageRank.

Links of interest:

How I can improve my PageRank website ?:
http://www.geamarketing.com/posicionamiento/mas_pagerank.php

Free Online Course Search Engine Optimization:
http://www.geamarketing.com/posicionamiento_buscadores.php

AJAX, a technique used sparingly

This article explains what AJAX is, when to use and what contraindications. We also show how to overcome some of the contraindications.

What is AJAX?

AJAX stands for: Asynchronous JavaScript And XML. That is, the combination of JavaScript and XML asynchronously.

It is a technique developed for interactive Web applications, which consists of a set of three existing technologies work together effectively.

These technologies are:

  1. (X)HTML y CSS (Cascading Style Sheets) to give a structure and present the information on the website.
  2. JavaScriptUsed for dynamic interaction with data.
  3. XMLUsed for interaction with the web server. Although it is not necessary to use XML simpre with AJAX applications, as for example, plain text files can also be stored information.

Like DHTML, AJAX is not itself an independent web technology, but a term that encompasses the three aforementioned technologies.

What is Ajax?

Ajax is used to make changes to a web page at the user, without having to reload the whole page again.

For example, on a web page the user requests any information that is offered from the same web page (such as a description of a product) and click on the link on the same page (without loading it again) the requested information appears .

The process of displaying HTML data made entirely consume significant bandwidth, since all HTML should again be loaded to only show the changes. Instead, an AJAX application is much faster and consumes bandwidth.

The JavaScript used in AJAX application is a dynamic, able to make changes to a Web page without re-charge language. AJAX makes sure that only the necessary information is requested and processed, using SOAP or other Web services language loosely based on XML.

Hence a technical level, 3 advantages are obtained: one time charge much lower, saving the user bandwidth and server load much less where the website is hosted.

AJAX problems

Problems with search engine indexing:

AJAX is used by Google, Yahoo, Amazon and a lot more than search engines, portals and content creators, but not general use and massive as some think. Google, for example, which supports webmasters to use AJAX in their programming, uses it himself in GMail, Google Suggest, Google Maps, but not absolutely all your web pages.

The problem with AJAX is that the content displayed within the application using AJAX, not indexed in search engines. This is because spiders (spiders) search engines are not able to interact with the AJAX application and get the command that displays the content is activated.

Hence, it is a bad idea for example, create a list with the names of our products and make an AJAX application by clicking on a product name, product description and photograph is displayed to the right of the list. If we do this, descriptions of products and their images will not be indexed or Google or any other search engine.

Although not all bad news, certain ways of working with AJAX itself that index, for example, playing a show or not show content using positive and negative margins. So just to bear in mind when scheduling if spiders may pass or not can spend.

Accessibility problems:

If we start from the basis that our website should always be accessible to all types of browsers and users and should at least meet the standard A W3C (http://www.w3.org), We find that most scripts that improve appearance and interactivity of a website have accessibility issues. The AJAX also has them.

As we have seen at the beginning of this article, the use of AJAX involves using JavaScript, and some browsers do not support this type of programming. Although as we shall see it is solvable.

But keep in mind that a large part of AJAX applications that we find in the libraries that exist on the Internet have not corrected this problem and therefore are apps that do not meet the W3C standares (at the end of these lines provide links to libraries code and articles dealing with the issue of accessibility and AJAX).

AJAX, to use sparingly

As we have seen in previous section, although AJAX applications provide dynamic, interactive and reduced bandwidth to a website, they also have drawbacks to search engine indexing level and level of accessibility. Therefore, we must consider and neutralize the following:

  1. If we use AJAX on our websites, we must be aware that the content displayed within the AJAX application will not be indexed by search engines. To remedy this detail, we can create this redundant content and make it accessible to the spiders through a sitemap or through links in the footer of the website.
  2. If we use AJAX to make our website interactive, we must keep in mind that will not meet the Level A accessibility, unless we use the code libraries adopted by the W3C or means to surf the web without using JavaScript.

related links

New information on Google indexing AJAX (March 2010): http://code.google.com/intl/es/web/ajaxcrawling/

Examples of Web pages that use AJAX and AJAX code libraries for use by webmasters:
http://ajaxpatterns.org/Ajax_Examples

Articles which explains how to get AJAX code that does comply with the level A W3C accessibility:
http://www.maxkiesler.com/

List common accessibility errors:
http://www.w3.org/TR/WCAG20-SCRIPT-TECHS/#N11799

Google updates the PageRank value of showing their bars

Google PageRank update in May 2007.

As planned, this weekend Google has updated the PageRank displayed in the Google bar that users have installed on their browsers. Google only updates this information every four months.

What is PageRank and its importance in the sorting algorithm results?

PageRank is the algorithm that Google uses to give a numerical value to the importance of a web page. This value is used as part of the algorithm that sets the order in which search results are displayed on Google.

The PageRank is named in honor of its creator Larry Page. PageRank does not mean "ranking of pages".

The purpose of PageRank is to assign a numerical value to web pages according to the number of times the recommended and other pages according to PageRank having these pages. That is, it establishes the importance of that website.

From January 24 (day when Google changed its sorting algorithm to neutralize some of the pitfalls that performed webmasters unscrupulous) in PageRank also affects the reliability of the website ... (but this does not explain now, but will be subject of another article).

How do the updates ?: difference between the actual PageRank and PageRank Google bar work

The PageRank shows the Google bar only updated once every 4 months or so.

The bar shows a PageRank base 10 on a logarithmic scale. That is, it is easy to climb from 0 to 1 or 2 to 3, but instead is very difficult to climb from 5 to 6, and even more up 6 to 7. But this is not the actual PageRank of our website, but value that Google assigned the last time you updated the PageRank bar.

The last update was carried out on January 24 and this time it has done on May 1, a few days before they met the mandatory four months.

During 2006 there were 4 updates PageRank: in February, April, July and late September. In other words, on 4 occasions during 2006 Google has calculated the value of PageRank in base 10 and has exported to the servers that power Google bars. During 2007 he is following the same pattern.

The PageRank that Google uses for its calculations is much more accurate and uses a much larger scale, we do not know how, and Google is a complete secrecy in this regard, although it seems that is based 100. Their internal servers update it daily.

When is the next update?

If all goes well, we should expect it by early September. So all actions to increase the PageRank we perform from now, will not be reflected in the Google bar until September.

This does not mean that before September, our actions are useless. Nothing is further from reality. Recall that Google works with a PageRank in real time.

How to know the PageRank in real time?

The exact numerical value and PageRank, we can not know in real time, but we can make an approximation to the real PageRank, although it is in base 4 instead of base 10 and relative values.

Permanent Link: Learn how to be a substitute Google PageRank: the TRUST RANK

Not much what we get with this, but at least we will know if we have PageRank assigned on every page, and we'll see, if you are increasing the number of pages of our web passing from medium to high, or low to medium.

In my bar I do not see Google PageRank, what I have to do?

By default, the Google bar does not include this information, but from bar options can include PageRank, and thus, while browsing the network will know the PageRank of the pages you visit. This will help you know what websites should try to include your links to increase your PageRank.

Links of interest:

How I can improve my PageRank website ?:
http://www.geamarketing.com/posicionamiento/mas_pagerank.php

Free Online Course Search Engine Optimization:
http://www.geamarketing.com/posicionamiento_buscadores.php

How indexes the https?

Https indexing is one of those mysteries that makes life more interesting SEO. While we know that it is possible to index it in most search engines, hardly anyone knows how to get it in the shortest possible time.

What is https?

The https is the secure version of the http protocol. The difference between one and the other is that the former transmits the encrypted data, and the second transmits unencrypted.

The system uses https based on Secure Socket Layers (SSL) encryption to send information.

The decoding of the information depends on the remote server and the browser used by the user.

It is mainly used by banks, online stores, and any service that requires sending personal data or passwords.

How does the https?

Contrary to what many people think, the https does not prevent access to information, only the encrypted when transmitted. Hence the content of a web page that uses the https protocol can be read by search engine spiders. What can not be read is the content that is sent from the website to your server, for example, the login and password for access to a private area of ​​the website.

The standard port for this protocol is 443.

How do we know the https is actually indexed?

Google indexes https since early 2002 and gradually other search engines have adapted their technology to also index the https.

The last search engine to do so was MSN, he got it in June 2006.

If we look for "https: // www." Or inurl: https in major search engines, we find https pages indexed in them.

How can we index our https?

In principle, naturally we can index our https pages, but as this protocol transmits information much slower, spiders sometimes fail to download the pages in the time they have established and will not index it. This is the main problem that we can find. We will resolve trying to reduce the download time of these pages.

How can we accelerate the indexing of https

There are two techniques:

  1. Google Sitemap: Include our sitemap https pages (we refer to the google sitemap, sitemap to not to humans), and register it in google sitemaps.
  2. Guerrilla: Internet links spread all over to go to our https pages, and thus achieve the spiders that are indexing the pages where the links have also come into the https part of our site.

How can we make our https being indexed

It is not as easy as it looks. It does not serve to include in our robots.txt https pages. Each port requires its own robots.txt, so we create a robot.txt to our http pages and another for our https pages. In other words, we also have a page called

https://www.nombredelapagina.com/robots.txt

If you need help or de-index to index your pages https, please contact us. We will encatados to assist you.

Additional information:

Blog MSN about indexing - Article where they explain that MSN index starts https
http://blogs.msdn.com/livesearch/archive/2006/06/28/649980.aspx

Information about how Google not index https:
http://www.google.es/support/webmasters/bin/answer.py?answer=35302

More information about Google Sitemaps:
SiteMaps de Google
http://www.geamarketing.com/articulos/Descubre_indexacion_futuro_Google_SiteMap.php

online course, free, search engine optimization: Course search engine positioning
http://www.geamarketing.com/posicionamiento_buscadores.php