Professional Documents
Culture Documents
Rich Internet Applications create new opportunities. The just link to a particular web site, but links directly to a specific
most fundamental of these is the ability to create Single Page page within the site. It is this MPI paradigm which informs the
Interfaces (SPIs). A SPI is an interface that consists of a single robots which are used by search engines such as Google or
HTML page. Additional information that is required, when Yahoo to index the information in web sites. Search bots are
the user clicks on a ‘link’ or when some other event occurs, software agents that ‘crawl’ through web sites; they start at
is not supplied by means of a traditional full page reload, the index page and, after categorizing all of the information
but is instead retrieved via an XML message. The original on the page; they follow the links on this page to other pages
page remains intact, its contents or state is simply updated on the site. In this way they crawl through the entire web site,
by the contents of the XML message. JavaScript is used to visiting any page that has been linked to using a link tag of
facilitate this whole process. Although it is not mandatory to the type:
create a SPI, when using Backbase’s software, a SPI provides a
more intuitive user interface and smoother user experience. <a href=”nextPage.html”>Next Page</a>
There are a few questions that need to be answered however
when you make use of this new paradigm. One of the main However in an SPI, the linked page structure that the search
questions is that of search engine accessibility and deep bot is expecting, has been extended with BXML commands,
linking. which indicate the use of include files, load commands and
form submissions, which only partially update the page,
The web sites that have been created up until now, consist instead of causing a full reload as is the case with normal
almost entirely of Multi Page Interfaces (MPIs). These web forms. Since search bots aren’t proper web browsers, they
sites and applications consist of multiple unique pages, which don’t understand or execute any JavaScript. This means that
may or may not have been dynamically generated. Since each a Backbase SPI needs to be specifically designed to work
page, and for dynamic pages every page state, has a unique with these search bots.
URI; it is very easy to link to any page or state within this site.
Navigation between pages is done by the user clicking on This article puts forward a set of guidelines, which you can
links or submitting forms, both of which contain the location use to design your SPI for maximal search engine accessibility
and state information for the new page. It is these unique and shows you techniques to allow for deep linking into your
URIs that make deep linking possible. Deep linking does not SPI.
Several approaches are available for making your web site Use a keywords meta element with a content attribute
accessible to search engines; these approaches differ in containing some appropriate keywords. For example:
the level of indexing, which is obtainable and how this is
achieved. For certain sites, it is not necessarily a requirement <meta name=”keywords” content=”WebMail, e-
that every part of the site can be indexed by search engines. mail, bxml, mail” />
For example, a site, which provides a web-based e-mail
service, does not require every single piece of information on Use a description meta element with a content attribute,
the site to be indexed by a search bot. Other sites, however, which contains a relevant description of the web page.
do require that every piece of information can easily be found The value of this element is often printed as part of a
and indexed by search engines. For example, a web site with search result by Google. For example:
information about the courses provided by a university is
such a case. Backbase has identified the following strategies <meta name=”description” content=”A Free
for getting a SPI indexed by search engines: BXML WebMail application. This unique
WebMail application offers the look and feel
of a normal Windows application, with the
Lightweight Indexing: no structurally changes are made ease and portability of a web-based client.”
to your site; existing tags such as meta, title and h1 are />
leveraged.
Place key content within the main HTML structure and
Extra Link Strategy: extra links are placed on the site, not in an include file, or some other dynamically loaded
which search bots can follow and thereby index the whole content. If possible, place this important content within
site. a h1, h2 or h3 element, since search bots deem these to
contain more important information. Remember that
Secondary Site Strategy: a secondary site is created, these tags can be styled in anyway you want using CSS.
which is fully accessible to the search engine.
It should be noted that these points can also be put to good
For each of these strategies the following questions will be use, in the design of your SPI, in conjunction with the extra
answered: link strategy or the secondary site strategy.
To what extent is the content of the page indexed? In summary by using this lightweight-indexing strategy
only the content supplied by the title and meta elements and
Can links be followed on the page (e.g. link elements (<a those elements that are directly located on the index page
href=”xx”>) or s:include elements)? is indexed. No links of type s:include are followed; therefore
there is no requirement to deal with redirection. This is not a
When a link is followed by the search bot, what is the very full indexing scheme, but it is extremely simple to apply
status of the URL that is being indexed. Can this URL be to your site.
displayed by browsers or will some type of redirection be
required?
The Extra Link Strategy
Lightweight Indexing There are two main approaches to making a site fully
indexable by search engines: the extra link strategy and the
This strategy should be used if only certain key information secondary site strategy. The extra link strategy is the easiest
needs to be indexed by search engines. In this case it is of these two to implement and it can make the site entirely
recommended that you take the following steps when indexable by search engines, but does not create a secondary
designing your SPI: site in normal HTML and is therefore not accessible to older
browser, which are incompatible with BXML. The essence of
Use a title element in the document head, preferably this strategy is to create an extra link on the main SPI index
containing one or more keywords that specifically relate page for each include file, whose contents you wish to be
to the contents of the site. For example: indexed. Some experimentation has revealed that the extra
links must be of the type:
<title>BXML WebMail – Sign In</title>
<a href=”include1.html”>include 1</a>
• Internet Explorer 5.0 and newer Obviously, it is also necessary to test for compatible versions
• Mozilla 1.5 and newer of Internet Explorer too. This can be done in a similar way,
• Firefox 1.0 and newer but there is one added complication. All compatible versions
• Netscape 7.2 and newer of Internet Explorer have a User-Agent string that contains
the sub-string: ‘MSIE’, which is directly followed by the ver-
User-agent detection can be done on the server using a PHP, sion number. Below is an example of such a header from an
ASP or JSP script. There are standard libraries, which help take Internet Explorer browser.
care of this. Alternatively if you cannot or do not wish to use
User-Agent: Mozilla/4.0 (compatible; MSIE 6.0;
server-side scripts to determine the user agent, it is possible
Windows NT 5.1; SV1; .NET CLR 1.1.4322)
to do this in JavaScript. If you take this approach, you should
be aware of the fact that search bots cannot be expected to
However unfortunately Opera browsers have a very similar
execute any JavaScript. Therefore if you are using the sec-
User-Agent string:
ondary site strategy in conjunction with JavaScript based
detection, the default page provided by the initial page re- User-Agent: Mozilla/4.0 (compatible; MSIE 6.0;
quest must be the non-BXML site, which is intended for the Windows NT 5.1; en) Opera 8.00
search engine bot. When you ascertain that the user agent is
a BXML-compatible browser, then JavaScript should redirect Therefore you must firstly test that the User-Agent string
the browser to the BXML version of your site. The following doesn’t contain the ‘Opera’ sub-string and once this has been
code fragment shows a simple JavaScript function, which ascertained, then simply parse out the version number which
tests whether a BXML-compatible Mozilla-based browser is follows the ‘MSIE’ sub-string.
in use and then redirects the browser based on this.
function testUA(){
var bCompatible = false;
var sUA = window.navigator.userAgent;
var iIOGecko = sUA.indexOf(“Gecko”);
//Test if the User-Agent string contains
//the string Gecko
function updateState(){
//first parse out the value of the referrer
//var sReferrer = document.referrer;
//do quick test to make sure that referrer
//is from the same host
if(sReferrer.indexOf(
window.location.hostname) >= 0){
var iLastSlash =
sReferrer.lastIndexOf(‘/’);
var sValue =
sReferrer.substr(iLastSlash + 1);
//trigger an event with the same name as
//the referrer
var sExecute = ‘<s:task
b:action=”trigger” b:event=”’ + sValue
+ ‘” b:target=”id(\’main\’)” />’;
bpc.execute(sExecute);
}
}