24 Jun 2002 mglazer   » (Journeyer)

Here's one of the most useful bookmarklets I have:
   Go Wayback  (Explorer 4+ and Netscape 4+)
Trigger it when you run into a 404. As time goes by, there will be increasingly more documents in the past than we have in the present. So the value of a service that solves 404s, such as Wayback (which archives old webpages) will increase.
Another good Wayback bookmarklet:
   Wayback Undo  (Explorer 4+ and Netscape 4+)
which is necessary because once you go into Wayback you stay in Wayback... this allows you to escape to the present.
To push this a bit further, I'll bring up the topic of XMLHTTP (Microsoft's documents and Mozilla's documents).
This allows scripts on a page to request data from other pages, so you can get a bookmarklet like:
  

<A
href="javascript:if(location.href.indexOf('http://web.archive.org/web/*')==0){h='';x=new 
ActiveXObject('Microsoft.XMLHTTP');W=open('','','width=500,height=300,scrollbars,resizable,status');W.status='Looking...';L=document.links;for(i=0;i<L.length;i++){if(L[i].href.lastIndexOf('http:')>1)try{x.open('GET',L[i].href,false);x.send();r=x.responseText;h+=r.length+' 
- '+L[i].innerText.link(L[i].href)+'<br>';W.document.body.innerHTML=h}catch(e){}}W.status='Done'}else{alert('Go 
Wayback first')}void(null)">Wayback Analyze</A> 
 (Explorer 5+ version)
which, if triggered on a result of Go Wayback, will list the sizes of the archived documents. This gives some sense of the overall variation in the page over time. It would be easy to modify this to allow other analyses of the archived pages... allowing searches or summarizations, etc.
The idea of XMLHTTP could be used in other ways. For example, this bookmarklet:
<A 
href="javascript:html='<head><base target=_blank></head><body>';str='';xmlhttp=new 
ActiveXObject('Microsoft.XMLHTTP');W=open('','','width=500,height=300,scrollbars,resizable,status');W.status='Looking...';lnks=document.links;for(i=0;i<lnks.length;i++){if(lnks[i].host.replace(/:.*/,'').indexOf(location.host)!=-1&amp;&amp;str.indexOf(lnks[i].href)==-1)try{xmlhttp.open('GET', 
lnks[i].href, false);xmlhttp.send();r=xmlhttp.responseText;html+=r.substring(r.search(/<title>/i)+7,r.search(/<\/title>/i)).link(lnks[i].href)+'<br>';str+=lnks[i].href;W.document.body.innerHTML=html}catch(e){}}W.status='Done';void(null)">Self-Link 
Titles</A>
 (Explorer 5+ version)
lists the titles of all pages that link off the current page into the same domain. Again, search and summarization are fairly easy. So you get a way to analyze the "cloud of meaning" around the current page.
From http://bookmarklets.com

Latest blog entries     Older blog entries

New Advogato Features

New HTML Parser: The long-awaited libxml2 based HTML parser code is live. It needs further work but already handles most markup better than the original parser.

Keep up with the latest Advogato features by reading the Advogato status blog.

If you're a C programmer with some spare time, take a look at the mod_virgule project page and help us with one of the tasks on the ToDo list!