a URL shortener with semantic and geo-spatial analysis

July 9th, 2008 is a URL-shortener like TinyURL with a host of interesting features, as enumerated in the switchAbit blog.

1. History — we remember the last 15 shortened URLs you’ve created. They’re displayed on the home page next time you go back. Cookie-based.
2. Click/Referrer tracking — Every time someone clicks on a short URL we add 1 to the count of clicks for that page and for the referring page.
3. There’s a simple API for creating short URLs from your web apps.
4. We automatically create three thumbnail images for each page you link through, small, medium and large size. You can use these in presenting choices to your users.
5. We automatically mirror each page, never know when you might need a backup.

A post in ReadWriteWeb, Please Use This TinyURL of the Future, points out some interesting ‘semantic’ features.

“In the background, is analyzing all of the pages that its users create shortcuts to using the Open Calais semantic analysis API from Reuters! Calais is something we’ve written about extensively here. will use Calais to determine the general category and specific subjects of all the pages its users create shortcuts to. That information will be freely available to the developer community using XML and JSON APIs as well.
    As if that’s not a whole lot of awesome already – is also using the MetaCarta GeoParsing API to draw geolocation data out of all the web pages it collects.
    You want to see all the web pages related to the US Presidential election, Barack Obama and Asheville, North Carolina? Or about Technology, Google and The Dalles, Oregon? That will be what delivers if it can build up a substantial database of pages. Once it does, it will open that data up to other developers as well.”

The idea of using a URL shortening service to identify significant or interesting Web pages for further processing is a new twist. It would be great of other services with catalogs of interesting pages, like, did this as well. Eventually, this will be done to the entire Web, but for now, it’s too expensive. This is an interesting intermediate step.

HealthMap mines text for a global disease alert map

July 8th, 2008

HealthMap is an interesting Web site that displays a “global disease alert map” based on information extracted from a variety of text sources on the Web, including news, WHO and NGOs. HealthMap was developed as a research project by Clark Freifeld and John Brownstein of the Children’s Hospital Informatics Program, part of the Harvard-MIT Division of Health Sciences & Technology.

HealthMap mines text for a global disease alert map

Their site says

“HealthMap brings together disparate data sources to achieve a unified and comprehensive view of the current global state of infectious diseases and their effect on human and animal health. This freely available Web site integrates outbreak data of varying reliability, ranging from news sources (such as Google News) to curated personal accounts (such as ProMED) to validated official alerts (such as World Health Organization). Through an automated text processing system, the data is aggregated by disease and displayed by location for user-friendly access to the original alert. HealthMap provides a jumping-off point for real-time information on emerging infectious diseases and has particular interest for public health officials and international travelers.”

The work was done in part with support from Google, as described in a story on ABC news, Researchers Track Disease With Google News, Money