
They take care to mention this is just a showcase outline demo of an ultimate system, not even a Alpha, whatever that means now. I installed it and played with their sample data, a DB of global satellite images, map data from the Library of Congress and other sample images. Transitions were smooth, but some of the zooming was slow, likely due to my WiFi connection reloading.
I much like the idea of using an infinite field of images that I can manipulate to study, especially from a mobile device. Could be useful for certain kinds of collaborative design capabilities. For example, comparing real objects and locations and large numbers of stored images. Would be further useful to include tag generation and image analysis capabilities to aid its exploration. Also a way to add a markup layer where I can collaborate with others on the field of image(s).
You can compare this to Google Earth, which also has a mobile viewer that gives fairly smooth drag and follow interaction on the IPhone. Compare it to Tag Galaxy for tagged images.
There is no way as yet to be able to search any of the image meta-data, like tags or title or dates. Quite fundamental. I would imagine that this would be easy to add at some point. So all I can do is browse in two dimensional spaces like maps. Some of the exploration methods, other than zoom in and out, are still primitive. The Photosynth connection, when working will also be interesting to examine for better exploration of recorded spaces.
Suppose we could generate images from our brain and then scan them this way? A backup of our own visual memory?
This is definitely worth a look and following to see how it evolves. There is also a Seadragon Ajax capability that you can place in your own web spaces to provide a viewer and related capabilities. I have not explored these yet.
Scanning just some of the comments, there is the usual cynical criticism and mostly childish poking at Microsoft. Ignore it.
No comments:
Post a Comment