Ever since Blaise Aguera y Arcas demoed Photosynth last year at the TED Conference, most people who saw it have wanted to get their hands on the software. The demo was, as the TED web site called it, simply jaw dropping. Photosynth was able to take hundreds of photos of the Notre Dame Cathedral gathered from Flickr and use them to create a virtual 3D environment by recognizing specific aspects of the cathedral in each picture how they related to the other photos. Each photo was placed in space accurate to where it was taken, allowing viewers to zoom around the cathedral and get an accurate picture of what it looks like from all angles.
Last week, Microsoft released Photosynth as a free, Windows-only browser plugin and application, along with 20GB of free online storage to store your “synths.” The plugin supposedly runs on Firefox 3 and IE 7, but I could only get it working under Internet Explorer. Photosynth fits into Microsoft’s philosophy of software + services, in which desktop software is augmented by web-based components.
Once I got it working, Photosynth was exceptionally easy to use. Just choose your photos, add them to the Photosynth creator app, and click a button to make your synth. The process takes a few minutes and the photos are uploaded to your account on Photosynth.net. The software grades the “synthiness” of your creations — which I gather is a numerical representation of how good the coverage of your subject is and how well your photos stitch together. My first attempt — the entire first floor of the SitePoint offices in about 40 pictures — was a complete failure, achieving just 32% “synthiness.” My second attempt, a depiction of just the break area in the office and which is embedded below, had a better 61% “synthiness” score utilizing 45 photos.
Ideally, synths should have 150-200 photos and no less than 20.
While the speed and ease of use of Photosynth is indeed impressive, I unfortunately found the actual synths to be a little underwhelming. Part of that might be because my subject matter for creating test synths wasn’t the best fodder for this technology — it excels at spatially relating photos taken from multiple angles of a single object, whereas my attempts at “synthing” used photos of a mostly empty space taken from essentially the same spot.
However, even the demos on the site from Microsoft employees and National Geographic weren’t as impressive as the original TED demo of Notre Dame. Perhaps that’s because of inferior photo sets, or perhaps it is due to slight changes in the viewer (I found it harder to zoom out and get the overall “3D” picture, for example, even with synths that should have been perfect for that, such as the Taj Mahal and the Sphinx).
Microsoft’s TED demo scraped hundreds of images from Flickr and was able to relate them spatially to create a 3D representation of an object. But right now, Photosynth is a personal technology — there’s no easy way to collaborate and gather multiple photos from outside sources. You can only create synths using pictures from your computer, and most people are unlikely to have the comprehensive photo sets available necessary to create high quality synths. Microsoft hinted that more collaboration tools are coming, though. When that happens, the technology will be less of a novelty, as people will (hopefully) be able to contribute photos to public synths and keep making them better.
Have you tried Photosynth? What were your thoughts? What sort of things have you synthed?