Tim Molderez wrote:
Hi Juan,
Those samples look quite amazing! At first I found it a little hard to
believe that current anti-aliasing (AA) techniques looked so poorly,
so I took a quick look at the aliasing in a bunch of different
programs: First, I opened your sample .svg picture of a tiger in
Firefox, Illustrator CS4, Adobe Reader 9 and Corel Paint Shop Pro X;
none of them seem to come close to your Morphic 3 solution. However,
to my knowledge, none of these programs are hardware-accelerated,
meaning the CPU is doing all the hard work. I wanted to see whether my
graphics card could do any better, so I booted up a couple of games
(Crysis and Supreme Commander :) ). You can still see some aliasing at
2x and 4x supersampling AA, but once you crank it up to 8x or higher,
there is very little aliasing left, if anything at all. The same goes
for the two 3D modeling tools I've tried (3ds Max 2010 and Blender
2.49b), which are hardware-accelerated as well. I don't know much
about anti-aliasing (well, pretty much nothing until after I read your
e-mail :) ), but I don't think these results are very surprising: You
just use the GPU's immense horsepower to supersample your image at
such a high resolution until you surpass, or come close to, the
image's Nyquist frequency.
Hehehe. I'm not using super sampling. And I'm not running on the GPU!
For this reason, I'm mainly interested in how Morphic 3 compares to
current AA techniques performance-wise. I think all graphics-related
tasks, not just 3D rendering tasks, will be handled by the GPU in the
(hopefully) near-future anyway. (As an example of this future
direction, Photoshop already supports hardware acceleration as of
version CS4.) However, even with all of the GPU's power, rendering
images at 8x supersampling or higher is an immense performance-drain,
so it would be very nice if Morphic 3's AA has much better performance.
Yes, it will have much higher performance to 8x super sampling.
If so, it would be even better if Morphic 3 eventually catches the
industry's attention, because I started paying attention to the
aliasing in every image I see since I saw your samples and it's
becoming quite bothersome. :)
Yes! The same happens to me after so much time looking at my experiments!
Cheers,
Tim Molderez
PS: To demonstrate how powerful a GPU really is, the image processing
research lab at our university has turned a plain old desktop computer
into a full-blown supercomputer by simply throwing in 4 graphics
cards: http://fastra.ua.ac.be I'm just mentioning it because their
demonstration video is such a fun watch. :)
Cool!
Cheers,
Juan Vuletich
At 2/06 16:15, Juan Vuletich wrote:
Hi Folks,
I am developing a novel way to do anti aliased 2d graphics that
breaks away from pixel coverage and super sampling, while being
simpler and providing higher quality. Please take a look at
http:www.jvuletich.org , especially at the samples.
I believe this work fits nicely in the FONC project, as a Nile
implementation of it would yield better quality results than Gezira.
Comments welcome.
Cheers,
Juan Vuletich
_______________________________________________
fonc mailing list
[email protected]
http://vpri.org/mailman/listinfo/fonc
------------------------------------------------------------------------
No virus found in this incoming message.
Checked by AVG - www.avg.com
Version: 9.0.819 / Virus Database: 271.1.1/2918 - Release Date: 06/04/10 15:25:00
_______________________________________________
fonc mailing list
[email protected]
http://vpri.org/mailman/listinfo/fonc