I would like to see some other animations done using the same technique. Baker wrote:Looks really good from what I can tell. That would be much better, as then it could be adaptive by using the amount of motion at a given frame to determine how many sub-frames to generate. Still not sure how to make videos look good on YouTube so you can download the original x264 AVI files from here:Īs you can see, it's actually not that complex, so I'm hoping that Mike will be able to integrate it into AS one day. It then took another 15 minutes to integrate them. Compose integrated PNGs into a video file.įor a 10 second animation (330 frames at 30 fps), it took about 15 minutes to render all 33000 subframes and the resulting PNG files took up about 500MB. I wrote this myself, and it's not too complicated, but if anyone's interested in obtaining it, they can PM me.Ĥ. Run Python Imaging Library script to perform the integration on the sub-frames. This gives 100 sub-frames per animation frame, meaning that an object can move up to 100 pixels without disrupting the blur effect.ģ. Rescale entire document by a factor of 100. As it turns out, this is already an existing technique called "temporal oversampling", that is used by most 3d animation programs that support high quality motion blur.ġ. I tried to emulate what happens with a real film camera by generating all bunch of sub-frames for each animation frame and integrating them all to get the resulting exposure. I was frustrated that there seem to be no free solutions for adding realistic looking motion blur to video, so I came up with my own.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |