Tag Archives: NVIDIA

Top 10 Slow Motion Video Editing Software List by Filmora!

Top 10 Slow Motion Video Editing Software

The video editing software Filmora has compounded a list of the top ten slow motion video editing software packages for personal computers. While you can clearly see this is in some way an article to attract an audience for their product, you can also learn about products you didn’t know existed as options. Even Filmora was to us a product that was not really on the radar but now is. Especially at the cheap price of $59.99 for a lifetime license with no subscriptions. We are not advertising any product, in particular, You can see the Top Ten Slow Mo Software List Here!

Our software of choice has always been Adobe After Effects with the Time Warp feature but we can understand why the subscription model for Adobe CC products can be a little too much to ask for many. The list here also shows completely free products like Avi Synth which is very powerful but not easy to use and other optical flow software that can really slow regular video and higher frame rates to a crawl. However, we consider interpolation fake slow motion as we noted in our article here!  Nvidia’s new Machine Learning Algorithm is very impressive and shows the way forward for converting regular video to higher frame rates with surprisingly amazing results. Tell us what you think – HSC

NVIDIA Slow Motion Interpolation With AI Deep Learning Tech!

NVIDIA Slow Motion Interpolation

NVIDIA has been hard at work on the problem posed by high frame rate interpolation of video data shot on lower fps.  We have had this tech since the late 1990s with the advent of Twixtor and refined over the decades in systems like Twixtor Pro and Adobe’s Optical Flow in After Effects. You are still not getting real temporal detail data since the frames are created by extrapolating velocity and direction vectors plus pixel values between frames to get the result.

We explored this technique in our post on interpolation here and why it is no substitute from a real slow motion camera solution.  NVIDIA’s new method uses machine learning along with 11,000 videos to arrive at a more convincing result. Considering the relatively small sample size we can imagine a future where hundreds of thousands or millions of footage samples are used to generate near flawless interpolation. This technique takes some serious computation and data sets so as of now it is not really ready for the mass market but that could change with the cloud very soon. → Continue Reading Full Post ←