Does 30fps feel more "cinematic" than 60fps?

Doesn't make any sense. Avatar 2 is going to need BR sales as much as any other movie, and there's no way to get enough TVs to support 72Hz and 3D (which requires 144 frames per second altogether) in time.
Not to mention storage related issues, a 72fps stereo movie requires six times as much space on the disc, I'm not sure if it's even possible at all.

Support for 48fps is going to be a challenge enough on its own and we should be happy to get just that much on our home 'cinemas' in 2 years..

72fps has the magical look that is pleasing to the eye. Pioneer plasmas supporrted 24hz material at 72hz yielding a nice silky smooth picture. Obviously there are many factors involved but 72 frames is the golden number.

72hz @ 4k would be the right thing to do. Sony constantly has pushed 4k resolution...give Cameron what he wants with 72hz.

1080p to 4k is roughly a 4x increase.

24hz to 72hz a 3x increase.

This is the logical path imho and more importantly it will look more natural.
 
72fps has the magical look that is pleasing to the eye. Pioneer plasmas supporrted 24hz material at 72hz yielding a nice silky smooth picture. Obviously there are many factors involved but 72 frames is the golden number.

72hz @ 4k would be the right thing to do. Sony constantly has pushed 4k resolution...give Cameron what he wants with 72hz.

1080p to 4k is roughly a 4x increase.

24hz to 72hz a 3x increase.

This is the logical path imho and more importantly it will look more natural.

The problem is that 72fps is not a standard for anything. 60fps or 120fps are a standard already, so the best choice would be one of these 2.

RED can shoot at 4k@120fps already, and the 4k projectors can already show at 4k@120fps, after server upgrade.
 
72hz for a proper cinematic standard. If your going to do something do it right.
But why?
120fps is the perfect standard, and encompasses the advantages of 72fps, plus having the same standard for everything (movies, tv shows, games, ecc...) allows huge money savings.
 
But why?
120fps is the perfect standard, and encompasses the advantages of 72fps, plus having the same standard for everything (movies, tv shows, games, ecc...) allows huge money savings.

72hz and the flicker fusion threshold.


72hz is a sweet spot for frame rate.
 
Well I guess at some point you have to define a standard. Plus there are costs with special effects.

Cameron has been about higher frame rates while Sony wanted 4k.

With Europe at 50hz and the USA and Japan at 60hz it is a good opportunity to unify at 72hz which research shows to be a good frame rate.
 
Well I guess at some point you have to define a standard. Plus there are costs with special effects.

Cameron has been about higher frame rates while Sony wanted 4k.

With Europe at 50hz and the USA and Japan at 60hz it is a good opportunity to unify at 72hz which research shows to be a good frame rate.

Again, 72hz is just the point where emotional response doesn't go up, and it's an arbitrary number.
And, again, 120hz would be the best unification framerate, since it's the same used by other media as well...
 
The first 48hz film hits studios and you expect them to nearly triple it immediately. You might want to hold on until the 2nd hfr movie hits theaters before arguing about a better standard.

Whatever it cost them making hobbit 48fps probably would have been better spent on better writing to make a less silly film.
 
The first 48hz film hits studios and you expect them to nearly triple it immediately. You might want to hold on until the 2nd hfr movie hits theaters before arguing about a better standard.
Again: the camera can already shoot at 4k@120fps 3D, and the 4K projectors can already play at 120fps (some just need server upgrade), so it's not a tech limitation.

Whatever it cost them making hobbit 48fps probably would have been better spent on better writing to make a less silly film.
Who cares about the movie, the important thing is the technology, in the same way Avatar brought us that incredible 3D and the-state-of-the-art CG and earned bilions, even if the story was pretty much irrelevant.
 
Again: the camera can already shoot at 4k@120fps 3D, and the 4K projectors can already play at 120fps (some just need server upgrade), so it's not a tech limitation.

Because that's not the only cost.

Who cares about the movie, the important thing is the technology, in the same way Avatar brought us that incredible 3D and the-state-of-the-art CG and earned bilions, even if the story was pretty much irrelevant.

I don't think the tech is important if it doesn't lend itself to a better movie. The Hobbit is currently making a crappy case for the tech.
 
The first 48hz film hits studios and you expect them to nearly triple it immediately. You might want to hold on until the 2nd hfr movie hits theaters before arguing about a better standard.

Whatever it cost them making hobbit 48fps probably would have been better spent on better writing to make a less silly film.

A 72hz spec doesn't mean studios have to shoot at that. Let the director decide 24...48...72...but put it in the specs. If Cameron wants 72hz at 2k reseloution let him.
 
The first 48hz film hits studios and you expect them to nearly triple it immediately. You might want to hold on until the 2nd hfr movie hits theaters before arguing about a better standard.

Yeah, has noone watched the documentaries on Weta's work?

The 4K source material helped them with all the 2D pre-processing, like rotoscoping elements, painting out actors and equipment and wires and such, but after that all the CG work was done at 2K. It's still more detailed than what you'd get on film stock, no grain and far less blur.

On the other hand they had to use 4 times as much rendering power, and 4 times as much storage space - 2x for stereo and 2x for the 48fps. They had like 5-6 petabytes of data for the show altogether, far more than Avatar and that had a lot more CG and full CG shots.

Now Weta's the largest or maybe second largest VFX studio with a lot of high profile work - smaller shops just don't have the money to upgrade their infrastructure. And you guys are dreaming about an increase of 24x or even 40x?? Even if the studios were willing to spend that amount of money on hardware, it'd still mean total blackouts in New Zealand, California, London and Canada, because of the power draw...
 
Again, 72hz is just the point where emotional response doesn't go up, and it's an arbitrary number.
And, again, 120hz would be the best unification framerate, since it's the same used by other media as well...

If you are refeering to Trumbolts (spelling?) experiments, he could not project at anything higher than 72 (IIRC).
 
Yeah, has noone watched the documentaries on Weta's work?

The 4K source material helped them with all the 2D pre-processing, like rotoscoping elements, painting out actors and equipment and wires and such, but after that all the CG work was done at 2K. It's still more detailed than what you'd get on film stock, no grain and far less blur.

On the other hand they had to use 4 times as much rendering power, and 4 times as much storage space - 2x for stereo and 2x for the 48fps. They had like 5-6 petabytes of data for the show altogether, far more than Avatar and that had a lot more CG and full CG shots.

Now Weta's the largest or maybe second largest VFX studio with a lot of high profile work - smaller shops just don't have the money to upgrade their infrastructure. And you guys are dreaming about an increase of 24x or even 40x?? Even if the studios were willing to spend that amount of money on hardware, it'd still mean total blackouts in New Zealand, California, London and Canada, because of the power draw...
Well, can avoid exxagerating with CG, and do more things the old way, until computing power catches up...
 
If you are refeering to Trumbolts (spelling?) experiments, he could not project at anything higher than 72 (IIRC).
From what I remember from the interview, he said he kept going higher, but there was no emotional response increase from the audience.
 
Well, can avoid exxagerating with CG, and do more things the old way, until computing power catches up...

What to avoid? If you want 72 fps, stereo, and all this at 4K, it's going to take 24 times as much space to store the images and about the same extra time to calculate them. You can't really skip on any part, and the quality expectations are far too high to experiment with interpolation techniques.

And keep in mind that during production, all intermediate steps have to be adjusted too. An average shot can be re-rendered 10 to 100 times and most versions are kept online as general practice, it's far too problematic to try to selectively delete previous versions. 4K usually also requires more detailed assets requiring more space and more work, and animating for 72 fps is more work compared to 24fps, too. It's quite different to a PC game where you just change a few options in the video settings.
 
Back
Top