You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have a relatively small YouTube channel and occasionally upload frame rate analysis videos. Like many, I capture lossless or as close to lossless as I can using OBS.
90% of the time my readings (to my knowledge) have been fairly accurate. However, I am currently analyzing a 30fps game play clip, and I may be wrong but something just doesn't feel quite right and I am not prepared to put out any content that isn't as accurate as it can possibly be.
There are a few instances in the clip where the frame rate is locked at 30 and then goes in a sequence 30-29-30-29-30-29-30-29-30-29 etc.
Looking at the delta rendering when this occurs, 2 black frames are displayed with the final black frame showing that the fps has now dropped to 29 and then begins the sequence white-black-white-black-white. I believe this is the correct behaviour? But just wanted to double check. As far as I know the footage is as lossless as I can get it, bit rate of 2607 mb/s. It just seemed odd that there are a few instances in separate clips of the same game where the fps just drops from 30 and runs in a sequence 30-29-30-29-30-29-30-29-30-29. I just wasn't sure if that was actually happening in the game or something wrong with the capture.
I have tried changing the pixel difference anywhere from 0-20 but it makes absolutely no difference to the frame rate that trdrop is detecting.
Any advice/assistance would be greatly appreciated.
The text was updated successfully, but these errors were encountered:
Hi all,
I am hoping someone may be able to assist me?
I have a relatively small YouTube channel and occasionally upload frame rate analysis videos. Like many, I capture lossless or as close to lossless as I can using OBS.
90% of the time my readings (to my knowledge) have been fairly accurate. However, I am currently analyzing a 30fps game play clip, and I may be wrong but something just doesn't feel quite right and I am not prepared to put out any content that isn't as accurate as it can possibly be.
There are a few instances in the clip where the frame rate is locked at 30 and then goes in a sequence 30-29-30-29-30-29-30-29-30-29 etc.
Looking at the delta rendering when this occurs, 2 black frames are displayed with the final black frame showing that the fps has now dropped to 29 and then begins the sequence white-black-white-black-white. I believe this is the correct behaviour? But just wanted to double check. As far as I know the footage is as lossless as I can get it, bit rate of 2607 mb/s. It just seemed odd that there are a few instances in separate clips of the same game where the fps just drops from 30 and runs in a sequence 30-29-30-29-30-29-30-29-30-29. I just wasn't sure if that was actually happening in the game or something wrong with the capture.
I have tried changing the pixel difference anywhere from 0-20 but it makes absolutely no difference to the frame rate that trdrop is detecting.
Any advice/assistance would be greatly appreciated.
The text was updated successfully, but these errors were encountered: