Hello guys ,
I’ve got a bit of a situation I was wondering if somebody out there could help me with . I have started playing the newly realeased title Shadow Of The Tomb Raider today . I have an i7 6700k and 2 x gtx 1080ti’s . I’ve utilised both cards when running benchmarks . The game shows 3 scenes in the test . For 2 of the scenes my cards use about 90-95-100% usage with 70-100 FPS which is great . However in 1 of the scenes my cards only use about 55-60-65% usage and my FPS is around 55 or so . After running the benchmark the following information was provided . If you look closely at the graph you will see the orange dip and geeen rise in the middle . This was when my cards were not using there 90-95-100% usage and FPS was lower . I see information written here I am not certain of such as cpu game , cpu render. I see milliseconds written on the left of the graph. I’m just trying to determine what all this means and perhaps find out why my cards dip mid benchmark run with the information I provided .
Thank you
I’ve got a bit of a situation I was wondering if somebody out there could help me with . I have started playing the newly realeased title Shadow Of The Tomb Raider today . I have an i7 6700k and 2 x gtx 1080ti’s . I’ve utilised both cards when running benchmarks . The game shows 3 scenes in the test . For 2 of the scenes my cards use about 90-95-100% usage with 70-100 FPS which is great . However in 1 of the scenes my cards only use about 55-60-65% usage and my FPS is around 55 or so . After running the benchmark the following information was provided . If you look closely at the graph you will see the orange dip and geeen rise in the middle . This was when my cards were not using there 90-95-100% usage and FPS was lower . I see information written here I am not certain of such as cpu game , cpu render. I see milliseconds written on the left of the graph. I’m just trying to determine what all this means and perhaps find out why my cards dip mid benchmark run with the information I provided .
Thank you