I''m using MFC in my OpenGL app to make development of dialogs, etc, easier. At the start of the app, I create a 1 millisecond timer to trigger my rendering function. Now I don''t expect 1000 fps, but I do expect it to move along as fast as it can. However, I''m only getting 20 fps regardless of video resolution: everything from 640x480 to 1800x1440 all run at the same speed, 20 fps, in both debug and release modes.
Why??? And how can I change this so it runs as fast as possible without giving up MFC?
512 Meg RAM
ATI All-in-Wonder Radeon, 32M DDR with latest drivers