Sign in to follow this  

OpenGL OpenGL Perspective Projection and the View Plane

Recommended Posts

RasterGuy    110
Hi guys,

I'm having a little trouble wrapping my mind completely around the perspective projection and the relationship to the view plane. A little background:[list]
[*]I'm working on a 3D Software Renderer.
[*]I'm reading Lengyel's "Mathematics for 3D Game Programming and Computer Graphics" (I've added a link at the end to a preview in Google Books, has a good chunk of what I'm reading at least, pp. 111 and on.)
So, I understand that a basic perspective projection can be achieved with (obviously the matrix is over kill):

[ 1 0 0 0 ]
[ 0 1 0 0 ]
[ 0 0 1 0 ]
[ 0 0 1/e 1 ]

Where e is the distance to the view-plane(focal length) as a function of the horizontal FOV.

This would correctly project points/vertices, but would not perform any clipping.

I know that the OpenGL projection matrix transforms all of the points into homogeneous clip-space (transforming the view-frustum to a cuboid), and makes clipping to the frustum much more simple. (The matrix can be seen on pp. 124 linked below, labeled Mfrustum)

I also see how the focal length determines the normals from the frustum planes, as well as the aspect ratio's role in the frustum plane normals.

What I don't understand completely, is that, both e(focal length) and the aspect ratio seem to have no contribution to the perspective projection matrix. Is the actual projection a final step after transforming to clip-space (and doing the clipping)? Is OpenGL doing something extra behind the scenes? (recall, I'm working on a software renderer) More over, what is the view-plane's relation to the view-frustum?

I hope my question makes sense, if not I will be happy to rephrase or elaborate. I appreciate and knowledge you can share.

Book Preview:

[url=""] normal&f=false[/url] Edited by RasterGuy

Share this post

Link to post
Share on other sites
Brother Bob    10344
The focal length and aspect ratio are determined from the parameters used in the equation in the book. For example, the aspect ration is [i](r-l)/(t-b)[/i] and the focal length (as you appear to be defining it) is [i]n/(t-b)[/i].

Share this post

Link to post
Share on other sites
RasterGuy    110
[quote name='Brother Bob' timestamp='1348700869' post='4984154']
The focal length and aspect ratio are determined from the parameters used in the equation in the book. For example, the aspect ration is [i](r-l)/(t-b)[/i] and the focal length (as you appear to be defining it) is [i]n/(t-b)[/i].

Thanks Brother Bob,

I guess I need to go over it some more. It seems like, in the book, the near plane is actually replacing the view/projection plane. Is that right? I apologize if this is obvious, but I seem to stumbling on it for one reason or another.

Share this post

Link to post
Share on other sites
Brother Bob    10344
There really is no [i]the[/i] projection plane, but [i]a[/i] projection plane. Within the view volume, you can slide the projection plane as you like along the depth axis without changing what is projected onto the plane relative the size of the view volume at the depth where the projection plane is positioned. You can think of the near plane as the projection plane, but you can also think of the far plane as the projection plane, or almost [i]any[/i] plane along the depth axis whether it's inside or outside the actual view volume.

If you want to think of a concrete projection plane to visualize how everything works, the near plane is probably the easiest plane to think of as the projection plane.

Keep in mind here that talking about an absolute focal length is not really relevant as it says nothing about how the view volume looks. The focal length only makes sense when you also have the size of a projection plane. In that case, the relevant number is the ratio between the focal length and the size of the projection plane. For a camera analogy, a 50mm focal length lens says nothing about what the picture from it will look like (I'm talking about perspective effects only), unless you also say that you either have, say, a 24mm or 35mm film/sensor. The ratio between the two is related to the tangent of the field of view.

If you look at the equation I mentioned about the focal length, you will see that it is a normalized focal length. The value [i]n/(b-t)[/i] is the ratio of the distance to the near plane and the size of the near clip plane (in the vertical direction in this case). This is the general case where you can define the size of the near clip plane and the distance of it from the view point. In your case, you have an implicit size of the projection plane equal to 2 (from -1 to +1), and your parameter [i]e[/i] is in relation to this size.

Share this post

Link to post
Share on other sites
Lightness1024    933
I'm pretty sure you are missing only one thing, because it is an obscure one...
everything in the projected vector is divided by w afterward by the rasterizer.

so, the projection matrix actually doesn't do anything really, apart from copying z into w. then the real "perspective division", like we call it, is done in a hidden step afterward.
you can do it yourself if you are coding a projection by hand on CPU (or even in shader) by doing Vproj = Vworld * WorldViewProjMatrix; Vproj /= Vproj.w;

there you go.

it is really easy to understand actually : the farther the points, the closer to the center of the view. that is how one creates parallax effects. (think of scolling planes with different speeds according to distances, or also, think of convergence lines in a drawing of a cube. further points are closer together.

then the focal enters in the equation as a multiplier for the perspective effect. the larger the FOV the stronger the "gather to the center division effect" will be. that is all :)

Share this post

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now

Sign in to follow this  

  • Similar Content

    • By pseudomarvin
      I assumed that if a shader is computationally expensive then the execution is just slower. But running the following GLSL FS instead just crashes
      void main() { float x = 0; float y = 0; int sum = 0; for (float x = 0; x < 10; x += 0.00005) { for (float y = 0; y < 10; y += 0.00005) { sum++; } } fragColor = vec4(1, 1, 1 , 1.0); } with unhandled exception in nvoglv32.dll. Are there any hard limits on the number of steps/time that a shader can take before it is shut down? I was thinking about implementing some time intensive computation in shaders where it would take on the order of seconds to compute a frame, is that possible? Thanks.
    • By Arulbabu Donbosco
      There are studios selling applications which is just copying any 3Dgraphic content and regenerating into another new window. especially for CAVE Virtual reality experience. so that the user opens REvite or CAD or any other 3D applications and opens a model. then when the user selects the rendered window the VR application copies the 3D model information from the OpenGL window. 
      I got the clue that the VR application replaces the windows opengl32.dll file. how this is possible ... how can we copy the 3d content from the current OpenGL window.
      anyone, please help me .. how to go further... to create an application like VR CAVE. 
    • By cebugdev
      hi all,

      i am trying to build an OpenGL 2D GUI system, (yeah yeah, i know i should not be re inventing the wheel, but this is for educational and some other purpose only),
      i have built GUI system before using 2D systems such as that of HTML/JS canvas, but in 2D system, i can directly match a mouse coordinates to the actual graphic coordinates with additional computation for screen size/ratio/scale ofcourse.
      now i want to port it to OpenGL, i know that to render a 2D object in OpenGL we specify coordiantes in Clip space or use the orthographic projection, now heres what i need help about.
      1. what is the right way of rendering the GUI? is it thru drawing in clip space or switching to ortho projection?
      2. from screen coordinates (top left is 0,0 nd bottom right is width height), how can i map the mouse coordinates to OpenGL 2D so that mouse events such as button click works? In consideration ofcourse to the current screen/size dimension.
      3. when let say if the screen size/dimension is different, how to handle this? in my previous javascript 2D engine using canvas, i just have my working coordinates and then just perform the bitblk or copying my working canvas to screen canvas and scale the mouse coordinates from there, in OpenGL how to work on a multiple screen sizes (more like an OpenGL ES question).
      lastly, if you guys know any books, resources, links or tutorials that handle or discuss this, i found one with marekknows opengl game engine website but its not free,
      Just let me know. Did not have any luck finding resource in google for writing our own OpenGL GUI framework.
      IF there are no any available online, just let me know, what things do i need to look into for OpenGL and i will study them one by one to make it work.
      thank you, and looking forward to positive replies.
    • By fllwr0491
      I have a few beginner questions about tesselation that I really have no clue.
      The opengl wiki doesn't seem to talk anything about the details.
      What is the relationship between TCS layout out and TES layout in?
      How does the tesselator know how control points are organized?
          e.g. If TES input requests triangles, but TCS can output N vertices.
             What happens in this case?
      In this article,
      the isoline example TCS out=4, but TES in=isoline.
      And gl_TessCoord is only a single one.
      So which ones are the control points?
      How are tesselator building primitives?
    • By Orella
      I've been developing a 2D Engine using SFML + ImGui.
      Here you can see an image
      The editor is rendered using ImGui and the scene window is a sf::RenderTexture where I draw the GameObjects and then is converted to ImGui::Image to render it in the editor.
      Now I need to create a 3D Engine during this year in my Bachelor Degree but using SDL2 + ImGui and I want to recreate what I did with the 2D Engine. 
      I've managed to render the editor like I did in the 2D Engine using this example that comes with ImGui. 
      3D Editor preview
      But I don't know how to create an equivalent of sf::RenderTexture in SDL2, so I can draw the 3D scene there and convert it to ImGui::Image to show it in the editor.
      If you can provide code will be better. And if you want me to provide any specific code tell me.
  • Popular Now