Does each vertex have its own normals or does each triangle have its own normals?

Like so often the answer is: “It depends”. Since a normal is defined as being the vector perpendicular to all vectors within a given plane (in N dimensions), you need a plane to calculate a normal. A vertex position is just a point and thus singular, so you actually need a face to calculate the normal. Thus, naively, one could assume that normals are *per face* as the first step in normal calculation is determining the face normals, by evaluating the cross product of the faces edges.

Say you have a triangle with points **A**, **B**, **C**, then these points have position vectors **↑A**, **↑B**, **↑C** and the edges have vectors **↑B – ↑A** and **↑C – ↑A** so the face normal vector is **↑N _{f} = (↑B – ↑A) × (↑C – ↑A)**

Note that the magnitude of **↑N _{f}** as it’s stated above is directly proportional to the face’s area.

In smooth surfaces vertices are shared between faces (or you could say those faces share a vertex). In that case the normal at the vertex is not one of the face normals of the faces it is part of, but a linear combination of them:

**↑N _{v} = ∑ p ↑N_{f}** ; where

**p**is a weighting for each face.

One could either assume a equal weighting between the participating face normals. But it makes more sense to assume that the larger a face is, the more it contributes to the normal.

Now recall that you normalize by a vector **↑v** by scaling it with it’s recipocal length: **↑v _{i} = ↑v/|↑v|**. But as already told the length of the face normals already depends on the face’s area. So the weighting factor

**p**given above is already contained in the vector itself: Its length, aka magnitude. So we can get the vertex normal vector by simply summing up all the face normals.

In lighting calculations the normal vector must be unit length, i.e. normalized to be useable. So after summing up, we normalize the newly found vertex normal and use that.

The carefull reader may have noticed I specifically said *smooth* surfaces share vertices. And in fact, if you have some creases / hard edges in your geometry, then the faces on either side don’t share vertices. In OpenGL a vertex is the whole combination of

- position
- normal
- (colour)
- N texture coordinates
- M further attributes

You change one of these and you got a completely different vertex. Now some 3D modelers see a vertex only as a point’s position and store the rest of those attributes per face (Blender is such a modeler). This saves some memory (or considerable memory, depending on the number of attributes). But OpenGL needs the whole thing, so if working with such a mixed paradigm file you will have to decompose it into OpenGL compatible data first. Have a look at one of Blender’s export scripts, like the PLY exporter to see how it’s done.

Now to cover some other thing. In your code you have this:

` glIndexPointer( GL_UNSIGNED_BYTE, 0, indices );`

The index pointer has **nothing** to do with vertex array indices! This is an anachronsim from the days, when graphics still used palettes instead of true color. A pixels colour wasn’t set by giving it’s RGB values, but by a single number offsetting into a limited palette of colours. Palette colours can still be found in several graphics file formats, but no decent piece of hardware uses them anymore.

Please erase *glIndexPointer* (and glIndex) from your memory and your code, they don’t do what you think they do The whole indexed color mode is arcane to used, and frankly I don’t know of any hardware built after 1998 that still supported it.

https://stackoverflow.com/questions/6656358/calculating-normals-in-a-triangle-mesh/6661242#6661242