I found no documentation for this behaviour. But tests with nvidia's drivers show that only GL_UNSIGNED_BYTE gets normalized.
Wow depends on this behaviour if GL_NV_vertex_program is present.
---
 src/mesa/main/varray.c |    3 ++-
 1 files changed, 2 insertions(+), 1 deletions(-)
diff --git a/src/mesa/main/varray.c b/src/mesa/main/varray.c
index 11720e4..f41ee73 100644
--- a/src/mesa/main/varray.c
+++ b/src/mesa/main/varray.c
@@ -443,7 +443,7 @@ void GLAPIENTRY
 _mesa_VertexAttribPointerNV(GLuint index, GLint size, GLenum type,
                             GLsizei stride, const GLvoid *ptr)
 {
-   const GLboolean normalized = GL_FALSE;
+   GLboolean normalized = GL_FALSE;
    GLsizei elementSize;
    GET_CURRENT_CONTEXT(ctx);
    ASSERT_OUTSIDE_BEGIN_END(ctx);
@@ -471,6 +471,7 @@ _mesa_VertexAttribPointerNV(GLuint index, GLint size, GLenum type,
    /* check for valid 'type' and compute StrideB right away */
    switch (type) {
       case GL_UNSIGNED_BYTE:
+         normalized = GL_TRUE;
          elementSize = size * sizeof(GLubyte);
          break;
       case GL_SHORT:

-------------------------------------------------------------------------
This SF.net email is sponsored by: Microsoft
Defy all challenges. Microsoft(R) Visual Studio 2008.
http://clk.atdmt.com/MRT/go/vse0120000070mrt/direct/01/
_______________________________________________
Mesa3d-dev mailing list
Mesa3d-dev@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/mesa3d-dev

Reply via email to