Normal Map的一些细节问题

原创 2006年05月22日 14:12:00
Normal Map的一些细节问题
1.Noraml Map的用处
一是用在Bump Map中,表现粗糙或凹凸的表面。
二是用在High Polygon Detail for Low Polygon,即用低多边形模型表现高多边形细节。
2.Normal Map的生成
(1)对于Bump Map,通常是由Height Map生成的。公式的前提是:square patch assumption,“this assumption is acceptable for many important surfaces such as flat polygons, spheres, tori, and surfaces of revolution.”因此基本上够用了。
N’= N + D =[0,0,c]+ D =[a,b,c]
a = -dF/du * 1/k
b = -dF/dv  * 1/k
c = 1
通常K取1。而偏导(partial differentials)用有限差分(finite differences)求。也可看成是图像处理中的“锐化滤波”。通常可用索伯尔(sobel filter)算子求(注意:通常边界不太好处理,但有各种解决办法)。

     -1 0 1        1  2  1 

dx = -2 0 2  dy =  0  0  0

     -1 0 1       -1 -2 -1

  注意结果要单位化,并且由于法线的范围是[-1, 1],所以还要转换到颜色的范围[0, 1] (又即0~255)

      for(y = 0; y < gHeight; y++)
         for(x = 0; x < gWidth; x++)
            // Do Y Sobel filter
            ReadPixel(srcImage, &pix, (x-1+gWidth) % gWidth, (y+1) % gHeight);
            dY = ((float) / 255.0f * -1.0f;
            ReadPixel(srcImage, &pix,   x   % gWidth, (y+1) % gHeight);
            dY += ((float) / 255.0f * -2.0f;
            ReadPixel(srcImage, &pix, (x+1) % gWidth, (y+1) % gHeight);
            dY += ((float) / 255.0f * -1.0f;
            ReadPixel(srcImage, &pix, (x-1+gWidth) % gWidth, (y-1+gHeight) % gHeight);
            dY += ((float) / 255.0f * 1.0f;
            ReadPixel(srcImage, &pix,   x   % gWidth, (y-1+gHeight) % gHeight);
            dY += ((float) / 255.0f * 2.0f;
            ReadPixel(srcImage, &pix, (x+1) % gWidth, (y-1+gHeight) % gHeight);
            dY += ((float) / 255.0f * 1.0f;
            // Do X Sobel filter
            ReadPixel(srcImage, &pix, (x-1+gWidth) % gWidth, (y-1+gHeight) % gHeight);
            dX = ((float) / 255.0f * -1.0f;
            ReadPixel(srcImage, &pix, (x-1+gWidth) % gWidth,   y   % gHeight);
            dX += ((float) / 255.0f * -2.0f;
            ReadPixel(srcImage, &pix, (x-1+gWidth) % gWidth, (y+1) % gHeight);
            dX += ((float) / 255.0f * -1.0f;
            ReadPixel(srcImage, &pix, (x+1) % gWidth, (y-1+gHeight) % gHeight);
            dX += ((float) / 255.0f * 1.0f;
            ReadPixel(srcImage, &pix, (x+1) % gWidth,   y   % gHeight);
            dX += ((float) / 255.0f * 2.0f;
            ReadPixel(srcImage, &pix, (x+1) % gWidth, (y+1) % gHeight);
            dX += ((float) / 255.0f * 1.0f;
            // Cross Product of components of gradient reduces to
           nX = -dX;
            nY = -dY;
            nZ = 1;
            // Normalize
            oolen = 1.0f/((float) sqrt(nX*nX + nY*nY + nZ*nZ));
            nX *= oolen;
            nY *= oolen;
            nZ *= oolen;
     = (BYTE) PackFloatInByte(nX);
   = (BYTE) PackFloatInByte(nY);
   = (BYTE) PackFloatInByte(nZ);
            WritePixel(dstImage, &pix, x, y);
BYTE PackFloatInByte(float in)
   return (BYTE) ((in+1.0f) / 2.0f * 255.0f);
注意ATI在这里最后一行和一列用了WRAP的方法,即循环: % gWidth, % gHeight
例如:Photoshop plug-in from NVIDIA
(2)对于High Poly Detail for Low Poly,则相对要复杂一些。(例如Z-Brush)
An even cooler use for a normal map is to make a low res model look almost exactly like a high res model. This type of normal map is generated by the computer instead of painted like a bump map. Here is how it works: First you create two versions of the model: a high polygon version (which can contain as much detail as you want) and a low polygon version that will actually get used in the game. Then you align the two models so that they occupy the same space and overlap each other.
Next you run a special program for generating the normal map. The program puts an empty texture map on the surface of the low res model. For each pixel of this empty texture map, the program casts a ray (draws a line) along surface normal of the low res model toward the high res model. At the point where that ray intersects with the surface of the high res model, the program finds the high res model normal. The idea is to figure out which direction the high res model surface is facing at that point and put that direction information (normal) in the texture map.
Once the program finds the normal from the high res model for the point, it encodes that normal into an RGB color and puts that color into the current pixel in the afore mentioned texture map. It repeats this process for all of the pixels in the texture map. When it`s done, you end up with a texture map that contains all of the normals calculated from the high res model. It`s ready to be applied to the low res model as a normal map. I抣l show you how to create this type of normal map in the second half of the tutorial.
3. TBN的计算
T = normalize(dx/du, dy/du, dz/du)
N = T × normalize(dx/dv, dy/dv, dz/dv)
B = N × T
(Tangent space is just such a local coordinate system. The orthonormal basis for the tangent space is the normalized unperturbed surface normal Nn, the tangent vector Tn defined by normalizing dP/du, and the binormal Bn defined as Nn×Tn. The orthonormal basis for a coordinate system is also sometimes
called the reference frame.)
(D3DX提供了两个帮助函数:D3DXComputeTangentFrame, D3DXComputeTangentFrameEx)
由仿射坐标系算法:V = x*X + y*Y + O:
例如:V1 = c1*T + c2*B + O。将V1,V2,V3分别写成公式后相减得:
v2v1 =   △c2c1T T + △c2c1B B
v3v1 =   △c3c1T T + △c3c1B B
问题1:这样得到的TB不一定正交(orthogonal)。是否需要通过:N = T X B, T = B X N(若保留B)B = N X T(若保留T)呢?
实际由于纹理坐标的映射通常是在正交的情况下作的(平面的坐标基,又称为reference frame),而Normal Map的映射与Base Texture的映射是一致的,所以可以认为TB是正交的。当然,这种假设不一对,最好看情况而定。
[in] Specifies the maximum cosine of the angle at which two partial derivatives are deemed to be incompatible with each other. If the dot product of the direction of the two partial derivatives in adjacent triangles is less than or equal to this threshold, then the vertices shared between these triangles will be split.
[in] Specifies the maximum magnitude of a partial derivative at which a vertex will be deemed singular. As multiple triangles are incident on a point that have nearby tangent frames, but altogether cancel each other out (such as at the top of a sphere), the magnitude of the partial derivative will decrease. If the magnitude is less than or equal to this threshold, then the vertex will be split for every triangle that contains it.
[in] Similar to fPartialEdgeThreshold, specifies the maximum cosine of the angle between two normals that is a threshold beyond which vertices shared between triangles will be split. If the dot product of the two normals is less than the threshold, the shared vertices will be split, forming a hard edge between neighboring triangles. If the dot product is more than the threshold, neighboring triangles will have their normals interpolated.

               Tx   Bx   Nx

Lts = Los ×  Ty   By   Ny

                       Tz    Bz   Nz

即两个坐标系之间的转换(object -> tangent),原理可参考world space -> view space。由于光线L和半角H是矢量,所以第四行的平移量就不用了。
这里N是平面原有的法线(unperturbed normal);N’是Normal Map中的perturbed normal
可以通过(Lz< 0) ? 0 : 1来解决。但效果不太好: Simple, but may result in “hard” shadow boundary。Possible winking and popping of highlights。NVIDIA的同志(文献[2])提出了这样一个经验公式:min( 8*max(Lz,0), 1)
Sself = min( 8*max(Lz,0), 1)。
I = Ia + Sself * ( Kd*max(0, L·N’) + Ks*max(0, N’·H)shininess )
5. 使用normalization cub map
由于要在VS中频繁地使用单位向量,而Normalize的花费比较大。所以通常使用预计算的normalization cub map来加速,即单位向量的查找表。
生成算法显然是: cube[x,y,z] = normalized(x, y, z),不要忘了range-compress:[-1, 1] => [0, 1]
6. 纹理过滤对法线的影响及diffuse与specular计算时的一些区别
在文献[1]中提到:因为从Normal Map中得到是经过纹理过滤后的扰动法线,所以它的结果并一定是单位化的。
Note that N’filtered is not generally normalized. Renormalizing N’filtered might seem reasonable and has been suggested in the literature [21], but not renormalizing when computing surface’s perceived diffuse intensity is actually more correct.
Note that if N’filtered is renormalized then bumpy diffuse surfaces incorrectly appear too bright. This result suggests that linear averaging of normalized perturbed normals is a reasonable filtering approach for the purpose of diffuse illumination. It also suggests that if bump maps are represented as perturbed normal maps then such normal maps can be reasonably filtered with conventional linear-mipmap-linear texturing hardware when used for diffuse lighting.
The single good thing about this approach is that there is an opportunity for sharing the same bump map encoding between the diffuse and specular illumination computations. The diffuse computation assumes an averaged perturbed normal that is not normalized while the specular computation requires the same normal normalized. Both needs are met by storing the normalized perturbed normal and a descaling factor. The specular computation uses the normalized normal directly while the diffuse computation uses the normalized normal but then fixes the diffuse illumination by multiplying by the descaling factor to get back to the unnormalized vector needed for computing the proper diffuse illumination.
A scaling factor for the normal. Mipmap level 0 has a constant magnitude of 1.0, but down-sampled mipmap levels keep track of  the unnormalized vector sum length. For diffuse per-pixel lighting, it is preferable to make N` be the _unnormalized_ vector, but for specular lighting to work reasonably, the normal vector should be normalized. In the diffuse case, we can multiply by the "mag" to get the possibly shortened unnormalized length.
7. mipmap的使用
通常Normal Map也应该使用mipmap,因为如果base/decal texture使用了mipmap,而normal map不使用则显然对应不上。normal map的mipmap级应该和base texture一致。
而由于第6点提到的技巧,因此normal map的mipmap应该由程序手动实现,而不要让d3d或硬件自动创建,因为:自动创建并不会单位化,且通常我们还要在alpha通道里存法线的长度(diffuse)
8. 什么情况下更新TBN?
  显然只有当object空间中顶点的位置发生变化时才更新(旋转,非均匀缩放Uniform scaling),通常为Mesh的动画。
 (1)如果是骨骼动画(Bone-Based Skinning),则直接用顶点变换矩阵Matrix作用于T和B,只算出T和B即可,然后
T X B = N
 (2)如果是基于Mesh的关键帧插值(Keyframe Interpolation),同样对T和B进行插值。
[1] Mark J. Kilgard, A Practical and Robust Bump-mapping Technique for Today`s GPUs, NVIDIA Corporation, July 2000
[2] Chris Wynn, Implementing Bump-Mapping using Register Combinerss, NVIDIA Corporation, August 2001
[3] Sim Dietrich, Texture Space Bump Mapping, NVIDIA Corporation, November 2000
[4] Jakob Gath, Derivation of the Tangent Space Matrix,
[5] BEN CLOWARD, Normal Mapping Tutorial,


凹凸贴图(Bump Map)实现原理以及与法线贴图(Normal Map)的区别

凹凸贴图(Bump Map)实现原理 以及与法线贴图(Normal Map)的区别 1 前言 翻译这篇教程的目的是为了帮助那些对图形渲染技术有兴趣却又苦于找不到免费中文学习资料的...

对于法线贴图(Normal Map) 的深入研究

前几篇文章写过有关法线贴图的内容,这次文章将讨论其原理及相关优化。回过头来看一下原来的文章真有种想删掉的感觉。。。 为什么叫法线贴图,我们知道法线(Normal)是垂直于一个面的直线,通过计算光...

OpenGL学习笔记:bump mapping

bump mapping是为了制作浮雕效果。(In 1978, James Blinn presented a method of performing what is called bump map...


Unity没有建模工具,大多是在其他3D软件中建模后导入Unity资源中使用。但通过脚本可以修改网格的顶点属性,所以理论上可以实现在Unity中从无到有地建模。 建模,就是建网格;建网格,就是画些...

《Unity Shader 与 计算机图形学》第二章

提示:本篇将会非常长~ 本系列文章分为硬件编程入门工程实践上一篇主要介绍了GPU的特征工作原理 以及渲染的底层流程 其实对于新架构而言还有所不同

C# 通过socket实现UDP 通信





今天突然发现操作Linux的时候不知道快捷键是多么的麻烦,于是在网上找了常用的几个快捷键。 1.前一个后一个工作区的切换 如果你经常使用工作区,那你就可以用Ctrl + Alt +上/下方向键很容易地...

网易面试总结(机器学习一些面试) 之前投过了网易内推,笔试过了,但是一直没通知我面试。我当时想着可能没戏了...

KlayGE Normal2Height工具

转载请注明出处为KlayGE游戏引擎,本文的永久链接为 多年前就想为KlayGE做个Normal2Height的工具,一直没时间,现在终于抽...