Getting poly normal and CreatePhongNormals()

Hi folks,
I'm trying to calculate a polygon's normal. I've read CreatePhongNormals() can return an array of (vertices point?) normals if there's a Phong tag. How do I use that to calculate a normal at poly UV?
// psuedo code po = (PolygonObject*)obj; SVector *norms = po>CreatePhongNormals(); if(norms) { int ply = 4; // let's get normal for polygon 4(5), points abc Vector uv; // the uv spot of the normal to get, on ply Vector normal = (norms>operator[](ply*4) + norms>operator[](ply*4+1) + norms>operator[](ply*4+2)) / 3; // do something with normal, etc. GeFree(norms); }
Would like to include blending across the polygon if possible (i.e. the normal at spot uv). Not worried about speed, just learning.
WP.

Hi,
CreatePhongNormals only create the vertex normals accordingly to the phong tag.
This tag, display those normals, you can change the phong tag setting and a plane's shape to understand them a bit better.import c4d #Welcome to the world of Python import random colors = [] def draw (bd): global colors obj = op.GetObject() if obj is None: raise ValueError("can't retrieve the object") # retrieves the information about the object points = obj.GetAllPoints() normals = obj.CreatePhongNormals() # scale the normals normals = [i * 100 for i in normals] polyCnt = obj.GetPolygonCount() for i in range(polyCnt): # gets the polygon cpoly structure cpoly = obj.GetPolygon(i) bd.SetPen(colors[i]) pointsIndexes = [cpoly.a, cpoly.b, cpoly.c, cpoly.d] for j, pointIndex in enumerate(pointsIndexes): # retrieves the point position vecStart = points[pointIndex] # retrieves the normal direction. (add the point position) vecEnd = vecStart + normals[i * 4 + j] bd.DrawLine (vecStart, vecEnd, c4d.NOCLIP_D ) def main(): # generates enough random color (1 per poly) global colors random.seed(183446) obj = op.GetObject() if obj is None: raise ValueError("can't retrieve the object") polyCnt = obj.GetPolygonCount() for i in range(polyCnt): colors.append(c4d.Vector (random.random(), random.random(), random.random()))
Those phong normals are an average of every vertex normals a vertex have. (a vertex have as many normal than the number of polygon it's part of)
According to the angle you set in the phong tag, the vertex will have 1 or several normals.doing it yourself :
 you have to calculate the normal of all your polygons.
 calculate the average of each vertex normal.
 for each uv's coordinate you want, calculate the barycentric coordinate. That will give you the ratio for each point.
 calculate the bilinear interpolation in 3D.
Cheers,
Manuel 
Just FYI: To calculate the normal of a single
CPolygon
, there is CalcFaceNormal().See also NormalTag Manual.
What version of Cinema 4D are you using?
GeFree()
is pretty outdated. 
i keep forgetting this CalcFaceNormal...

Thanks Manuel and PStudent,
I'm going around in circles. When I try blending between the 3 Phong normals, it comes out in variants like this:
The following is the code I use to calc the Phong normal:
//... hit test function here... // f = 1 / area Vector uvw; uvw.x = f * s.Dot(h); uvw.y = f * dir.Dot(q); uvw.z = 1  uvw.x  uvw.y; //... end of hit test, return true if hit, uvw passed over... // if hit function returns true, use uvw to calc normal at hit spot // norms = CreatePhongNormals() // let's test first polygon in poly 4(5) Vector normal = uvw.x * norms>operator[](4*4) + uvw.y * norms>operator[](4*4+1) + uvw.z * norms>operator[](4*4+2); // repeat for second tri in quad, if exists // etc
I can get the plain polygon face normal fine. But blended ones I can't get my head around. Where am I going wrong here?
WP.

Hi,
at least to me it is not quite clear what you are trying to do. One reason is that there are quite a few symbols floating around in your code which are not selfexplanatory (what are
s, q, h
anddir
?). I also do not quite understand why you are computing something into the vectoruvw
what seems to be a linear combination.If you are interested just in the polygonal normal, you can compute that by evaluating the cross product of the vertices of a corner of your choice in the polygon in a clockwise fashion (e.g.
da % ba
for the vertexa
in a four point polygon). In nonplanar polygons you will have to evaluate two vertices, or more realistically all four, since that is faster than finding "the right ones", and compute the arithmetic mean of them.If you want to to interpolate between the already interpolated point normals (a.k.a. Phong normals), a simple bilinear interpolation between these four normals should be enough. Where the
u
and thev
coordinates of your texture data (or wherever they are coming from) are the parameters for the respective interpolation on that axis. You won't need barycentric coordinates or something like that.Cheers,
zipit 
same here, not sure what you are trying to achieve at the end.

Thanks Zipit / Manuel,
@zipit said in Getting poly normal and CreatePhongNormals():
If you want to interpolate between the already interpolated point normals (a.k.a. Phong normals), a simple bilinear interpolation between these four normals should be enough
Yep  that's what I'm trying to do!! But I have a way with code that makes simple things really difficult . Whatever I try isn't working. I got to this:
but can't figure out the bilinear interpolation. This is where I'm going around in circles. Maybe we take a step back for a moment and do this in two steps.
Step 1: I have a function that tests each polygon triangle using the MollerTrumbore algorithm. Assuming the hit is true, how do I calculate the UV? I'm asking this step to check if I'm setting UV right (which, is where some of the code above comes from).
Step 2: how do I use UV to interpolate between Phong normals a,b and c (triangle)?
The end goal is to draw a normals map.
WP.

Hi,
a bilinear interpolation is quite straight forward. If you have the quadrilateral Q with the the points,
cd   ab
then the bilinear interpolation is just,
ab = lerp(a, b, t0) cd = lerp(c, d, t0) res = lerp(ab, cd, t1)
where
t0, t1
are the interpolation offset(s), i.e. the texture coordinates in your case (the ordering/orientation of the quad is obviously not set in stone). I am not quite sure what you do when rendering normals, but when you render a color gradient, in a value noise for example, you actually want to avoid linear interpolation, because it will give you these ugly starpatterns. So you might need something like a biquadratic, bicubic or bicosine interpolation, i.e. preinterpolate your interpolation offsets.If I am not overlooking something, this should also work for triangles when you treat them as quasiquadrilaterals like Cinema does in its polygon type.
Cheers,
zipit 
Hi,
@WickedP said in Getting poly normal and CreatePhongNormals():
MollerTrumbore
what I forgot to mention, but I already hinted at in my previous posting, is that you did not state in which space your interpolation coordinates are formulated. When you talk about uvcoordinates I am (and probably also everyone else is) assuming that you have cartesian coordinates, just like Cinemas texture coordinate system is formulated in.
If you have another coordinate system, some kind of linear/affine combination like your code snippet suggests, then the interpolated normal is just the linear combination of the neighbouring normals (I think that was what you were trying to do in that code).
If this fails, you should either check your MΓΆllerTrumbore code for errors (ScratchPixel has a nice article on it and how to calculate correct affine space/barycentric coordinates for it), or more pragmatically use Cinema's
GeRayCollider
instead, which will also give you texture coordinates.I am not so well versed in the C++ SDK, maybe there is even a more low level (i.e. triangle level) version of
GeRayCollider
there, i.e. something where you do not have the overhead of casting against a whole mesh.Cheers,
zipit 
Oooh..... I feel like such a dope. I had almost everything right to begin with, except I was accessing the returned CreatePhongNormals() SVector incorrectly. I was using operator()[] when I should have been using ToRV(). Here's a smoothed result:
Thanks guys, we got there
One last thing, how do I 'globalise' the normals? They look local to me (see floating cube, it's slightly rotated but has the same shading as walls in the background)?
WP.

Hi,
all polygonal data is in object space. So if you want your normals to be in global/world space, you will have to multiply them with the frame of the object they are attached to. To get the frame, you will only have to zero out the offset of the global matrix of the object and then normalise the axis components.
Cheers,
zipit 
Thanks zipit,
makes sense being local. Zeroing out and multiplying by the frame matrix does it:
Thanks for your time contributors, been a big help.
WP.

hi
can we considered this thread as resolved ?
Cheers,
Manuel 
Thanks Manuel,
I didn't realise I had to mark the thread as solved. Done.
WP.

@WickedP Uhh, Ohh what are you coding there is that a brdf map shader ? I am searching for something like this !
https://alastaira.wordpress.com/2013/11/26/lightingmodelsandbrdfmaps/

Hi @mogh,
apologies I didn't see your reply.
That display is for a normals engine. I had a 2D draw engine but needed to expand on it because the native renderers wouldn't let me get what I wanted in code. I didn't want to make a render engine, it's just kind of how it's turning out...!
It's only meant to be a basic one that serves the needs though. There's a spiel about it here on my website: [https://wickedp.com/hyperion/](link url). You'll have to excuse the site layout, I'm stuck in the middle of revamping it until I deliver some other big projects.
WP.

Thanks for the reply, Wicked.
interessting, but not what i hoped for ...
kepp on chrunching