exporting bind and keyframe bone poses from blender to use in OpenGL

Posted by SaldaVonSchwartz on Game Development See other posts from Game Development or by SaldaVonSchwartz
Published on 2012-11-20T05:47:11Z Indexed on 2012/11/20 11:28 UTC
Read the original article Hit count: 647

I'm having a hard time trying to understand how exactly Blender's concept of bone transforms maps to the usual math of skinning (which I'm implementing in an OpenGL-based engine of sorts). Or I'm missing out something in the math.. It's gonna be long, but here's as much background as I can think of.

First, a few notes and assumptions:

  • I'm using column-major order and multiply from right to left. So for instance, vertex v transformed by matrix A and then further transformed by matrix B would be: v' = BAv. This also means whenever I export a matrix from blender through python, I export it (in text format) in 4 lines, each representing a column. This is so I can then I can read them back into my engine like this:

    if (fscanf(fileHandle, "%f %f %f %f", &skeleton.joints[currentJointIndex].inverseBindTransform.m[0], &skeleton.joints[currentJointIndex].inverseBindTransform.m[1], &skeleton.joints[currentJointIndex].inverseBindTransform.m[2], &skeleton.joints[currentJointIndex].inverseBindTransform.m[3])) { if (fscanf(fileHandle, "%f %f %f %f", &skeleton.joints[currentJointIndex].inverseBindTransform.m[4], &skeleton.joints[currentJointIndex].inverseBindTransform.m[5], &skeleton.joints[currentJointIndex].inverseBindTransform.m[6], &skeleton.joints[currentJointIndex].inverseBindTransform.m[7])) { if (fscanf(fileHandle, "%f %f %f %f", &skeleton.joints[currentJointIndex].inverseBindTransform.m[8], &skeleton.joints[currentJointIndex].inverseBindTransform.m[9], &skeleton.joints[currentJointIndex].inverseBindTransform.m[10], &skeleton.joints[currentJointIndex].inverseBindTransform.m[11])) { if (fscanf(fileHandle, "%f %f %f %f", &skeleton.joints[currentJointIndex].inverseBindTransform.m[12], &skeleton.joints[currentJointIndex].inverseBindTransform.m[13], &skeleton.joints[currentJointIndex].inverseBindTransform.m[14], &skeleton.joints[currentJointIndex].inverseBindTransform.m[15])) {

  • I'm simplifying the code I show because otherwise it would make things unnecessarily harder (in the context of my question) to explain / follow.

  • Please refrain from making remarks related to optimizations. This is not final code.

Having said that, if I understand correctly, the basic idea of skinning/animation is:

  • I have a a mesh made up of vertices

  • I have the mesh model->world transform W

  • I have my joints, which are really just transforms from each joint's space to its parent's space. I'll call these transforms Bj meaning matrix which takes from joint j's bind pose to joint j-1's bind pose. For each of these, I actually import their inverse to the engine, Bj^-1.

  • I have keyframes each containing a set of current poses Cj for each joint J. These are initially imported to my engine in TQS format but after (S)LERPING them I compose them into Cj matrices which are equivalent to the Bjs (not the Bj^-1 ones) only that for the current spacial configurations of each joint at that frame.

Given the above, the "skeletal animation algorithm is"

On each frame:
   check how much time has elpased and compute the resulting current time 
   in the animation, from 0 meaning frame 0 to 1, meaning the end of the animation. 
   (Oh and I'm looping forever so the time is mod(total duration))

    for each joint:
       1 -calculate its world inverse bind pose, that is Bj_w^-1 = Bj^-1 Bj-1^-1 ... B0^-1
       2 -use the current animation time to LERP the componets of the TQS and come up
          with an interpolated current pose matrix Cj which should transform from the
          joints current configuration space to world space. Similar to what I did to
          get the world version of the inverse bind poses, I come up with the joint's 
          world current pose, Cj_w = C0 C1 ... Cj
       3 -now that I have world versions of Bj and Cj, I store this joint's world-
          skinning matrix K_wj = Cj_w Bj_w^-1.

The above is roughly implemented like so:

- (void)update:(NSTimeInterval)elapsedTime
{
            static double time = 0;
            time =  fmod((time + elapsedTime),1.);
            uint16_t LERPKeyframeNumber = 60 * time;
            uint16_t lkeyframeNumber = 0;
            uint16_t lkeyframeIndex = 0;
            uint16_t rkeyframeNumber = 0;
            uint16_t rkeyframeIndex = 0;


            for (int i = 0; i < aClip.keyframesCount; i++) {
                uint16_t keyframeNumber = aClip.keyframes[i].number;
                if (keyframeNumber <= LERPKeyframeNumber) {
                    lkeyframeIndex = i;
                    lkeyframeNumber = keyframeNumber;
                }
                else {
                    rkeyframeIndex = i;
                    rkeyframeNumber = keyframeNumber;
                    break;
                }
            }

            double lTime = lkeyframeNumber / 60.;
            double rTime = rkeyframeNumber / 60.;
            double blendFactor = (time - lTime) / (rTime - lTime);


            GLKMatrix4 bindPosePalette[aSkeleton.jointsCount];
            GLKMatrix4 currentPosePalette[aSkeleton.jointsCount];

            for (int i = 0; i < aSkeleton.jointsCount; i++) {
                F3DETQSType& lPose = aClip.keyframes[lkeyframeIndex].skeletonPose.jointPoses[i];
                F3DETQSType& rPose = aClip.keyframes[rkeyframeIndex].skeletonPose.jointPoses[i];
                GLKVector3 LERPTranslation = GLKVector3Lerp(lPose.t, rPose.t, blendFactor);
                GLKQuaternion SLERPRotation = GLKQuaternionSlerp(lPose.q, rPose.q, blendFactor);
                GLKVector3 LERPScaling = GLKVector3Lerp(lPose.s, rPose.s, blendFactor);
                GLKMatrix4  currentTransform = GLKMatrix4MakeWithQuaternion(SLERPRotation);
                currentTransform =  GLKMatrix4Multiply(currentTransform,
                                                       GLKMatrix4MakeTranslation(LERPTranslation.x, LERPTranslation.y, LERPTranslation.z));
                currentTransform =  GLKMatrix4Multiply(currentTransform,
                                                       GLKMatrix4MakeScale(LERPScaling.x, LERPScaling.y, LERPScaling.z));

                if (aSkeleton.joints[i].parentIndex == -1) {
                    bindPosePalette[i] = aSkeleton.joints[i].inverseBindTransform;
                    currentPosePalette[i] = currentTransform;
                }

                else {
                    bindPosePalette[i] = GLKMatrix4Multiply(aSkeleton.joints[i].inverseBindTransform, bindPosePalette[aSkeleton.joints[i].parentIndex]);
                    currentPosePalette[i] = GLKMatrix4Multiply(currentPosePalette[aSkeleton.joints[i].parentIndex], currentTransform);
                }

                aSkeleton.skinningPalette[i] = GLKMatrix4Multiply(currentPosePalette[i], bindPosePalette[i]);
            }

     }

At this point, I should have my skinning palette. So on each frame in my vertex shader, I do:

uniform mat4 modelMatrix;
uniform mat4 projectionMatrix;
uniform mat3 normalMatrix;
uniform mat4 skinningPalette[6];

attribute vec4 position;
attribute vec3 normal;
attribute vec2 tCoordinates;
attribute vec4 jointsWeights;
attribute vec4 jointsIndices;

varying highp vec2 tCoordinatesVarying;
varying highp float lIntensity;

void main()
{
    vec3 eyeNormal = normalize(normalMatrix * normal);
    vec3 lightPosition = vec3(0., 0., 2.);
    lIntensity = max(0.0, dot(eyeNormal, normalize(lightPosition)));
    tCoordinatesVarying = tCoordinates;

    vec4 skinnedVertexPosition = vec4(0.);
    for (int i = 0; i < 4; i++) {
        skinnedVertexPosition +=  jointsWeights[i] * skinningPalette[int(jointsIndices[i])] * position;
    }

    gl_Position = projectionMatrix * modelMatrix * skinnedVertexPosition;
}

The result:

The mesh parts that are supposed to animate do animate and follow the expected motion, however, the rotations are messed up in terms of orientations. That is, the mesh is not translated somewhere else or scaled in any way, but the orientations of rotations seem to be off.

So a few observations:

  1. In the above shader notice I actually did not multiply the vertices by the mesh modelMatrix (the one which would take them to model or world or global space, whichever you prefer, since there is no parent to the mesh itself other than "the world") until after skinning. This is contrary to what I implied in the theory: if my skinning matrix takes vertices from model to joint and back to model space, I'd think the vertices should already be premultiplied by the mesh transform. But if I do so, I just get a black screen.
  2. As far as exporting the joints from Blender, my python script exports for each armature bone in bind pose, it's matrix in this way:

    def DFSJointTraversal(file, skeleton, jointList):

    for joint in jointList:
        poseJoint = skeleton.pose.bones[joint.name]
    
        jointTransform = poseJoint.matrix.inverted()
    
        file.write('Joint ' + joint.name + ' Transform {\n')
        for col in jointTransform.col:
            file.write('{:9f} {:9f} {:9f} {:9f}\n'.format(col[0], col[1], col[2], col[3]))
    
        DFSJointTraversal(file, skeleton, joint.children)
        file.write('}\n')
    

And for current / keyframe poses (assuming I'm in the right keyframe):

def exportAnimations(filepath):

    # Only one skeleton per scene
    objList = [object for object in bpy.context.scene.objects if object.type == 'ARMATURE']

    if len(objList) == 0:
        return
    elif len(objList) > 1:
        return
    #raise exception? dialog box?

    skeleton = objList[0]
    jointNames = [bone.name for bone in skeleton.data.bones]

    for action in bpy.data.actions:
        # One animation clip per action in Blender, named as the action
        animationClipFilePath = filepath[0 : filepath.rindex('/') + 1] + action.name + ".aClip"
        file = open(animationClipFilePath, 'w')
        file.write('target skeleton: ' + skeleton.name + '\n')
        file.write('joints count: {:d}'.format(len(jointNames)) + '\n')

        skeleton.animation_data.action = action
        keyframeNum = max([len(fcurve.keyframe_points) for fcurve in action.fcurves])

        keyframes = []
        for fcurve in action.fcurves:
            for keyframe in fcurve.keyframe_points:
                keyframes.append(keyframe.co[0])

        keyframes = set(keyframes)
        keyframes = [kf for kf in keyframes]
        keyframes.sort()
        file.write('keyframes count: {:d}'.format(len(keyframes)) + '\n')

        for kfIndex in keyframes:
            bpy.context.scene.frame_set(kfIndex)
            file.write('keyframe: {:d}\n'.format(int(kfIndex)))
            for i in range(0, len(skeleton.data.bones)):
                file.write('joint: {:d}\n'.format(i))

                joint = skeleton.pose.bones[i]

                jointCurrentPoseTransform = joint.matrix

                translationV = jointCurrentPoseTransform.to_translation()
                rotationQ = jointCurrentPoseTransform.to_3x3().to_quaternion()
                scaleV = jointCurrentPoseTransform.to_scale()
                file.write('T {:9f} {:9f} {:9f}\n'.format(translationV[0], translationV[1], translationV[2]))
                file.write('Q {:9f} {:9f} {:9f} {:9f}\n'.format(rotationQ[1], rotationQ[2], rotationQ[3], rotationQ[0]))
                file.write('S {:9f} {:9f} {:9f}\n'.format(scaleV[0], scaleV[1], scaleV[2]))

                file.write('\n')

        file.close()

Which I believe follow the theory explained at the beginning of my question. But then I checked out Blender's directX .x exporter for reference.. and what threw me off was that in the .x script they are exporting bind poses like so (transcribed using the same variable names I used so you can compare):

if joint.parent:
            jointTransform = poseJoint.parent.matrix.inverted()
        else:
            jointTransform = Matrix()

        jointTransform *= poseJoint.matrix

and exporting current keyframe poses like this:

if joint.parent:
                    jointCurrentPoseTransform = joint.parent.matrix.inverted()
                else:
                    jointCurrentPoseTransform = Matrix()
                jointCurrentPoseTransform *= joint.matrix
  • why are they using the parent's transform instead of the joint in question's? isn't the join transform assumed to exist in the context of a parent transform since after all it transforms from this joint's space to its parent's?
  • Why are they concatenating in the same order for both bind poses and keyframe poses? If these two are then supposed to be concatenated with each other to cancel out the change of basis?

Anyway, any ideas are appreciated.

© Game Development or respective owner

Related posts about opengl

Related posts about opengl-es