If there are any sync issues it's usually more about the tech stuff and not speed of light vs sound.
I actually think that this presentation was a little bit rushed maybe, having a set deadline for this citizencon or whatever; and also perhaps driven by a need to prove that the budget is well spent, as there's been some article on the Escapist recently with some pretty strong claims about management problems at the studio. So they basically decided to show of stuff that was probably not as polished as it could have been.
You see, as far as I'm aware there's no facial mocap solution that can just deliver perfect quality without a lot of manual animation on top. Not even for places like Weta or ILM. It's basically the old 20-80 rule, which means that usually the first 20% of effort will get you to 80% of the quality you seek, but then you have to spend 80% of your effort to get that last 20% of the result.
Coincidentally the same is true for body mocap, but that tech is a lot more mature and more understood by now, so we usually don't get to see the rougher first pass results.
So what I believe to be the case here is that they didn't have enough time to fully polish the animation and get rid of these tiny little issues, which however are pretty damn important to sell the character. Both the tech and the acquired datasets seem to indicate that their system can produce great results - but no matter how much R&D goes into it, there still has to be animators who can provide that final but important touch to bring the characters to life.
But also, as I've mentioned above, this effort takes a lot of human resources, which are quite expensive; and a game like this would probably require hours and hours of facial animation. So no matter how big the budget is, there's always going to be a limited amount of work that's possible to do on the rough p-cap and so we can't reasonably expect every second of the results to be perfect. Here, it comes down to the proficiency of the supervisors to allocate the resources in the most efficient way and make sure that the important and most visible moments get the greatest effort.
So I guess it's a bit too early to tell just how much of the potential of their tech they will actually be able to realize. What has been shown is certainly promising.