t1 / TFDContents / Assets / KinectScripts / FacetrackingManager.cs @ 3
이력 | 보기 | 이력해설 | 다운로드 (42.6 KB)
| 1 |
using UnityEngine; |
|---|---|
| 2 |
using System; |
| 3 |
using System.Collections; |
| 4 |
using System.Collections.Generic; |
| 5 |
//using System.Runtime.InteropServices; |
| 6 |
using System.Text; |
| 7 |
|
| 8 |
|
| 9 |
/// <summary> |
| 10 |
/// Facetracking manager is the component that manages the head and face tracking. |
| 11 |
/// </summary> |
| 12 |
public class FacetrackingManager : MonoBehaviour |
| 13 |
{
|
| 14 |
[Tooltip("Index of the player, tracked by this component. 0 means the 1st player, 1 - the 2nd one, 2 - the 3rd one, etc.")]
|
| 15 |
public int playerIndex = 0; |
| 16 |
|
| 17 |
[Tooltip("Whether to poll the HD-face model data or not.")]
|
| 18 |
public bool getFaceModelData = false; |
| 19 |
|
| 20 |
[Tooltip("Whether to display the face rectangle over the color camera feed.")]
|
| 21 |
public bool displayFaceRect = false; |
| 22 |
|
| 23 |
[Tooltip("Time tolerance in seconds, when the face may not to be tracked, without considering it lost.")]
|
| 24 |
public float faceTrackingTolerance = 0.5f; |
| 25 |
|
| 26 |
[Tooltip("Game object that will be used to display the HD-face model mesh in the scene.")]
|
| 27 |
public GameObject faceModelMesh = null; |
| 28 |
|
| 29 |
[Tooltip("Whether the HD-face model mesh should be mirrored or not.")]
|
| 30 |
private bool mirroredModelMesh = true; |
| 31 |
|
| 32 |
//[Tooltip("Whether to skip the continuous updates of the HD-face model mesh, or not.")]
|
| 33 |
//public bool dontUpdateModelMesh = false; |
| 34 |
|
| 35 |
[Tooltip("Whether to pause the updates of the HD-face model mesh.")]
|
| 36 |
public bool pauseModelMeshUpdates = false; |
| 37 |
|
| 38 |
public enum TextureType : int { None, ColorMap, FaceRectangle }
|
| 39 |
[Tooltip("How the HD-face model mesh should be textured.")]
|
| 40 |
public TextureType texturedModelMesh = TextureType.ColorMap; |
| 41 |
|
| 42 |
[Tooltip("Whether to move the face model mesh, to be the same as user's head position.")]
|
| 43 |
public bool moveModelMesh = false; |
| 44 |
|
| 45 |
[Tooltip("Camera that may be used to overlay face mesh over the color background.")]
|
| 46 |
public Camera foregroundCamera; |
| 47 |
|
| 48 |
[Tooltip("Scale factor for the face mesh.")]
|
| 49 |
[Range(0.1f, 2.0f)] |
| 50 |
public float modelMeshScale = 1f; |
| 51 |
|
| 52 |
[Tooltip("Vertical offset of the mesh above the head (in meters).")]
|
| 53 |
[Range(-0.5f, 0.5f)] |
| 54 |
public float verticalMeshOffset = 0f; |
| 55 |
|
| 56 |
[Tooltip("GUI-Text to display the FT-manager debug messages.")]
|
| 57 |
public GUIText debugText; |
| 58 |
|
| 59 |
// // nose and head transforms |
| 60 |
// public Transform noseTransform; |
| 61 |
// public Transform headTransform; |
| 62 |
// public GUIText debugText2; |
| 63 |
|
| 64 |
|
| 65 |
// Is currently tracking user's face |
| 66 |
private bool isTrackingFace = false; |
| 67 |
private float lastFaceTrackedTime = 0f; |
| 68 |
|
| 69 |
// Skeleton ID of the tracked face |
| 70 |
//private long faceTrackingID = 0; |
| 71 |
|
| 72 |
// Animation units |
| 73 |
private Dictionary<KinectInterop.FaceShapeAnimations, float> dictAU = new Dictionary<KinectInterop.FaceShapeAnimations, float>(); |
| 74 |
private bool bGotAU = false; |
| 75 |
|
| 76 |
// Shape units |
| 77 |
private Dictionary<KinectInterop.FaceShapeDeformations, float> dictSU = new Dictionary<KinectInterop.FaceShapeDeformations, float>(); |
| 78 |
private bool bGotSU = false; |
| 79 |
|
| 80 |
// whether the face model mesh was initialized |
| 81 |
private bool bFaceModelMeshInited = false; |
| 82 |
private Vector3[] vMeshVertices = null; |
| 83 |
|
| 84 |
// Vertices, UV and triangles of the face model |
| 85 |
private Vector3[] avModelVertices = null; |
| 86 |
private Vector2[] avModelUV = null; |
| 87 |
private bool bGotModelVertices = false; |
| 88 |
//private bool bGotModelVerticesFromDC = false; |
| 89 |
|
| 90 |
private int[] avModelTriangles = null; |
| 91 |
private bool bGotModelTriangles = false; |
| 92 |
private bool bGotModelTrianglesFromDC = false; |
| 93 |
|
| 94 |
// Head position and rotation |
| 95 |
private Vector3 headPos = Vector3.zero; |
| 96 |
private bool bGotHeadPos = false; |
| 97 |
|
| 98 |
private Quaternion headRot = Quaternion.identity; |
| 99 |
private bool bGotHeadRot = false; |
| 100 |
|
| 101 |
// offset vector from head to face center |
| 102 |
private Vector3 faceHeadOffset = Vector3.zero; |
| 103 |
|
| 104 |
// Tracked face rectangle |
| 105 |
private Rect faceRect = new Rect(); |
| 106 |
//private bool bGotFaceRect; |
| 107 |
|
| 108 |
// primary user ID, as reported by KinectManager |
| 109 |
private long primaryUserID = 0; |
| 110 |
private long lastUserID = 0; |
| 111 |
|
| 112 |
// primary sensor data structure |
| 113 |
private KinectInterop.SensorData sensorData = null; |
| 114 |
|
| 115 |
// Bool to keep track of whether face-tracking system has been initialized |
| 116 |
private bool isFacetrackingInitialized = false; |
| 117 |
private bool wasFacetrackingActive = false; |
| 118 |
|
| 119 |
// The single instance of FacetrackingManager |
| 120 |
private static FacetrackingManager instance; |
| 121 |
|
| 122 |
// update times |
| 123 |
private float facePosUpdateTime = 0f; |
| 124 |
private float faceMeshUpdateTime = 0f; |
| 125 |
|
| 126 |
// used when dontUpdateModelMesh is true |
| 127 |
//private bool faceMeshGotOnce = false; |
| 128 |
|
| 129 |
// whether UpdateFaceModelMesh() is running |
| 130 |
private bool updateFaceMeshStarted = false; |
| 131 |
|
| 132 |
private RenderTexture faceMeshTexture = null; |
| 133 |
private Vector3 nosePos = Vector3.zero; |
| 134 |
|
| 135 |
/// <summary> |
| 136 |
/// Gets the single FacetrackingManager instance. |
| 137 |
/// </summary> |
| 138 |
/// <value>The FacetrackingManager instance.</value> |
| 139 |
public static FacetrackingManager Instance |
| 140 |
{
|
| 141 |
get |
| 142 |
{
|
| 143 |
return instance; |
| 144 |
} |
| 145 |
} |
| 146 |
|
| 147 |
/// <summary> |
| 148 |
/// Determines the facetracking system was successfully initialized, false otherwise. |
| 149 |
/// </summary> |
| 150 |
/// <returns><c>true</c> if the facetracking system was successfully initialized; otherwise, <c>false</c>.</returns> |
| 151 |
public bool IsFaceTrackingInitialized() |
| 152 |
{
|
| 153 |
return isFacetrackingInitialized; |
| 154 |
} |
| 155 |
|
| 156 |
/// <summary> |
| 157 |
/// Determines whether this the sensor is currently tracking a face. |
| 158 |
/// </summary> |
| 159 |
/// <returns><c>true</c> if the sensor is tracking a face; otherwise, <c>false</c>.</returns> |
| 160 |
public bool IsTrackingFace() |
| 161 |
{
|
| 162 |
return isTrackingFace; |
| 163 |
} |
| 164 |
|
| 165 |
/// <summary> |
| 166 |
/// Gets the current user ID, or 0 if no user is currently tracked. |
| 167 |
/// </summary> |
| 168 |
/// <returns>The face tracking I.</returns> |
| 169 |
public long GetFaceTrackingID() |
| 170 |
{
|
| 171 |
return isTrackingFace ? primaryUserID : 0; |
| 172 |
} |
| 173 |
|
| 174 |
/// <summary> |
| 175 |
/// Determines whether the sensor is currently tracking the face of the specified user. |
| 176 |
/// </summary> |
| 177 |
/// <returns><c>true</c> if the sensor is currently tracking the face of the specified user; otherwise, <c>false</c>.</returns> |
| 178 |
/// <param name="userId">User ID</param> |
| 179 |
public bool IsTrackingFace(long userId) |
| 180 |
{
|
| 181 |
if(sensorData != null && sensorData.sensorInterface != null) |
| 182 |
{
|
| 183 |
return sensorData.sensorInterface.IsFaceTracked(userId); |
| 184 |
} |
| 185 |
|
| 186 |
return false; |
| 187 |
} |
| 188 |
|
| 189 |
/// <summary> |
| 190 |
/// Gets the last face position & rotation update time, in seconds since game start. |
| 191 |
/// </summary> |
| 192 |
/// <returns>The last face position & rotation update time.</returns> |
| 193 |
public float GetFacePosUpdateTime() |
| 194 |
{
|
| 195 |
return facePosUpdateTime; |
| 196 |
} |
| 197 |
|
| 198 |
/// <summary> |
| 199 |
/// Gets the last face mesh update time, in seconds since game start. |
| 200 |
/// </summary> |
| 201 |
/// <returns>The last face mesh update time.</returns> |
| 202 |
public float GetFaceMeshUpdateTime() |
| 203 |
{
|
| 204 |
return faceMeshUpdateTime; |
| 205 |
} |
| 206 |
|
| 207 |
/// <summary> |
| 208 |
/// Gets the head position of the currently tracked user. |
| 209 |
/// </summary> |
| 210 |
/// <returns>The head position.</returns> |
| 211 |
/// <param name="bMirroredMovement">If set to <c>true</c> returns mirorred head position.</param> |
| 212 |
public Vector3 GetHeadPosition(bool bMirroredMovement) |
| 213 |
{
|
| 214 |
Vector3 vHeadPos = headPos; // bGotHeadPos ? headPos : Vector3.zero; |
| 215 |
|
| 216 |
if(!bMirroredMovement) |
| 217 |
{
|
| 218 |
vHeadPos.z = -vHeadPos.z; |
| 219 |
} |
| 220 |
|
| 221 |
return vHeadPos; |
| 222 |
} |
| 223 |
|
| 224 |
/// <summary> |
| 225 |
/// Gets the head position of the specified user. |
| 226 |
/// </summary> |
| 227 |
/// <returns>The head position.</returns> |
| 228 |
/// <param name="userId">User ID</param> |
| 229 |
/// <param name="bMirroredMovement">If set to <c>true</c> returns mirorred head position.</param> |
| 230 |
public Vector3 GetHeadPosition(long userId, bool bMirroredMovement) |
| 231 |
{
|
| 232 |
Vector3 vHeadPos = Vector3.zero; |
| 233 |
bool bGotPosition = sensorData.sensorInterface.GetHeadPosition(userId, ref vHeadPos); |
| 234 |
|
| 235 |
if(bGotPosition) |
| 236 |
{
|
| 237 |
if(!bMirroredMovement) |
| 238 |
{
|
| 239 |
vHeadPos.z = -vHeadPos.z; |
| 240 |
} |
| 241 |
|
| 242 |
return vHeadPos; |
| 243 |
} |
| 244 |
|
| 245 |
return Vector3.zero; |
| 246 |
} |
| 247 |
|
| 248 |
/// <summary> |
| 249 |
/// Gets the head rotation of the currently tracked user. |
| 250 |
/// </summary> |
| 251 |
/// <returns>The head rotation.</returns> |
| 252 |
/// <param name="bMirroredMovement">If set to <c>true</c> returns mirorred head rotation.</param> |
| 253 |
public Quaternion GetHeadRotation(bool bMirroredMovement) |
| 254 |
{
|
| 255 |
Vector3 rotAngles = headRot.eulerAngles; // bGotHeadRot ? headRot.eulerAngles : Vector3.zero; |
| 256 |
|
| 257 |
if(bMirroredMovement) |
| 258 |
{
|
| 259 |
rotAngles.x = -rotAngles.x; |
| 260 |
rotAngles.z = -rotAngles.z; |
| 261 |
} |
| 262 |
else |
| 263 |
{
|
| 264 |
rotAngles.x = -rotAngles.x; |
| 265 |
rotAngles.y = -rotAngles.y; |
| 266 |
} |
| 267 |
|
| 268 |
return Quaternion.Euler(rotAngles); |
| 269 |
} |
| 270 |
|
| 271 |
/// <summary> |
| 272 |
/// Gets the head rotation of the specified user. |
| 273 |
/// </summary> |
| 274 |
/// <returns>The head rotation.</returns> |
| 275 |
/// <param name="userId">User ID</param> |
| 276 |
/// <param name="bMirroredMovement">If set to <c>true</c> returns mirorred head rotation.</param> |
| 277 |
public Quaternion GetHeadRotation(long userId, bool bMirroredMovement) |
| 278 |
{
|
| 279 |
Quaternion vHeadRot = Quaternion.identity; |
| 280 |
bool bGotRotation = sensorData.sensorInterface.GetHeadRotation(userId, ref vHeadRot); |
| 281 |
|
| 282 |
if(bGotRotation) |
| 283 |
{
|
| 284 |
Vector3 rotAngles = vHeadRot.eulerAngles; |
| 285 |
|
| 286 |
if(bMirroredMovement) |
| 287 |
{
|
| 288 |
rotAngles.x = -rotAngles.x; |
| 289 |
rotAngles.z = -rotAngles.z; |
| 290 |
} |
| 291 |
else |
| 292 |
{
|
| 293 |
rotAngles.x = -rotAngles.x; |
| 294 |
rotAngles.y = -rotAngles.y; |
| 295 |
} |
| 296 |
|
| 297 |
return Quaternion.Euler(rotAngles); |
| 298 |
} |
| 299 |
|
| 300 |
return Quaternion.identity; |
| 301 |
} |
| 302 |
|
| 303 |
/// <summary> |
| 304 |
/// Gets the tracked face rectangle of the specified user in color image coordinates, or zero-rect if the user's face is not tracked. |
| 305 |
/// </summary> |
| 306 |
/// <returns>The face rectangle, in color image coordinates.</returns> |
| 307 |
/// <param name="userId">User ID</param> |
| 308 |
public Rect GetFaceColorRect(long userId) |
| 309 |
{
|
| 310 |
Rect faceColorRect = new Rect(); |
| 311 |
sensorData.sensorInterface.GetFaceRect(userId, ref faceColorRect); |
| 312 |
|
| 313 |
return faceColorRect; |
| 314 |
} |
| 315 |
|
| 316 |
/// <summary> |
| 317 |
/// Determines whether there are valid anim units. |
| 318 |
/// </summary> |
| 319 |
/// <returns><c>true</c> if there are valid anim units; otherwise, <c>false</c>.</returns> |
| 320 |
public bool IsGotAU() |
| 321 |
{
|
| 322 |
return bGotAU; |
| 323 |
} |
| 324 |
|
| 325 |
/// <summary> |
| 326 |
/// Gets the animation unit value at given index, or 0 if the index is invalid. |
| 327 |
/// </summary> |
| 328 |
/// <returns>The animation unit value.</returns> |
| 329 |
/// <param name="faceAnimKey">Face animation unit.</param> |
| 330 |
public float GetAnimUnit(KinectInterop.FaceShapeAnimations faceAnimKey) |
| 331 |
{
|
| 332 |
if(dictAU.ContainsKey(faceAnimKey)) |
| 333 |
{
|
| 334 |
return dictAU[faceAnimKey]; |
| 335 |
} |
| 336 |
|
| 337 |
return 0.0f; |
| 338 |
} |
| 339 |
|
| 340 |
/// <summary> |
| 341 |
/// Gets all animation units for the specified user. |
| 342 |
/// </summary> |
| 343 |
/// <returns><c>true</c>, if the user's face is tracked, <c>false</c> otherwise.</returns> |
| 344 |
/// <param name="userId">User ID</param> |
| 345 |
/// <param name="dictAnimUnits">Animation units dictionary, to get the results.</param> |
| 346 |
public bool GetUserAnimUnits(long userId, ref Dictionary<KinectInterop.FaceShapeAnimations, float> dictAnimUnits) |
| 347 |
{
|
| 348 |
if(sensorData != null && sensorData.sensorInterface != null) |
| 349 |
{
|
| 350 |
bool bGotIt = sensorData.sensorInterface.GetAnimUnits(userId, ref dictAnimUnits); |
| 351 |
return bGotIt; |
| 352 |
} |
| 353 |
|
| 354 |
return false; |
| 355 |
} |
| 356 |
|
| 357 |
/// <summary> |
| 358 |
/// Determines whether there are valid shape units. |
| 359 |
/// </summary> |
| 360 |
/// <returns><c>true</c> if there are valid shape units; otherwise, <c>false</c>.</returns> |
| 361 |
public bool IsGotSU() |
| 362 |
{
|
| 363 |
return bGotSU; |
| 364 |
} |
| 365 |
|
| 366 |
/// <summary> |
| 367 |
/// Gets the shape unit value at given index, or 0 if the index is invalid. |
| 368 |
/// </summary> |
| 369 |
/// <returns>The shape unit value.</returns> |
| 370 |
/// <param name="faceShapeKey">Face shape unit.</param> |
| 371 |
public float GetShapeUnit(KinectInterop.FaceShapeDeformations faceShapeKey) |
| 372 |
{
|
| 373 |
if(dictSU.ContainsKey(faceShapeKey)) |
| 374 |
{
|
| 375 |
return dictSU[faceShapeKey]; |
| 376 |
} |
| 377 |
|
| 378 |
return 0.0f; |
| 379 |
} |
| 380 |
|
| 381 |
/// <summary> |
| 382 |
/// Gets all animation units for the specified user. |
| 383 |
/// </summary> |
| 384 |
/// <returns><c>true</c>, if the user's face is tracked, <c>false</c> otherwise.</returns> |
| 385 |
/// <param name="userId">User ID</param> |
| 386 |
/// <param name="dictShapeUnits">Shape units dictionary, to get the results.</param> |
| 387 |
public bool GetUserShapeUnits(long userId, ref Dictionary<KinectInterop.FaceShapeDeformations, float> dictShapeUnits) |
| 388 |
{
|
| 389 |
if(sensorData != null && sensorData.sensorInterface != null) |
| 390 |
{
|
| 391 |
bool bGotIt = sensorData.sensorInterface.GetShapeUnits(userId, ref dictShapeUnits); |
| 392 |
return bGotIt; |
| 393 |
} |
| 394 |
|
| 395 |
return false; |
| 396 |
} |
| 397 |
|
| 398 |
/// <summary> |
| 399 |
/// Gets the count of face model vertices. |
| 400 |
/// </summary> |
| 401 |
/// <returns>The count of face model vertices.</returns> |
| 402 |
public int GetFaceModelVertexCount() |
| 403 |
{
|
| 404 |
if (avModelVertices != null) |
| 405 |
{
|
| 406 |
return avModelVertices.Length; |
| 407 |
} |
| 408 |
|
| 409 |
return 0; |
| 410 |
} |
| 411 |
|
| 412 |
/// <summary> |
| 413 |
/// Gets the face model vertex, if a face model is available and the index is in range; Vector3.zero otherwise. |
| 414 |
/// </summary> |
| 415 |
/// <returns>The face model vertex.</returns> |
| 416 |
/// <param name="index">Vertex index, or Vector3.zero</param> |
| 417 |
public Vector3 GetFaceModelVertex(int index) |
| 418 |
{
|
| 419 |
if (avModelVertices != null) |
| 420 |
{
|
| 421 |
if(index >= 0 && index < avModelVertices.Length) |
| 422 |
{
|
| 423 |
return avModelVertices[index]; |
| 424 |
} |
| 425 |
} |
| 426 |
|
| 427 |
return Vector3.zero; |
| 428 |
} |
| 429 |
|
| 430 |
/// <summary> |
| 431 |
/// Gets all face model vertices, if a face model is available; null otherwise. |
| 432 |
/// </summary> |
| 433 |
/// <returns>The face model vertices, or null.</returns> |
| 434 |
public Vector3[] GetFaceModelVertices() |
| 435 |
{
|
| 436 |
return avModelVertices; |
| 437 |
} |
| 438 |
|
| 439 |
/// <summary> |
| 440 |
/// Gets the count of face model vertices for the specified user |
| 441 |
/// </summary> |
| 442 |
/// <returns>The count of face model vertices.</returns> |
| 443 |
/// <param name="userId">User ID</param> |
| 444 |
public int GetUserFaceVertexCount(long userId) |
| 445 |
{
|
| 446 |
if(sensorData != null && sensorData.sensorInterface != null) |
| 447 |
{
|
| 448 |
int iVertCount = sensorData.sensorInterface.GetFaceModelVerticesCount(userId); |
| 449 |
return iVertCount; |
| 450 |
} |
| 451 |
|
| 452 |
return 0; |
| 453 |
} |
| 454 |
|
| 455 |
/// <summary> |
| 456 |
/// Gets all face model vertices for the specified user. |
| 457 |
/// </summary> |
| 458 |
/// <returns><c>true</c>, if the user's face is tracked, <c>false</c> otherwise.</returns> |
| 459 |
/// <param name="userId">User ID</param> |
| 460 |
/// <param name="avVertices">Reference to array of vertices, to get the result.</param> |
| 461 |
public bool GetUserFaceVertices(long userId, ref Vector3[] avVertices) |
| 462 |
{
|
| 463 |
if(sensorData != null && sensorData.sensorInterface != null) |
| 464 |
{
|
| 465 |
bool bGotIt = sensorData.sensorInterface.GetFaceModelVertices(userId, ref avVertices); |
| 466 |
return bGotIt; |
| 467 |
} |
| 468 |
|
| 469 |
return false; |
| 470 |
} |
| 471 |
|
| 472 |
/// <summary> |
| 473 |
/// Gets the count of face model triangles. |
| 474 |
/// </summary> |
| 475 |
/// <returns>The count of face model triangles.</returns> |
| 476 |
public int GetFaceModelTriangleCount() |
| 477 |
{
|
| 478 |
if (avModelTriangles != null) |
| 479 |
{
|
| 480 |
return avModelTriangles.Length; |
| 481 |
} |
| 482 |
|
| 483 |
return 0; |
| 484 |
} |
| 485 |
|
| 486 |
/// <summary> |
| 487 |
/// Gets the face model triangle indices, if a face model is available; null otherwise. |
| 488 |
/// </summary> |
| 489 |
/// <returns>The face model triangle indices, or null.</returns> |
| 490 |
/// <param name="bMirroredModel">If set to <c>true</c> gets mirorred model indices.</param> |
| 491 |
public int[] GetFaceModelTriangleIndices(bool bMirroredModel) |
| 492 |
{
|
| 493 |
if (avModelTriangles != null) |
| 494 |
{
|
| 495 |
return avModelTriangles; |
| 496 |
} |
| 497 |
|
| 498 |
return null; |
| 499 |
} |
| 500 |
|
| 501 |
|
| 502 |
//----------------------------------- end of public functions --------------------------------------// |
| 503 |
|
| 504 |
void Awake() |
| 505 |
{
|
| 506 |
instance = this; |
| 507 |
} |
| 508 |
|
| 509 |
void Start() |
| 510 |
{
|
| 511 |
try |
| 512 |
{
|
| 513 |
// get sensor data |
| 514 |
KinectManager kinectManager = KinectManager.Instance; |
| 515 |
if(kinectManager && kinectManager.IsInitialized()) |
| 516 |
{
|
| 517 |
sensorData = kinectManager.GetSensorData(); |
| 518 |
} |
| 519 |
|
| 520 |
if(sensorData == null || sensorData.sensorInterface == null) |
| 521 |
{
|
| 522 |
throw new Exception("Face tracking cannot be started, because KinectManager is missing or not initialized.");
|
| 523 |
} |
| 524 |
|
| 525 |
if(debugText != null) |
| 526 |
{
|
| 527 |
debugText.text = "Please, wait..."; |
| 528 |
} |
| 529 |
|
| 530 |
// ensure the needed dlls are in place and face tracking is available for this interface |
| 531 |
bool bNeedRestart = false; |
| 532 |
if(sensorData.sensorInterface.IsFaceTrackingAvailable(ref bNeedRestart)) |
| 533 |
{
|
| 534 |
if(bNeedRestart) |
| 535 |
{
|
| 536 |
KinectInterop.RestartLevel(gameObject, "FM"); |
| 537 |
return; |
| 538 |
} |
| 539 |
} |
| 540 |
else |
| 541 |
{
|
| 542 |
string sInterfaceName = sensorData.sensorInterface.GetType().Name; |
| 543 |
throw new Exception(sInterfaceName + ": Face tracking is not supported!"); |
| 544 |
} |
| 545 |
|
| 546 |
// Initialize the face tracker |
| 547 |
wasFacetrackingActive = sensorData.sensorInterface.IsFaceTrackingActive(); |
| 548 |
if(!wasFacetrackingActive) |
| 549 |
{
|
| 550 |
if (!sensorData.sensorInterface.InitFaceTracking(getFaceModelData, displayFaceRect)) |
| 551 |
{
|
| 552 |
throw new Exception("Face tracking could not be initialized.");
|
| 553 |
} |
| 554 |
} |
| 555 |
|
| 556 |
isFacetrackingInitialized = true; |
| 557 |
|
| 558 |
//DontDestroyOnLoad(gameObject); |
| 559 |
|
| 560 |
if(debugText != null) |
| 561 |
{
|
| 562 |
debugText.text = "Ready."; |
| 563 |
} |
| 564 |
} |
| 565 |
catch(DllNotFoundException ex) |
| 566 |
{
|
| 567 |
Debug.LogError(ex.ToString()); |
| 568 |
if(debugText != null) |
| 569 |
debugText.text = "Please check the Kinect and FT-Library installations."; |
| 570 |
} |
| 571 |
catch (Exception ex) |
| 572 |
{
|
| 573 |
Debug.LogError(ex.ToString()); |
| 574 |
if(debugText != null) |
| 575 |
debugText.text = ex.Message; |
| 576 |
} |
| 577 |
} |
| 578 |
|
| 579 |
void OnDestroy() |
| 580 |
{
|
| 581 |
if(isFacetrackingInitialized && !wasFacetrackingActive && sensorData != null && sensorData.sensorInterface != null) |
| 582 |
{
|
| 583 |
// finish face tracking |
| 584 |
sensorData.sensorInterface.FinishFaceTracking(); |
| 585 |
} |
| 586 |
|
| 587 |
if (faceMeshTexture != null) |
| 588 |
{
|
| 589 |
faceMeshTexture.Release(); |
| 590 |
faceMeshTexture = null; |
| 591 |
} |
| 592 |
|
| 593 |
// // clean up |
| 594 |
// Resources.UnloadUnusedAssets(); |
| 595 |
// GC.Collect(); |
| 596 |
|
| 597 |
isFacetrackingInitialized = false; |
| 598 |
instance = null; |
| 599 |
} |
| 600 |
|
| 601 |
void Update() |
| 602 |
{
|
| 603 |
if(isFacetrackingInitialized) |
| 604 |
{
|
| 605 |
KinectManager kinectManager = KinectManager.Instance; |
| 606 |
if(kinectManager && kinectManager.IsInitialized()) |
| 607 |
{
|
| 608 |
lastUserID = primaryUserID; |
| 609 |
primaryUserID = kinectManager.GetUserIdByIndex(playerIndex); |
| 610 |
|
| 611 |
if (primaryUserID != lastUserID && primaryUserID != 0) |
| 612 |
{
|
| 613 |
//faceMeshGotOnce = false; |
| 614 |
} |
| 615 |
} |
| 616 |
|
| 617 |
// update the face tracker |
| 618 |
isTrackingFace = false; |
| 619 |
|
| 620 |
bool bFacetrackingUpdated = !wasFacetrackingActive ? sensorData.sensorInterface.UpdateFaceTracking() : true; |
| 621 |
if(bFacetrackingUpdated) |
| 622 |
{
|
| 623 |
// estimate the tracking state |
| 624 |
isTrackingFace = sensorData.sensorInterface.IsFaceTracked(primaryUserID); |
| 625 |
|
| 626 |
if(!isTrackingFace && (Time.realtimeSinceStartup - lastFaceTrackedTime) <= faceTrackingTolerance) |
| 627 |
{
|
| 628 |
// allow tolerance in tracking |
| 629 |
isTrackingFace = true; |
| 630 |
} |
| 631 |
|
| 632 |
// get the facetracking parameters |
| 633 |
if(isTrackingFace) |
| 634 |
{
|
| 635 |
lastFaceTrackedTime = Time.realtimeSinceStartup; |
| 636 |
facePosUpdateTime = Time.time; |
| 637 |
|
| 638 |
// get face rectangle |
| 639 |
/**bGotFaceRect =*/ sensorData.sensorInterface.GetFaceRect(primaryUserID, ref faceRect); |
| 640 |
|
| 641 |
// get head position |
| 642 |
bGotHeadPos = sensorData.sensorInterface.GetHeadPosition(primaryUserID, ref headPos); |
| 643 |
|
| 644 |
// get head rotation |
| 645 |
bGotHeadRot = sensorData.sensorInterface.GetHeadRotation(primaryUserID, ref headRot); |
| 646 |
|
| 647 |
// get the animation units |
| 648 |
bGotAU = sensorData.sensorInterface.GetAnimUnits(primaryUserID, ref dictAU); |
| 649 |
|
| 650 |
// get the shape units |
| 651 |
bGotSU = sensorData.sensorInterface.GetShapeUnits(primaryUserID, ref dictSU); |
| 652 |
|
| 653 |
//if(faceModelMesh != null && faceModelMesh.activeInHierarchy) |
| 654 |
{
|
| 655 |
// apply model vertices to the mesh |
| 656 |
if(!bFaceModelMeshInited) |
| 657 |
{
|
| 658 |
bFaceModelMeshInited = CreateFaceModelMesh(); |
| 659 |
} |
| 660 |
} |
| 661 |
|
| 662 |
if (getFaceModelData && bFaceModelMeshInited && primaryUserID != 0) |
| 663 |
{
|
| 664 |
if (!pauseModelMeshUpdates && !updateFaceMeshStarted) |
| 665 |
{
|
| 666 |
StartCoroutine(UpdateFaceModelMesh()); |
| 667 |
} |
| 668 |
} |
| 669 |
} |
| 670 |
} |
| 671 |
|
| 672 |
// // set mesh activity flag |
| 673 |
// bool bFaceMeshActive = isTrackingFace && primaryUserID != 0; |
| 674 |
// if(faceModelMesh != null && bFaceModelMeshInited && faceModelMesh.activeSelf != bFaceMeshActive) |
| 675 |
// {
|
| 676 |
// faceModelMesh.SetActive(bFaceMeshActive); |
| 677 |
// } |
| 678 |
} |
| 679 |
} |
| 680 |
|
| 681 |
void OnGUI() |
| 682 |
{
|
| 683 |
if(isFacetrackingInitialized) |
| 684 |
{
|
| 685 |
if(debugText != null) |
| 686 |
{
|
| 687 |
if(isTrackingFace) |
| 688 |
{
|
| 689 |
debugText.text = "Tracking - BodyID: " + primaryUserID; |
| 690 |
} |
| 691 |
else |
| 692 |
{
|
| 693 |
debugText.text = "Not tracking..."; |
| 694 |
} |
| 695 |
} |
| 696 |
} |
| 697 |
} |
| 698 |
|
| 699 |
|
| 700 |
protected bool CreateFaceModelMesh() |
| 701 |
{
|
| 702 |
// if(faceModelMesh == null) |
| 703 |
// return false; |
| 704 |
|
| 705 |
if (avModelVertices == null /**&& !bGotModelVerticesFromDC*/) |
| 706 |
{
|
| 707 |
int iNumVertices = sensorData.sensorInterface.GetFaceModelVerticesCount(0); |
| 708 |
if(iNumVertices <= 0) |
| 709 |
return false; |
| 710 |
|
| 711 |
avModelVertices = new Vector3[iNumVertices]; |
| 712 |
bGotModelVertices = sensorData.sensorInterface.GetFaceModelVertices(0, ref avModelVertices); |
| 713 |
|
| 714 |
avModelUV = new Vector2[iNumVertices]; |
| 715 |
|
| 716 |
if(!bGotModelVertices) |
| 717 |
return false; |
| 718 |
} |
| 719 |
|
| 720 |
// estimate face mesh vertices with respect to the head joint |
| 721 |
Vector3[] vMeshVertices = new Vector3[avModelVertices.Length]; |
| 722 |
|
| 723 |
//if (!bGotModelVerticesFromDC) |
| 724 |
{
|
| 725 |
Vector3 vFaceCenter = Vector3.zero; |
| 726 |
for (int i = 0; i < avModelVertices.Length; i++) |
| 727 |
{
|
| 728 |
vFaceCenter += avModelVertices[i]; |
| 729 |
} |
| 730 |
|
| 731 |
vFaceCenter /= (float)avModelVertices.Length; |
| 732 |
|
| 733 |
faceHeadOffset = Vector3.zero; |
| 734 |
if (vFaceCenter.sqrMagnitude >= 1f) |
| 735 |
{
|
| 736 |
Vector3 vHeadToFace = (vFaceCenter - headPos); |
| 737 |
|
| 738 |
faceHeadOffset = Quaternion.Inverse(headRot) * vHeadToFace; |
| 739 |
faceHeadOffset.y += verticalMeshOffset; |
| 740 |
} |
| 741 |
|
| 742 |
vFaceCenter -= headRot * faceHeadOffset; |
| 743 |
|
| 744 |
for(int i = 0; i < avModelVertices.Length; i++) |
| 745 |
{
|
| 746 |
//avModelVertices[i] = kinectToWorld.MultiplyPoint3x4(avModelVertices[i]) - headPosWorld; |
| 747 |
//avModelVertices[i] -= vFaceCenter; |
| 748 |
|
| 749 |
vMeshVertices[i] = avModelVertices[i] - vFaceCenter; |
| 750 |
} |
| 751 |
} |
| 752 |
|
| 753 |
if (avModelTriangles == null && !bGotModelTrianglesFromDC) |
| 754 |
{
|
| 755 |
int iNumTriangles = sensorData.sensorInterface.GetFaceModelTrianglesCount(); |
| 756 |
if(iNumTriangles <= 0) |
| 757 |
return false; |
| 758 |
|
| 759 |
avModelTriangles = new int[iNumTriangles]; |
| 760 |
bGotModelTriangles = sensorData.sensorInterface.GetFaceModelTriangles(mirroredModelMesh, ref avModelTriangles); |
| 761 |
|
| 762 |
if(!bGotModelTriangles) |
| 763 |
return false; |
| 764 |
} |
| 765 |
|
| 766 |
if (faceModelMesh) |
| 767 |
{
|
| 768 |
Mesh mesh = new Mesh(); |
| 769 |
mesh.name = "FaceMesh"; |
| 770 |
faceModelMesh.GetComponent<MeshFilter>().mesh = mesh; |
| 771 |
|
| 772 |
mesh.vertices = vMeshVertices; // avModelVertices; |
| 773 |
//mesh.uv = avModelUV; |
| 774 |
|
| 775 |
mesh.triangles = avModelTriangles; |
| 776 |
mesh.RecalculateNormals(); |
| 777 |
|
| 778 |
// if (moveModelMesh) |
| 779 |
// {
|
| 780 |
// faceModelMesh.transform.position = headPos; |
| 781 |
// //faceModelMesh.transform.rotation = faceModelRot; |
| 782 |
// } |
| 783 |
|
| 784 |
SetFaceModelMeshTexture(); |
| 785 |
} |
| 786 |
|
| 787 |
//bFaceModelMeshInited = true; |
| 788 |
return true; |
| 789 |
} |
| 790 |
|
| 791 |
// sets the proper face mesh texture |
| 792 |
protected void SetFaceModelMeshTexture() |
| 793 |
{
|
| 794 |
if (texturedModelMesh == TextureType.ColorMap) |
| 795 |
{
|
| 796 |
KinectManager kinectManager = KinectManager.Instance; |
| 797 |
Texture texColorMap = kinectManager ? kinectManager.GetUsersClrTex() : null; |
| 798 |
|
| 799 |
if (!faceMeshTexture && kinectManager && texColorMap) |
| 800 |
{
|
| 801 |
faceMeshTexture = new RenderTexture (texColorMap.width, texColorMap.height, 0); |
| 802 |
faceModelMesh.GetComponent<MeshRenderer>().material.mainTexture = faceMeshTexture; // kinectManager.GetUsersClrTex(); |
| 803 |
} |
| 804 |
|
| 805 |
if (faceMeshTexture && texColorMap) |
| 806 |
{
|
| 807 |
// update the color texture |
| 808 |
Graphics.Blit(texColorMap, faceMeshTexture); |
| 809 |
} |
| 810 |
} |
| 811 |
else if (texturedModelMesh == TextureType.FaceRectangle) |
| 812 |
{
|
| 813 |
if (faceMeshTexture != null) |
| 814 |
{
|
| 815 |
faceMeshTexture = null; |
| 816 |
} |
| 817 |
} |
| 818 |
else if(texturedModelMesh == TextureType.None) |
| 819 |
{
|
| 820 |
if (faceModelMesh.GetComponent<MeshRenderer>().material.mainTexture != null) |
| 821 |
{
|
| 822 |
faceMeshTexture = null; |
| 823 |
faceModelMesh.GetComponent<MeshRenderer>().material.mainTexture = null; |
| 824 |
} |
| 825 |
} |
| 826 |
} |
| 827 |
|
| 828 |
|
| 829 |
protected IEnumerator UpdateFaceModelMesh() |
| 830 |
{
|
| 831 |
updateFaceMeshStarted = true; |
| 832 |
|
| 833 |
//if (!dontUpdateModelMesh || !faceMeshGotOnce /**&& !bGotModelVerticesFromDC*/) |
| 834 |
{
|
| 835 |
// init the vertices array if needed |
| 836 |
if(avModelVertices == null) |
| 837 |
{
|
| 838 |
int iNumVertices = sensorData.sensorInterface.GetFaceModelVerticesCount(primaryUserID); |
| 839 |
avModelVertices = new Vector3[iNumVertices]; |
| 840 |
} |
| 841 |
|
| 842 |
// get face model vertices |
| 843 |
bGotModelVertices = sensorData.sensorInterface.GetFaceModelVertices(primaryUserID, ref avModelVertices); |
| 844 |
} |
| 845 |
|
| 846 |
if(bGotModelVertices && faceModelMesh != null) |
| 847 |
{
|
| 848 |
//Quaternion faceModelRot = faceModelMesh.transform.rotation; |
| 849 |
//faceModelMesh.transform.rotation = Quaternion.identity; |
| 850 |
|
| 851 |
bool bFaceMeshUpdated = false; |
| 852 |
//if (!dontUpdateModelMesh || !faceMeshGotOnce) |
| 853 |
{
|
| 854 |
AsyncTask<bool> task = new AsyncTask<bool>(() => {
|
| 855 |
// estimate face mesh vertices with respect to the head joint |
| 856 |
vMeshVertices = null; |
| 857 |
|
| 858 |
KinectManager kinectManager = KinectManager.Instance; |
| 859 |
Matrix4x4 kinectToWorld = kinectManager ? kinectManager.GetKinectToWorldMatrix() : Matrix4x4.identity; |
| 860 |
Vector3 headPosWorld = kinectToWorld.MultiplyPoint3x4(headPos); |
| 861 |
|
| 862 |
Vector3 lastNosePos = nosePos; |
| 863 |
//if (!bGotModelVerticesFromDC) |
| 864 |
{
|
| 865 |
// Vector3 vFaceCenter = Vector3.zero; |
| 866 |
// for (int i = 0; i < avModelVertices.Length; i++) |
| 867 |
// {
|
| 868 |
// vFaceCenter += avModelVertices[i]; |
| 869 |
// } |
| 870 |
// |
| 871 |
// vFaceCenter /= (float)avModelVertices.Length; |
| 872 |
// |
| 873 |
// Vector3 vHeadToFace = (vFaceCenter - headPos); |
| 874 |
// if (vHeadToFace.sqrMagnitude < 0.015f) // max 0.12 x 0.12 |
| 875 |
// {
|
| 876 |
// faceHeadOffset = Quaternion.Inverse(headRot) * vHeadToFace; |
| 877 |
// faceHeadOffset.y += verticalMeshOffset; |
| 878 |
// } |
| 879 |
|
| 880 |
nosePos = GetFaceModelNosePos(); |
| 881 |
Vector3 vHeadToNose = Quaternion.Inverse(headRot) * (nosePos - headPos); |
| 882 |
float headToNoseLen = vHeadToNose.magnitude; |
| 883 |
|
| 884 |
// string sHeadToNose = string.Format("({0:F2}, {0:F2}, {0:F2})", vHeadToNose.x, vHeadToNose.y, vHeadToNose.z);
|
| 885 |
// Debug.Log("U-Face nosePos: " + nosePos + ", headPos: " + headPos + "\noffset: " + sHeadToNose + ", len: " + headToNoseLen);
|
| 886 |
|
| 887 |
if(headToNoseLen >= 0.08f && headToNoseLen <= 0.18f) |
| 888 |
{
|
| 889 |
//vFaceCenter -= headRot * faceHeadOffset; |
| 890 |
|
| 891 |
vMeshVertices = new Vector3[avModelVertices.Length]; |
| 892 |
for(int i = 0; i < avModelVertices.Length; i++) |
| 893 |
{
|
| 894 |
//avModelVertices[i] = kinectToWorld.MultiplyPoint3x4(avModelVertices[i]) - headPosWorld; |
| 895 |
//avModelVertices[i] -= vFaceCenter; |
| 896 |
|
| 897 |
//vMeshVertices[i] = avModelVertices[i] - vFaceCenter; |
| 898 |
vMeshVertices[i] = kinectToWorld.MultiplyPoint3x4(avModelVertices[i]) - headPosWorld; // avModelVertices[i] - headPos; |
| 899 |
} |
| 900 |
} |
| 901 |
} |
| 902 |
|
| 903 |
if(vMeshVertices == null || lastNosePos == nosePos) |
| 904 |
{
|
| 905 |
return false; |
| 906 |
} |
| 907 |
|
| 908 |
//if (!bGotModelVerticesFromDC) |
| 909 |
{
|
| 910 |
if(texturedModelMesh != TextureType.None) |
| 911 |
{
|
| 912 |
float colorWidth = (float)kinectManager.GetColorImageWidth(); |
| 913 |
float colorHeight = (float)kinectManager.GetColorImageHeight(); |
| 914 |
|
| 915 |
//bool bGotFaceRect = sensorData.sensorInterface.GetFaceRect(userId, ref faceRect); |
| 916 |
bool faceRectValid = /**bGotFaceRect &&*/ faceRect.width > 0 && faceRect.height > 0; |
| 917 |
|
| 918 |
for(int i = 0; i < avModelVertices.Length; i++) |
| 919 |
{
|
| 920 |
Vector2 posDepth = kinectManager.MapSpacePointToDepthCoords(avModelVertices[i]); |
| 921 |
|
| 922 |
bool bUvSet = false; |
| 923 |
if(posDepth != Vector2.zero) |
| 924 |
{
|
| 925 |
ushort depth = kinectManager.GetDepthForPixel((int)posDepth.x, (int)posDepth.y); |
| 926 |
Vector2 posColor = kinectManager.MapDepthPointToColorCoords(posDepth, depth); |
| 927 |
|
| 928 |
if(posColor != Vector2.zero && !float.IsInfinity(posColor.x) && !float.IsInfinity(posColor.y)) |
| 929 |
{
|
| 930 |
if(texturedModelMesh == TextureType.ColorMap) |
| 931 |
{
|
| 932 |
avModelUV[i] = new Vector2(posColor.x / colorWidth, posColor.y / colorHeight); |
| 933 |
bUvSet = true; |
| 934 |
} |
| 935 |
else if(texturedModelMesh == TextureType.FaceRectangle && faceRectValid) |
| 936 |
{
|
| 937 |
avModelUV[i] = new Vector2(Mathf.Clamp01((posColor.x - faceRect.x) / faceRect.width), |
| 938 |
-Mathf.Clamp01((posColor.y - faceRect.y) / faceRect.height)); |
| 939 |
bUvSet = true; |
| 940 |
} |
| 941 |
} |
| 942 |
} |
| 943 |
|
| 944 |
if(!bUvSet) |
| 945 |
{
|
| 946 |
avModelUV[i] = Vector2.zero; |
| 947 |
} |
| 948 |
} |
| 949 |
} |
| 950 |
} |
| 951 |
|
| 952 |
return true; |
| 953 |
}); |
| 954 |
|
| 955 |
task.Start(); |
| 956 |
|
| 957 |
while (task.State == AsyncTaskState.Running) |
| 958 |
{
|
| 959 |
yield return null; |
| 960 |
} |
| 961 |
|
| 962 |
// // show nose & head positions |
| 963 |
// Matrix4x4 kinectToWorld2 = KinectManager.Instance.GetKinectToWorldMatrix(); |
| 964 |
// if (noseTransform) |
| 965 |
// noseTransform.position = kinectToWorld2.MultiplyPoint3x4(nosePos); |
| 966 |
// if(headTransform) |
| 967 |
// headTransform.position = kinectToWorld2.MultiplyPoint3x4(headPos); |
| 968 |
// |
| 969 |
// Vector3 vHeadToNose2 = Quaternion.Inverse(headRot) * (nosePos - headPos); |
| 970 |
// string sHeadToNose2 = string.Format("({0:F2}, {0:F2}, {0:F2})", vHeadToNose2.x, vHeadToNose2.y, vHeadToNose2.z);
|
| 971 |
// if(debugText2) |
| 972 |
// debugText2.text = "h2n: " + sHeadToNose2 + ", len: " + vHeadToNose2.magnitude; |
| 973 |
|
| 974 |
bFaceMeshUpdated = task.Result; |
| 975 |
if(bFaceMeshUpdated) |
| 976 |
{
|
| 977 |
Mesh mesh = faceModelMesh.GetComponent<MeshFilter>().mesh; |
| 978 |
mesh.vertices = vMeshVertices; // avModelVertices; |
| 979 |
vMeshVertices = null; |
| 980 |
|
| 981 |
if(texturedModelMesh != TextureType.None && avModelUV != null) |
| 982 |
{
|
| 983 |
mesh.uv = avModelUV; |
| 984 |
} |
| 985 |
|
| 986 |
faceMeshUpdateTime = Time.time; |
| 987 |
//faceMeshGotOnce = true; |
| 988 |
|
| 989 |
mesh.RecalculateNormals(); |
| 990 |
mesh.RecalculateBounds(); |
| 991 |
|
| 992 |
// set the face mesh texture |
| 993 |
SetFaceModelMeshTexture(); |
| 994 |
} |
| 995 |
} |
| 996 |
|
| 997 |
if (moveModelMesh) |
| 998 |
{
|
| 999 |
KinectManager kinectManager = KinectManager.Instance; |
| 1000 |
Matrix4x4 kinectToWorld = kinectManager ? kinectManager.GetKinectToWorldMatrix() : Matrix4x4.identity; |
| 1001 |
Vector3 newHeadPos = kinectToWorld.MultiplyPoint3x4(headPos); |
| 1002 |
|
| 1003 |
// check for head pos overlay |
| 1004 |
if(foregroundCamera) |
| 1005 |
{
|
| 1006 |
// get the background rectangle (use the portrait background, if available) |
| 1007 |
Rect backgroundRect = foregroundCamera.pixelRect; |
| 1008 |
PortraitBackground portraitBack = PortraitBackground.Instance; |
| 1009 |
|
| 1010 |
if(portraitBack && portraitBack.enabled) |
| 1011 |
{
|
| 1012 |
backgroundRect = portraitBack.GetBackgroundRect(); |
| 1013 |
} |
| 1014 |
|
| 1015 |
if(kinectManager) |
| 1016 |
{
|
| 1017 |
Vector3 posColorOverlay = kinectManager.GetJointPosColorOverlay(primaryUserID, (int)KinectInterop.JointType.Head, foregroundCamera, backgroundRect); |
| 1018 |
|
| 1019 |
if(posColorOverlay != Vector3.zero) |
| 1020 |
{
|
| 1021 |
newHeadPos = posColorOverlay; |
| 1022 |
} |
| 1023 |
} |
| 1024 |
} |
| 1025 |
|
| 1026 |
faceModelMesh.transform.position = newHeadPos; // Vector3.Lerp(faceModelMesh.transform.position, newHeadPos, 20f * Time.deltaTime); |
| 1027 |
//faceModelMesh.transform.rotation = faceModelRot; |
| 1028 |
} |
| 1029 |
|
| 1030 |
// don't rotate the transform - mesh follows the head rotation |
| 1031 |
if (faceModelMesh.transform.rotation != Quaternion.identity) |
| 1032 |
{
|
| 1033 |
faceModelMesh.transform.rotation = Quaternion.identity; |
| 1034 |
} |
| 1035 |
|
| 1036 |
// apply scale factor |
| 1037 |
if(faceModelMesh.transform.localScale.x != modelMeshScale) |
| 1038 |
{
|
| 1039 |
faceModelMesh.transform.localScale = new Vector3(modelMeshScale, modelMeshScale, modelMeshScale); |
| 1040 |
} |
| 1041 |
|
| 1042 |
if(!faceModelMesh.activeSelf) |
| 1043 |
{
|
| 1044 |
faceModelMesh.SetActive(true); |
| 1045 |
} |
| 1046 |
} |
| 1047 |
else |
| 1048 |
{
|
| 1049 |
if(faceModelMesh && faceModelMesh.activeSelf) |
| 1050 |
{
|
| 1051 |
faceModelMesh.SetActive(false); |
| 1052 |
} |
| 1053 |
} |
| 1054 |
|
| 1055 |
updateFaceMeshStarted = false; |
| 1056 |
} |
| 1057 |
|
| 1058 |
// returns the nose tip position, or Vector3.zero if not found |
| 1059 |
private Vector3 GetFaceModelNosePos() |
| 1060 |
{
|
| 1061 |
if (avModelVertices != null) |
| 1062 |
{
|
| 1063 |
int iNoseIndex = -1; |
| 1064 |
if (sensorData.sensorIntPlatform == KinectInterop.DepthSensorPlatform.KinectSDKv2 || |
| 1065 |
sensorData.sensorIntPlatform == KinectInterop.DepthSensorPlatform.KinectUWPv2 || |
| 1066 |
sensorData.sensorIntPlatform == KinectInterop.DepthSensorPlatform.DummyK2) |
| 1067 |
{
|
| 1068 |
iNoseIndex = 18; // Microsoft.Kinect.Face.HighDetailFacePoints.NoseTip |
| 1069 |
} |
| 1070 |
else if (sensorData.sensorIntPlatform == KinectInterop.DepthSensorPlatform.KinectSDKv1 || |
| 1071 |
sensorData.sensorIntPlatform == KinectInterop.DepthSensorPlatform.DummyK1) |
| 1072 |
{
|
| 1073 |
iNoseIndex = 89; // |
| 1074 |
} |
| 1075 |
|
| 1076 |
if (iNoseIndex >= 0 && iNoseIndex < avModelVertices.Length) |
| 1077 |
{
|
| 1078 |
return avModelVertices[iNoseIndex]; |
| 1079 |
} |
| 1080 |
} |
| 1081 |
|
| 1082 |
return Vector3.zero; |
| 1083 |
} |
| 1084 |
|
| 1085 |
// gets face basic parameters as csv line |
| 1086 |
public string GetFaceParamsAsCsv() |
| 1087 |
{
|
| 1088 |
// create the output string |
| 1089 |
StringBuilder sbBuf = new StringBuilder(); |
| 1090 |
const char delimiter = ','; |
| 1091 |
|
| 1092 |
if (bGotHeadPos || bGotHeadRot) |
| 1093 |
{
|
| 1094 |
sbBuf.Append("fp").Append(delimiter);
|
| 1095 |
|
| 1096 |
// head pos |
| 1097 |
sbBuf.Append (bGotHeadPos ? "1" : "0").Append(delimiter); |
| 1098 |
|
| 1099 |
if (bGotHeadPos) |
| 1100 |
{
|
| 1101 |
sbBuf.AppendFormat ("{0:F3}", headPos.x).Append (delimiter);
|
| 1102 |
sbBuf.AppendFormat ("{0:F3}", headPos.y).Append (delimiter);
|
| 1103 |
sbBuf.AppendFormat ("{0:F3}", headPos.z).Append (delimiter);
|
| 1104 |
} |
| 1105 |
|
| 1106 |
// head rot |
| 1107 |
sbBuf.Append (bGotHeadRot ? "1" : "0").Append(delimiter); |
| 1108 |
Vector3 vheadRot = headRot.eulerAngles; |
| 1109 |
|
| 1110 |
if (bGotHeadRot) |
| 1111 |
{
|
| 1112 |
sbBuf.AppendFormat ("{0:F3}", vheadRot.x).Append (delimiter);
|
| 1113 |
sbBuf.AppendFormat ("{0:F3}", vheadRot.y).Append (delimiter);
|
| 1114 |
sbBuf.AppendFormat ("{0:F3}", vheadRot.z).Append (delimiter);
|
| 1115 |
} |
| 1116 |
|
| 1117 |
// face rect |
| 1118 |
sbBuf.Append ("1").Append(delimiter);
|
| 1119 |
sbBuf.AppendFormat ("{0:F0}", faceRect.x).Append (delimiter);
|
| 1120 |
sbBuf.AppendFormat ("{0:F0}", faceRect.y).Append (delimiter);
|
| 1121 |
sbBuf.AppendFormat ("{0:F0}", faceRect.width).Append (delimiter);
|
| 1122 |
sbBuf.AppendFormat ("{0:F0}", faceRect.height).Append (delimiter);
|
| 1123 |
|
| 1124 |
// animation units |
| 1125 |
sbBuf.Append (bGotAU ? "1" : "0").Append(delimiter); |
| 1126 |
|
| 1127 |
if (bGotAU) |
| 1128 |
{
|
| 1129 |
int enumCount = Enum.GetNames (typeof(KinectInterop.FaceShapeAnimations)).Length; |
| 1130 |
sbBuf.Append (enumCount).Append(delimiter); |
| 1131 |
|
| 1132 |
for (int i = 0; i < enumCount; i++) |
| 1133 |
{
|
| 1134 |
float dictValue = dictAU [(KinectInterop.FaceShapeAnimations)i]; |
| 1135 |
sbBuf.AppendFormat ("{0:F3}", dictValue).Append (delimiter);
|
| 1136 |
} |
| 1137 |
} |
| 1138 |
|
| 1139 |
// shape units |
| 1140 |
sbBuf.Append (bGotSU ? "1" : "0").Append(delimiter); |
| 1141 |
|
| 1142 |
if (bGotSU) |
| 1143 |
{
|
| 1144 |
int enumCount = Enum.GetNames (typeof(KinectInterop.FaceShapeDeformations)).Length; |
| 1145 |
sbBuf.Append (enumCount).Append(delimiter); |
| 1146 |
|
| 1147 |
for (int i = 0; i < enumCount; i++) |
| 1148 |
{
|
| 1149 |
float dictValue = dictSU [(KinectInterop.FaceShapeDeformations)i]; |
| 1150 |
sbBuf.AppendFormat ("{0:F3}", dictValue).Append (delimiter);
|
| 1151 |
} |
| 1152 |
} |
| 1153 |
|
| 1154 |
// any other parameters... |
| 1155 |
} |
| 1156 |
|
| 1157 |
// remove the last delimiter |
| 1158 |
if(sbBuf.Length > 0 && sbBuf[sbBuf.Length - 1] == delimiter) |
| 1159 |
{
|
| 1160 |
sbBuf.Remove(sbBuf.Length - 1, 1); |
| 1161 |
} |
| 1162 |
|
| 1163 |
return sbBuf.ToString(); |
| 1164 |
} |
| 1165 |
|
| 1166 |
// sets basic face parameters from a csv line |
| 1167 |
public bool SetFaceParamsFromCsv(string sCsvLine) |
| 1168 |
{
|
| 1169 |
if(sCsvLine.Length == 0) |
| 1170 |
return false; |
| 1171 |
|
| 1172 |
// split the csv line in parts |
| 1173 |
char[] delimiters = { ',' };
|
| 1174 |
string[] alCsvParts = sCsvLine.Split(delimiters); |
| 1175 |
|
| 1176 |
if(alCsvParts.Length < 1 || alCsvParts[0] != "fp") |
| 1177 |
return false; |
| 1178 |
|
| 1179 |
int iIndex = 1; |
| 1180 |
int iLength = alCsvParts.Length; |
| 1181 |
|
| 1182 |
if (iLength < (iIndex + 1)) |
| 1183 |
return false; |
| 1184 |
|
| 1185 |
// head pos |
| 1186 |
bGotHeadPos = (alCsvParts[iIndex] == "1"); |
| 1187 |
iIndex++; |
| 1188 |
|
| 1189 |
if (bGotHeadPos && iLength >= (iIndex + 3)) |
| 1190 |
{
|
| 1191 |
float x = 0f, y = 0f, z = 0f; |
| 1192 |
|
| 1193 |
float.TryParse(alCsvParts[iIndex], out x); |
| 1194 |
float.TryParse(alCsvParts[iIndex + 1], out y); |
| 1195 |
float.TryParse(alCsvParts[iIndex + 2], out z); |
| 1196 |
iIndex += 3; |
| 1197 |
|
| 1198 |
headPos = new Vector3(x, y, z); |
| 1199 |
} |
| 1200 |
|
| 1201 |
// head rot |
| 1202 |
bGotHeadRot = (alCsvParts[iIndex] == "1"); |
| 1203 |
iIndex++; |
| 1204 |
|
| 1205 |
if (bGotHeadRot && iLength >= (iIndex + 3)) |
| 1206 |
{
|
| 1207 |
float x = 0f, y = 0f, z = 0f; |
| 1208 |
|
| 1209 |
float.TryParse(alCsvParts[iIndex], out x); |
| 1210 |
float.TryParse(alCsvParts[iIndex + 1], out y); |
| 1211 |
float.TryParse(alCsvParts[iIndex + 2], out z); |
| 1212 |
iIndex += 3; |
| 1213 |
|
| 1214 |
headRot = Quaternion.Euler(x, y, z); |
| 1215 |
} |
| 1216 |
|
| 1217 |
// face rect |
| 1218 |
bool bGotFaceRect = (alCsvParts[iIndex] == "1"); |
| 1219 |
iIndex++; |
| 1220 |
|
| 1221 |
if (bGotFaceRect && iLength >= (iIndex + 4)) |
| 1222 |
{
|
| 1223 |
float x = 0f, y = 0f, w = 0f, h = 0f; |
| 1224 |
|
| 1225 |
float.TryParse(alCsvParts[iIndex], out x); |
| 1226 |
float.TryParse(alCsvParts[iIndex + 1], out y); |
| 1227 |
float.TryParse(alCsvParts[iIndex + 2], out w); |
| 1228 |
float.TryParse(alCsvParts[iIndex + 3], out h); |
| 1229 |
iIndex += 4; |
| 1230 |
|
| 1231 |
faceRect.x = x; faceRect.y = y; |
| 1232 |
faceRect.width = w; faceRect.height = h; |
| 1233 |
} |
| 1234 |
|
| 1235 |
// animation units |
| 1236 |
bGotAU = (alCsvParts[iIndex] == "1"); |
| 1237 |
iIndex++; |
| 1238 |
|
| 1239 |
if (bGotAU && iLength >= (iIndex + 1)) |
| 1240 |
{
|
| 1241 |
int count = 0; |
| 1242 |
int.TryParse(alCsvParts[iIndex], out count); |
| 1243 |
iIndex++; |
| 1244 |
|
| 1245 |
for (int i = 0; i < count && iLength >= (iIndex + 1); i++) |
| 1246 |
{
|
| 1247 |
float v = 0; |
| 1248 |
float.TryParse(alCsvParts[iIndex], out v); |
| 1249 |
iIndex++; |
| 1250 |
|
| 1251 |
dictAU [(KinectInterop.FaceShapeAnimations)i] = v; |
| 1252 |
} |
| 1253 |
} |
| 1254 |
|
| 1255 |
// shape units |
| 1256 |
bGotSU = (alCsvParts[iIndex] == "1"); |
| 1257 |
iIndex++; |
| 1258 |
|
| 1259 |
if (bGotSU && iLength >= (iIndex + 1)) |
| 1260 |
{
|
| 1261 |
int count = 0; |
| 1262 |
int.TryParse(alCsvParts[iIndex], out count); |
| 1263 |
iIndex++; |
| 1264 |
|
| 1265 |
for (int i = 0; i < count && iLength >= (iIndex + 1); i++) |
| 1266 |
{
|
| 1267 |
float v = 0; |
| 1268 |
float.TryParse(alCsvParts[iIndex], out v); |
| 1269 |
iIndex++; |
| 1270 |
|
| 1271 |
dictSU [(KinectInterop.FaceShapeDeformations)i] = v; |
| 1272 |
} |
| 1273 |
} |
| 1274 |
|
| 1275 |
// any other parameters here... |
| 1276 |
|
| 1277 |
// emulate face tracking |
| 1278 |
lastFaceTrackedTime = Time.realtimeSinceStartup; |
| 1279 |
facePosUpdateTime = Time.time; |
| 1280 |
|
| 1281 |
return true; |
| 1282 |
} |
| 1283 |
|
| 1284 |
// gets face model vertices as csv line |
| 1285 |
public string GetFaceVerticesAsCsv() |
| 1286 |
{
|
| 1287 |
// create the output string |
| 1288 |
StringBuilder sbBuf = new StringBuilder(); |
| 1289 |
const char delimiter = ','; |
| 1290 |
|
| 1291 |
if (bGotModelVertices && avModelVertices != null) |
| 1292 |
{
|
| 1293 |
sbBuf.Append("fv").Append(delimiter);
|
| 1294 |
|
| 1295 |
// model vertices |
| 1296 |
int vertCount = avModelVertices.Length; |
| 1297 |
sbBuf.Append (vertCount).Append(delimiter); |
| 1298 |
|
| 1299 |
for (int i = 0; i < vertCount; i++) |
| 1300 |
{
|
| 1301 |
sbBuf.AppendFormat ("{0:F3}", avModelVertices[i].x).Append (delimiter);
|
| 1302 |
sbBuf.AppendFormat ("{0:F3}", avModelVertices[i].y).Append (delimiter);
|
| 1303 |
sbBuf.AppendFormat ("{0:F3}", avModelVertices[i].z).Append (delimiter);
|
| 1304 |
} |
| 1305 |
} |
| 1306 |
|
| 1307 |
// remove the last delimiter |
| 1308 |
if(sbBuf.Length > 0 && sbBuf[sbBuf.Length - 1] == delimiter) |
| 1309 |
{
|
| 1310 |
sbBuf.Remove(sbBuf.Length - 1, 1); |
| 1311 |
} |
| 1312 |
|
| 1313 |
return sbBuf.ToString(); |
| 1314 |
} |
| 1315 |
|
| 1316 |
// sets face model vertices from a csv line |
| 1317 |
public bool SetFaceVerticesFromCsv(string sCsvLine) |
| 1318 |
{
|
| 1319 |
if(sCsvLine.Length == 0) |
| 1320 |
return false; |
| 1321 |
|
| 1322 |
// split the csv line in parts |
| 1323 |
char[] delimiters = { ',' };
|
| 1324 |
string[] alCsvParts = sCsvLine.Split(delimiters); |
| 1325 |
|
| 1326 |
if(alCsvParts.Length < 1 || alCsvParts[0] != "fv") |
| 1327 |
return false; |
| 1328 |
|
| 1329 |
int iIndex = 1; |
| 1330 |
int iLength = alCsvParts.Length; |
| 1331 |
|
| 1332 |
if (iLength < (iIndex + 1)) |
| 1333 |
return false; |
| 1334 |
|
| 1335 |
// model vertices |
| 1336 |
int vertCount = 0; |
| 1337 |
int.TryParse(alCsvParts[iIndex], out vertCount); |
| 1338 |
iIndex++; |
| 1339 |
|
| 1340 |
if (vertCount > 0) |
| 1341 |
{
|
| 1342 |
if (avModelVertices == null || avModelVertices.Length != vertCount) |
| 1343 |
{
|
| 1344 |
avModelVertices = new Vector3[vertCount]; |
| 1345 |
} |
| 1346 |
|
| 1347 |
for (int i = 0; i < vertCount && iLength >= (iIndex + 3); i++) |
| 1348 |
{
|
| 1349 |
float x = 0f, y = 0f, z = 0f; |
| 1350 |
|
| 1351 |
float.TryParse(alCsvParts[iIndex], out x); |
| 1352 |
float.TryParse(alCsvParts[iIndex + 1], out y); |
| 1353 |
float.TryParse(alCsvParts[iIndex + 2], out z); |
| 1354 |
iIndex += 3; |
| 1355 |
|
| 1356 |
avModelVertices[i] = new Vector3(x, y, z); |
| 1357 |
} |
| 1358 |
|
| 1359 |
bGotModelVertices = true; |
| 1360 |
//bGotModelVerticesFromDC = true; |
| 1361 |
} |
| 1362 |
|
| 1363 |
faceMeshUpdateTime = Time.time; |
| 1364 |
|
| 1365 |
return true; |
| 1366 |
} |
| 1367 |
|
| 1368 |
// gets face model UVs as csv line |
| 1369 |
public string GetFaceUvsAsCsv() |
| 1370 |
{
|
| 1371 |
// create the output string |
| 1372 |
StringBuilder sbBuf = new StringBuilder(); |
| 1373 |
const char delimiter = ','; |
| 1374 |
|
| 1375 |
if (bGotModelVertices && avModelUV != null) |
| 1376 |
{
|
| 1377 |
sbBuf.Append("fu").Append(delimiter);
|
| 1378 |
|
| 1379 |
// face rect width & height |
| 1380 |
sbBuf.AppendFormat ("{0:F0}", faceRect.width).Append (delimiter);
|
| 1381 |
sbBuf.AppendFormat ("{0:F0}", faceRect.height).Append (delimiter);
|
| 1382 |
|
| 1383 |
// model UVs |
| 1384 |
int uvCount = avModelUV.Length; |
| 1385 |
sbBuf.Append (uvCount).Append(delimiter); |
| 1386 |
|
| 1387 |
for (int i = 0; i < uvCount; i++) |
| 1388 |
{
|
| 1389 |
sbBuf.AppendFormat ("{0:F3}", avModelUV[i].x).Append (delimiter);
|
| 1390 |
sbBuf.AppendFormat ("{0:F3}", avModelUV[i].y).Append (delimiter);
|
| 1391 |
} |
| 1392 |
} |
| 1393 |
|
| 1394 |
// remove the last delimiter |
| 1395 |
if(sbBuf.Length > 0 && sbBuf[sbBuf.Length - 1] == delimiter) |
| 1396 |
{
|
| 1397 |
sbBuf.Remove(sbBuf.Length - 1, 1); |
| 1398 |
} |
| 1399 |
|
| 1400 |
return sbBuf.ToString(); |
| 1401 |
} |
| 1402 |
|
| 1403 |
// sets face model UVs from a csv line |
| 1404 |
public bool SetFaceUvsFromCsv(string sCsvLine) |
| 1405 |
{
|
| 1406 |
if(sCsvLine.Length == 0) |
| 1407 |
return false; |
| 1408 |
|
| 1409 |
// split the csv line in parts |
| 1410 |
char[] delimiters = { ',' };
|
| 1411 |
string[] alCsvParts = sCsvLine.Split(delimiters); |
| 1412 |
|
| 1413 |
if(alCsvParts.Length < 1 || alCsvParts[0] != "fu") |
| 1414 |
return false; |
| 1415 |
|
| 1416 |
int iIndex = 1; |
| 1417 |
int iLength = alCsvParts.Length; |
| 1418 |
|
| 1419 |
if (iLength < (iIndex + 2)) |
| 1420 |
return false; |
| 1421 |
|
| 1422 |
// face width & height |
| 1423 |
float w = 0f, h = 0f; |
| 1424 |
|
| 1425 |
float.TryParse(alCsvParts[iIndex], out w); |
| 1426 |
float.TryParse(alCsvParts[iIndex + 1], out h); |
| 1427 |
iIndex += 2; |
| 1428 |
|
| 1429 |
faceRect.width = w; faceRect.height = h; |
| 1430 |
|
| 1431 |
// model UVs |
| 1432 |
int uvCount = 0; |
| 1433 |
if (iLength >= (iIndex + 1)) |
| 1434 |
{
|
| 1435 |
int.TryParse(alCsvParts[iIndex], out uvCount); |
| 1436 |
iIndex++; |
| 1437 |
} |
| 1438 |
|
| 1439 |
if (uvCount > 0) |
| 1440 |
{
|
| 1441 |
if (avModelUV == null || avModelUV.Length != uvCount) |
| 1442 |
{
|
| 1443 |
avModelUV = new Vector2[uvCount]; |
| 1444 |
} |
| 1445 |
|
| 1446 |
for (int i = 0; i < uvCount && iLength >= (iIndex + 2); i++) |
| 1447 |
{
|
| 1448 |
float x = 0f, y = 0f; |
| 1449 |
|
| 1450 |
float.TryParse(alCsvParts[iIndex], out x); |
| 1451 |
float.TryParse(alCsvParts[iIndex + 1], out y); |
| 1452 |
iIndex += 2; |
| 1453 |
|
| 1454 |
avModelUV[i] = new Vector2(x, y); |
| 1455 |
} |
| 1456 |
} |
| 1457 |
|
| 1458 |
return true; |
| 1459 |
} |
| 1460 |
|
| 1461 |
// gets face model triangles as csv line |
| 1462 |
public string GetFaceTrianglesAsCsv() |
| 1463 |
{
|
| 1464 |
// create the output string |
| 1465 |
StringBuilder sbBuf = new StringBuilder(); |
| 1466 |
const char delimiter = ','; |
| 1467 |
|
| 1468 |
if (avModelTriangles != null) |
| 1469 |
{
|
| 1470 |
sbBuf.Append("ft").Append(delimiter);
|
| 1471 |
|
| 1472 |
// model triangles |
| 1473 |
int triCount = avModelTriangles.Length; |
| 1474 |
sbBuf.Append (triCount).Append(delimiter); |
| 1475 |
|
| 1476 |
for (int i = 0; i < triCount; i++) |
| 1477 |
{
|
| 1478 |
sbBuf.Append(avModelTriangles[i]).Append (delimiter); |
| 1479 |
} |
| 1480 |
} |
| 1481 |
|
| 1482 |
// remove the last delimiter |
| 1483 |
if(sbBuf.Length > 0 && sbBuf[sbBuf.Length - 1] == delimiter) |
| 1484 |
{
|
| 1485 |
sbBuf.Remove(sbBuf.Length - 1, 1); |
| 1486 |
} |
| 1487 |
|
| 1488 |
return sbBuf.ToString(); |
| 1489 |
} |
| 1490 |
|
| 1491 |
// sets face model model from a csv line |
| 1492 |
public bool SetFaceTrianglesFromCsv(string sCsvLine) |
| 1493 |
{
|
| 1494 |
if(sCsvLine.Length == 0) |
| 1495 |
return false; |
| 1496 |
|
| 1497 |
// split the csv line in parts |
| 1498 |
char[] delimiters = { ',' };
|
| 1499 |
string[] alCsvParts = sCsvLine.Split(delimiters); |
| 1500 |
|
| 1501 |
if(alCsvParts.Length < 1 || alCsvParts[0] != "ft") |
| 1502 |
return false; |
| 1503 |
|
| 1504 |
int iIndex = 1; |
| 1505 |
int iLength = alCsvParts.Length; |
| 1506 |
|
| 1507 |
if (iLength < (iIndex + 1)) |
| 1508 |
return false; |
| 1509 |
|
| 1510 |
// model triangles |
| 1511 |
int triCount = 0; |
| 1512 |
int.TryParse(alCsvParts[iIndex], out triCount); |
| 1513 |
iIndex++; |
| 1514 |
|
| 1515 |
if (triCount > 0) |
| 1516 |
{
|
| 1517 |
if (avModelTriangles == null || avModelTriangles.Length != triCount) |
| 1518 |
{
|
| 1519 |
avModelTriangles = new int[triCount]; |
| 1520 |
} |
| 1521 |
|
| 1522 |
for (int i = 0; i < triCount && iLength >= (iIndex + 1); i++) |
| 1523 |
{
|
| 1524 |
int v = 0; |
| 1525 |
|
| 1526 |
int.TryParse(alCsvParts[iIndex], out v); |
| 1527 |
iIndex++; |
| 1528 |
|
| 1529 |
avModelTriangles[i] = v; |
| 1530 |
} |
| 1531 |
|
| 1532 |
bGotModelTriangles = true; |
| 1533 |
bGotModelTrianglesFromDC = true; |
| 1534 |
} |
| 1535 |
|
| 1536 |
return true; |
| 1537 |
} |
| 1538 |
|
| 1539 |
|
| 1540 |
} |