t1 / TFDContents / Assets / KinectScripts / Interfaces / Kinect2Interface.cs @ 3
이력 | 보기 | 이력해설 | 다운로드 (52.8 KB)
1 | 3 | KTH | #if !(UNITY_WSA_10_0 && NETFX_CORE) |
---|---|---|---|
2 | using UnityEngine; |
||
3 | using System.Collections; |
||
4 | using Windows.Kinect; |
||
5 | using System.Runtime.InteropServices; |
||
6 | using Microsoft.Kinect.Face; |
||
7 | using System.Collections.Generic; |
||
8 | using System; |
||
9 | |||
10 | public class Kinect2Interface : DepthSensorInterface |
||
11 | { |
||
12 | // change this to false, if you aren't using Kinect-v2 only and want KM to check for available sensors |
||
13 | public static bool sensorAlwaysAvailable = true; |
||
14 | |||
15 | private KinectInterop.FrameSource sensorFlags; |
||
16 | public KinectSensor kinectSensor; |
||
17 | public CoordinateMapper coordMapper; |
||
18 | |||
19 | private BodyFrameReader bodyFrameReader; |
||
20 | private BodyIndexFrameReader bodyIndexFrameReader; |
||
21 | private ColorFrameReader colorFrameReader; |
||
22 | private DepthFrameReader depthFrameReader; |
||
23 | private InfraredFrameReader infraredFrameReader; |
||
24 | |||
25 | private MultiSourceFrameReader multiSourceFrameReader; |
||
26 | private MultiSourceFrame multiSourceFrame; |
||
27 | |||
28 | private BodyFrame msBodyFrame = null; |
||
29 | private BodyIndexFrame msBodyIndexFrame = null; |
||
30 | private ColorFrame msColorFrame = null; |
||
31 | private DepthFrame msDepthFrame = null; |
||
32 | private InfraredFrame msInfraredFrame = null; |
||
33 | |||
34 | private int bodyCount; |
||
35 | private Body[] bodyData; |
||
36 | |||
37 | private bool bFaceTrackingInited = false; |
||
38 | public FaceFrameSource[] faceFrameSources = null; |
||
39 | public FaceFrameReader[] faceFrameReaders = null; |
||
40 | public FaceFrameResult[] faceFrameResults = null; |
||
41 | |||
42 | // private int faceDisplayWidth; |
||
43 | // private int faceDisplayHeight; |
||
44 | |||
45 | private bool isDrawFaceRect = false; |
||
46 | public HighDefinitionFaceFrameSource[] hdFaceFrameSources = null; |
||
47 | public HighDefinitionFaceFrameReader[] hdFaceFrameReaders = null; |
||
48 | public FaceAlignment[] hdFaceAlignments = null; |
||
49 | public FaceModel[] hdFaceModels = null; |
||
50 | |||
51 | private bool bBackgroundRemovalInited = false; |
||
52 | |||
53 | |||
54 | // DLL Imports for speech wrapper functions |
||
55 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "InitSpeechRecognizer")] |
||
56 | private static extern int InitSpeechRecognizerNative([MarshalAs(UnmanagedType.LPWStr)]string sRecoCriteria, bool bUseKinect, bool bAdaptationOff); |
||
57 | |||
58 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "FinishSpeechRecognizer")] |
||
59 | private static extern void FinishSpeechRecognizerNative(); |
||
60 | |||
61 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "UpdateSpeechRecognizer")] |
||
62 | private static extern int UpdateSpeechRecognizerNative(); |
||
63 | |||
64 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "LoadSpeechGrammar")] |
||
65 | private static extern int LoadSpeechGrammarNative([MarshalAs(UnmanagedType.LPWStr)]string sFileName, short iNewLangCode, bool bDynamic); |
||
66 | |||
67 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "AddGrammarPhrase")] |
||
68 | private static extern int AddGrammarPhraseNative([MarshalAs(UnmanagedType.LPWStr)]string sFromRule, [MarshalAs(UnmanagedType.LPWStr)]string sToRule, [MarshalAs(UnmanagedType.LPWStr)]string sPhrase, bool bClearRule, bool bCommitGrammar); |
||
69 | |||
70 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "AddSpeechGrammar")] |
||
71 | private static extern int AddSpeechGrammarNative([MarshalAs(UnmanagedType.LPWStr)]string sFileName, short iNewLangCode, bool bDynamic); |
||
72 | |||
73 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "AddPhraseToGrammar")] |
||
74 | private static extern int AddPhraseToGrammarNative([MarshalAs(UnmanagedType.LPWStr)]string sGrammarName, [MarshalAs(UnmanagedType.LPWStr)]string sFromRule, [MarshalAs(UnmanagedType.LPWStr)]string sToRule, [MarshalAs(UnmanagedType.LPWStr)]string sPhrase, bool bClearRule, bool bCommitGrammar); |
||
75 | |||
76 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "SetGrammarState")] |
||
77 | private static extern int SetGrammarStateNative([MarshalAs(UnmanagedType.LPWStr)]string sGrammarName, bool bEnableGrammar); |
||
78 | |||
79 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "SetRequiredConfidence")] |
||
80 | private static extern void SetSpeechConfidenceNative(float fConfidence); |
||
81 | |||
82 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "IsSoundStarted")] |
||
83 | private static extern bool IsSpeechStartedNative(); |
||
84 | |||
85 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "IsSoundEnded")] |
||
86 | private static extern bool IsSpeechEndedNative(); |
||
87 | |||
88 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "IsPhraseRecognized")] |
||
89 | private static extern bool IsPhraseRecognizedNative(); |
||
90 | |||
91 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "GetPhraseConfidence")] |
||
92 | private static extern float GetPhraseConfidenceNative(); |
||
93 | |||
94 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "GetRecognizedTag")] |
||
95 | private static extern IntPtr GetRecognizedPhraseTagNative(); |
||
96 | |||
97 | [DllImport("Kinect2SpeechWrapper", EntryPoint = "ClearPhraseRecognized")] |
||
98 | private static extern void ClearRecognizedPhraseNative(); |
||
99 | |||
100 | |||
101 | public KinectInterop.DepthSensorPlatform GetSensorPlatform() |
||
102 | { |
||
103 | return KinectInterop.DepthSensorPlatform.KinectSDKv2; |
||
104 | } |
||
105 | |||
106 | public bool InitSensorInterface (bool bCopyLibs, ref bool bNeedRestart) |
||
107 | { |
||
108 | bool bOneCopied = false, bAllCopied = true; |
||
109 | string sTargetPath = KinectInterop.GetTargetDllPath(".", KinectInterop.Is64bitArchitecture()) + "/"; |
||
110 | |||
111 | if(!bCopyLibs) |
||
112 | { |
||
113 | // check if the native library is there |
||
114 | string sTargetLib = sTargetPath + "KinectUnityAddin.dll"; |
||
115 | bNeedRestart = false; |
||
116 | |||
117 | string sZipFileName = !KinectInterop.Is64bitArchitecture() ? "KinectV2UnityAddin.x86.zip" : "KinectV2UnityAddin.x64.zip"; |
||
118 | long iTargetSize = KinectInterop.GetUnzippedEntrySize(sZipFileName, "KinectUnityAddin.dll"); |
||
119 | |||
120 | // System.IO.FileInfo targetFile = new System.IO.FileInfo(sTargetLib); |
||
121 | // return targetFile.Exists && targetFile.Length == iTargetSize; |
||
122 | return KinectInterop.IsFileExists(sTargetLib, iTargetSize); |
||
123 | } |
||
124 | |||
125 | if(!KinectInterop.Is64bitArchitecture()) |
||
126 | { |
||
127 | Debug.Log("x32-architecture detected."); |
||
128 | |||
129 | //KinectInterop.CopyResourceFile(sTargetPath + "KinectUnityAddin.dll", "KinectUnityAddin.dll", ref bOneCopied, ref bAllCopied); |
||
130 | |||
131 | Dictionary<string, string> dictFilesToUnzip = new Dictionary<string, string>(); |
||
132 | dictFilesToUnzip["KinectUnityAddin.dll"] = sTargetPath + "KinectUnityAddin.dll"; |
||
133 | dictFilesToUnzip["Kinect20.Face.dll"] = sTargetPath + "Kinect20.Face.dll"; |
||
134 | dictFilesToUnzip["KinectFaceUnityAddin.dll"] = sTargetPath + "KinectFaceUnityAddin.dll"; |
||
135 | dictFilesToUnzip["Kinect2SpeechWrapper.dll"] = sTargetPath + "Kinect2SpeechWrapper.dll"; |
||
136 | dictFilesToUnzip["Kinect20.VisualGestureBuilder.dll"] = sTargetPath + "Kinect20.VisualGestureBuilder.dll"; |
||
137 | dictFilesToUnzip["KinectVisualGestureBuilderUnityAddin.dll"] = sTargetPath + "KinectVisualGestureBuilderUnityAddin.dll"; |
||
138 | dictFilesToUnzip["vgbtechs/AdaBoostTech.dll"] = sTargetPath + "vgbtechs/AdaBoostTech.dll"; |
||
139 | dictFilesToUnzip["vgbtechs/RFRProgressTech.dll"] = sTargetPath + "vgbtechs/RFRProgressTech.dll"; |
||
140 | dictFilesToUnzip["msvcp110.dll"] = sTargetPath + "msvcp110.dll"; |
||
141 | dictFilesToUnzip["msvcr110.dll"] = sTargetPath + "msvcr110.dll"; |
||
142 | |||
143 | KinectInterop.UnzipResourceFiles(dictFilesToUnzip, "KinectV2UnityAddin.x86.zip", ref bOneCopied, ref bAllCopied); |
||
144 | } |
||
145 | else |
||
146 | { |
||
147 | Debug.Log("x64-architecture detected."); |
||
148 | |||
149 | //KinectInterop.CopyResourceFile(sTargetPath + "KinectUnityAddin.dll", "KinectUnityAddin.dll.x64", ref bOneCopied, ref bAllCopied); |
||
150 | |||
151 | Dictionary<string, string> dictFilesToUnzip = new Dictionary<string, string>(); |
||
152 | dictFilesToUnzip["KinectUnityAddin.dll"] = sTargetPath + "KinectUnityAddin.dll"; |
||
153 | dictFilesToUnzip["Kinect20.Face.dll"] = sTargetPath + "Kinect20.Face.dll"; |
||
154 | dictFilesToUnzip["KinectFaceUnityAddin.dll"] = sTargetPath + "KinectFaceUnityAddin.dll"; |
||
155 | dictFilesToUnzip["Kinect2SpeechWrapper.dll"] = sTargetPath + "Kinect2SpeechWrapper.dll"; |
||
156 | dictFilesToUnzip["Kinect20.VisualGestureBuilder.dll"] = sTargetPath + "Kinect20.VisualGestureBuilder.dll"; |
||
157 | dictFilesToUnzip["KinectVisualGestureBuilderUnityAddin.dll"] = sTargetPath + "KinectVisualGestureBuilderUnityAddin.dll"; |
||
158 | dictFilesToUnzip["vgbtechs/AdaBoostTech.dll"] = sTargetPath + "vgbtechs/AdaBoostTech.dll"; |
||
159 | dictFilesToUnzip["vgbtechs/RFRProgressTech.dll"] = sTargetPath + "vgbtechs/RFRProgressTech.dll"; |
||
160 | dictFilesToUnzip["msvcp110.dll"] = sTargetPath + "msvcp110.dll"; |
||
161 | dictFilesToUnzip["msvcr110.dll"] = sTargetPath + "msvcr110.dll"; |
||
162 | |||
163 | KinectInterop.UnzipResourceFiles(dictFilesToUnzip, "KinectV2UnityAddin.x64.zip", ref bOneCopied, ref bAllCopied); |
||
164 | } |
||
165 | |||
166 | KinectInterop.UnzipResourceDirectory(sTargetPath, "NuiDatabase.zip", sTargetPath + "NuiDatabase"); |
||
167 | |||
168 | bNeedRestart = (bOneCopied && bAllCopied); |
||
169 | |||
170 | return true; |
||
171 | } |
||
172 | |||
173 | public void FreeSensorInterface (bool bDeleteLibs) |
||
174 | { |
||
175 | if(bDeleteLibs) |
||
176 | { |
||
177 | KinectInterop.DeleteNativeLib("KinectUnityAddin.dll", true); |
||
178 | KinectInterop.DeleteNativeLib("msvcp110.dll", false); |
||
179 | KinectInterop.DeleteNativeLib("msvcr110.dll", false); |
||
180 | } |
||
181 | } |
||
182 | |||
183 | public bool IsSensorAvailable() |
||
184 | { |
||
185 | KinectSensor sensor = KinectSensor.GetDefault(); |
||
186 | |||
187 | if(sensor != null) |
||
188 | { |
||
189 | if(sensorAlwaysAvailable) |
||
190 | { |
||
191 | sensor = null; |
||
192 | return true; |
||
193 | } |
||
194 | |||
195 | if(!sensor.IsOpen) |
||
196 | { |
||
197 | sensor.Open(); |
||
198 | } |
||
199 | |||
200 | float fWaitTime = Time.realtimeSinceStartup + 3f; |
||
201 | while(!sensor.IsAvailable && Time.realtimeSinceStartup < fWaitTime) |
||
202 | { |
||
203 | // wait for availability |
||
204 | } |
||
205 | |||
206 | bool bAvailable = sensor.IsAvailable; |
||
207 | |||
208 | if(sensor.IsOpen) |
||
209 | { |
||
210 | sensor.Close(); |
||
211 | } |
||
212 | |||
213 | fWaitTime = Time.realtimeSinceStartup + 3f; |
||
214 | while(sensor.IsOpen && Time.realtimeSinceStartup < fWaitTime) |
||
215 | { |
||
216 | // wait for sensor to close |
||
217 | } |
||
218 | |||
219 | sensor = null; |
||
220 | return bAvailable; |
||
221 | } |
||
222 | |||
223 | return false; |
||
224 | } |
||
225 | |||
226 | public int GetSensorsCount() |
||
227 | { |
||
228 | int numSensors = 0; |
||
229 | |||
230 | KinectSensor sensor = KinectSensor.GetDefault(); |
||
231 | if(sensor != null) |
||
232 | { |
||
233 | if(!sensor.IsOpen) |
||
234 | { |
||
235 | sensor.Open(); |
||
236 | } |
||
237 | |||
238 | float fWaitTime = Time.realtimeSinceStartup + 3f; |
||
239 | while(!sensor.IsAvailable && Time.realtimeSinceStartup < fWaitTime) |
||
240 | { |
||
241 | // wait for availability |
||
242 | } |
||
243 | |||
244 | numSensors = sensor.IsAvailable ? 1 : 0; |
||
245 | |||
246 | if(sensor.IsOpen) |
||
247 | { |
||
248 | sensor.Close(); |
||
249 | } |
||
250 | |||
251 | fWaitTime = Time.realtimeSinceStartup + 3f; |
||
252 | while(sensor.IsOpen && Time.realtimeSinceStartup < fWaitTime) |
||
253 | { |
||
254 | // wait for sensor to close |
||
255 | } |
||
256 | } |
||
257 | |||
258 | return numSensors; |
||
259 | } |
||
260 | |||
261 | public KinectInterop.SensorData OpenDefaultSensor (KinectInterop.FrameSource dwFlags, float sensorAngle, bool bUseMultiSource) |
||
262 | { |
||
263 | KinectInterop.SensorData sensorData = new KinectInterop.SensorData(); |
||
264 | sensorFlags = dwFlags; |
||
265 | |||
266 | kinectSensor = KinectSensor.GetDefault(); |
||
267 | if(kinectSensor == null) |
||
268 | return null; |
||
269 | |||
270 | coordMapper = kinectSensor.CoordinateMapper; |
||
271 | |||
272 | this.bodyCount = kinectSensor.BodyFrameSource.BodyCount; |
||
273 | sensorData.bodyCount = this.bodyCount; |
||
274 | sensorData.jointCount = 25; |
||
275 | |||
276 | sensorData.depthCameraFOV = 60f; |
||
277 | sensorData.colorCameraFOV = 53.8f; |
||
278 | sensorData.depthCameraOffset = -0.05f; |
||
279 | sensorData.faceOverlayOffset = -0.04f; |
||
280 | |||
281 | if((dwFlags & KinectInterop.FrameSource.TypeBody) != 0) |
||
282 | { |
||
283 | if(!bUseMultiSource) |
||
284 | bodyFrameReader = kinectSensor.BodyFrameSource.OpenReader(); |
||
285 | |||
286 | bodyData = new Body[sensorData.bodyCount]; |
||
287 | } |
||
288 | |||
289 | var frameDesc = kinectSensor.ColorFrameSource.CreateFrameDescription(ColorImageFormat.Rgba); |
||
290 | sensorData.colorImageWidth = frameDesc.Width; |
||
291 | sensorData.colorImageHeight = frameDesc.Height; |
||
292 | |||
293 | if((dwFlags & KinectInterop.FrameSource.TypeColor) != 0) |
||
294 | { |
||
295 | if(!bUseMultiSource) |
||
296 | colorFrameReader = kinectSensor.ColorFrameSource.OpenReader(); |
||
297 | |||
298 | sensorData.colorImage = new byte[frameDesc.BytesPerPixel * frameDesc.LengthInPixels]; |
||
299 | } |
||
300 | |||
301 | sensorData.depthImageWidth = kinectSensor.DepthFrameSource.FrameDescription.Width; |
||
302 | sensorData.depthImageHeight = kinectSensor.DepthFrameSource.FrameDescription.Height; |
||
303 | |||
304 | if((dwFlags & KinectInterop.FrameSource.TypeDepth) != 0) |
||
305 | { |
||
306 | if(!bUseMultiSource) |
||
307 | depthFrameReader = kinectSensor.DepthFrameSource.OpenReader(); |
||
308 | |||
309 | sensorData.depthImage = new ushort[kinectSensor.DepthFrameSource.FrameDescription.LengthInPixels]; |
||
310 | } |
||
311 | |||
312 | if((dwFlags & KinectInterop.FrameSource.TypeBodyIndex) != 0) |
||
313 | { |
||
314 | if(!bUseMultiSource) |
||
315 | bodyIndexFrameReader = kinectSensor.BodyIndexFrameSource.OpenReader(); |
||
316 | |||
317 | sensorData.bodyIndexImage = new byte[kinectSensor.BodyIndexFrameSource.FrameDescription.LengthInPixels]; |
||
318 | } |
||
319 | |||
320 | if((dwFlags & KinectInterop.FrameSource.TypeInfrared) != 0) |
||
321 | { |
||
322 | if(!bUseMultiSource) |
||
323 | infraredFrameReader = kinectSensor.InfraredFrameSource.OpenReader(); |
||
324 | |||
325 | sensorData.infraredImage = new ushort[kinectSensor.InfraredFrameSource.FrameDescription.LengthInPixels]; |
||
326 | } |
||
327 | |||
328 | //if(!kinectSensor.IsOpen) |
||
329 | { |
||
330 | //Debug.Log("Opening sensor, available: " + kinectSensor.IsAvailable); |
||
331 | kinectSensor.Open(); |
||
332 | } |
||
333 | |||
334 | float fWaitTime = Time.realtimeSinceStartup + 3f; |
||
335 | while(!kinectSensor.IsAvailable && Time.realtimeSinceStartup < fWaitTime) |
||
336 | { |
||
337 | // wait for sensor to open |
||
338 | } |
||
339 | |||
340 | Debug.Log("K2-sensor " + (kinectSensor.IsOpen ? "opened" : "closed") + |
||
341 | ", available: " + kinectSensor.IsAvailable); |
||
342 | |||
343 | if(bUseMultiSource && dwFlags != KinectInterop.FrameSource.TypeNone && kinectSensor.IsOpen) |
||
344 | { |
||
345 | multiSourceFrameReader = kinectSensor.OpenMultiSourceFrameReader((FrameSourceTypes)((int)dwFlags & 0x3F)); |
||
346 | } |
||
347 | |||
348 | return sensorData; |
||
349 | } |
||
350 | |||
351 | public void CloseSensor (KinectInterop.SensorData sensorData) |
||
352 | { |
||
353 | if(coordMapper != null) |
||
354 | { |
||
355 | coordMapper = null; |
||
356 | } |
||
357 | |||
358 | if(bodyFrameReader != null) |
||
359 | { |
||
360 | bodyFrameReader.Dispose(); |
||
361 | bodyFrameReader = null; |
||
362 | } |
||
363 | |||
364 | if(bodyIndexFrameReader != null) |
||
365 | { |
||
366 | bodyIndexFrameReader.Dispose(); |
||
367 | bodyIndexFrameReader = null; |
||
368 | } |
||
369 | |||
370 | if(colorFrameReader != null) |
||
371 | { |
||
372 | colorFrameReader.Dispose(); |
||
373 | colorFrameReader = null; |
||
374 | } |
||
375 | |||
376 | if(depthFrameReader != null) |
||
377 | { |
||
378 | depthFrameReader.Dispose(); |
||
379 | depthFrameReader = null; |
||
380 | } |
||
381 | |||
382 | if(infraredFrameReader != null) |
||
383 | { |
||
384 | infraredFrameReader.Dispose(); |
||
385 | infraredFrameReader = null; |
||
386 | } |
||
387 | |||
388 | if(multiSourceFrameReader != null) |
||
389 | { |
||
390 | multiSourceFrameReader.Dispose(); |
||
391 | multiSourceFrameReader = null; |
||
392 | } |
||
393 | |||
394 | if(kinectSensor != null) |
||
395 | { |
||
396 | //if (kinectSensor.IsOpen) |
||
397 | { |
||
398 | //Debug.Log("Closing sensor, available: " + kinectSensor.IsAvailable); |
||
399 | kinectSensor.Close(); |
||
400 | } |
||
401 | |||
402 | float fWaitTime = Time.realtimeSinceStartup + 3f; |
||
403 | while(kinectSensor.IsOpen && Time.realtimeSinceStartup < fWaitTime) |
||
404 | { |
||
405 | // wait for sensor to close |
||
406 | } |
||
407 | |||
408 | Debug.Log("K2-sensor " + (kinectSensor.IsOpen ? "opened" : "closed") + |
||
409 | ", available: " + kinectSensor.IsAvailable); |
||
410 | |||
411 | kinectSensor = null; |
||
412 | } |
||
413 | } |
||
414 | |||
415 | public bool UpdateSensorData (KinectInterop.SensorData sensorData) |
||
416 | { |
||
417 | return true; |
||
418 | } |
||
419 | |||
420 | public bool GetMultiSourceFrame (KinectInterop.SensorData sensorData) |
||
421 | { |
||
422 | if(multiSourceFrameReader != null) |
||
423 | { |
||
424 | multiSourceFrame = multiSourceFrameReader.AcquireLatestFrame(); |
||
425 | |||
426 | if(multiSourceFrame != null) |
||
427 | { |
||
428 | // try to get all frames at once |
||
429 | msBodyFrame = (sensorFlags & KinectInterop.FrameSource.TypeBody) != 0 ? multiSourceFrame.BodyFrameReference.AcquireFrame() : null; |
||
430 | msBodyIndexFrame = (sensorFlags & KinectInterop.FrameSource.TypeBodyIndex) != 0 ? multiSourceFrame.BodyIndexFrameReference.AcquireFrame() : null; |
||
431 | msColorFrame = (sensorFlags & KinectInterop.FrameSource.TypeColor) != 0 ? multiSourceFrame.ColorFrameReference.AcquireFrame() : null; |
||
432 | msDepthFrame = (sensorFlags & KinectInterop.FrameSource.TypeDepth) != 0 ? multiSourceFrame.DepthFrameReference.AcquireFrame() : null; |
||
433 | msInfraredFrame = (sensorFlags & KinectInterop.FrameSource.TypeInfrared) != 0 ? multiSourceFrame.InfraredFrameReference.AcquireFrame() : null; |
||
434 | |||
435 | bool bAllSet = |
||
436 | ((sensorFlags & KinectInterop.FrameSource.TypeBody) == 0 || msBodyFrame != null) && |
||
437 | ((sensorFlags & KinectInterop.FrameSource.TypeBodyIndex) == 0 || msBodyIndexFrame != null) && |
||
438 | ((sensorFlags & KinectInterop.FrameSource.TypeColor) == 0 || msColorFrame != null) && |
||
439 | ((sensorFlags & KinectInterop.FrameSource.TypeDepth) == 0 || msDepthFrame != null) && |
||
440 | ((sensorFlags & KinectInterop.FrameSource.TypeInfrared) == 0 || msInfraredFrame != null); |
||
441 | |||
442 | if(!bAllSet) |
||
443 | { |
||
444 | // release all frames |
||
445 | if(msBodyFrame != null) |
||
446 | { |
||
447 | msBodyFrame.Dispose(); |
||
448 | msBodyFrame = null; |
||
449 | } |
||
450 | |||
451 | if(msBodyIndexFrame != null) |
||
452 | { |
||
453 | msBodyIndexFrame.Dispose(); |
||
454 | msBodyIndexFrame = null; |
||
455 | } |
||
456 | |||
457 | if(msColorFrame != null) |
||
458 | { |
||
459 | msColorFrame.Dispose(); |
||
460 | msColorFrame = null; |
||
461 | } |
||
462 | |||
463 | if(msDepthFrame != null) |
||
464 | { |
||
465 | msDepthFrame.Dispose(); |
||
466 | msDepthFrame = null; |
||
467 | } |
||
468 | |||
469 | if(msInfraredFrame != null) |
||
470 | { |
||
471 | msInfraredFrame.Dispose(); |
||
472 | msInfraredFrame = null; |
||
473 | } |
||
474 | } |
||
475 | // else |
||
476 | // { |
||
477 | // bool bNeedBody = (sensorFlags & KinectInterop.FrameSource.TypeBody) != 0; |
||
478 | // bool bNeedBodyIndex = (sensorFlags & KinectInterop.FrameSource.TypeBodyIndex) != 0; |
||
479 | // bool bNeedColor = (sensorFlags & KinectInterop.FrameSource.TypeColor) != 0; |
||
480 | // bool bNeedDepth = (sensorFlags & KinectInterop.FrameSource.TypeDepth) != 0; |
||
481 | // bool bNeedInfrared = (sensorFlags & KinectInterop.FrameSource.TypeInfrared) != 0; |
||
482 | // |
||
483 | // bAllSet = true; |
||
484 | // } |
||
485 | } |
||
486 | |||
487 | return (multiSourceFrame != null); |
||
488 | } |
||
489 | |||
490 | return false; |
||
491 | } |
||
492 | |||
493 | public void FreeMultiSourceFrame (KinectInterop.SensorData sensorData) |
||
494 | { |
||
495 | // release all frames |
||
496 | if(msBodyFrame != null) |
||
497 | { |
||
498 | msBodyFrame.Dispose(); |
||
499 | msBodyFrame = null; |
||
500 | } |
||
501 | |||
502 | if(msBodyIndexFrame != null) |
||
503 | { |
||
504 | msBodyIndexFrame.Dispose(); |
||
505 | msBodyIndexFrame = null; |
||
506 | } |
||
507 | |||
508 | if(msColorFrame != null) |
||
509 | { |
||
510 | msColorFrame.Dispose(); |
||
511 | msColorFrame = null; |
||
512 | } |
||
513 | |||
514 | if(msDepthFrame != null) |
||
515 | { |
||
516 | msDepthFrame.Dispose(); |
||
517 | msDepthFrame = null; |
||
518 | } |
||
519 | |||
520 | if(msInfraredFrame != null) |
||
521 | { |
||
522 | msInfraredFrame.Dispose(); |
||
523 | msInfraredFrame = null; |
||
524 | } |
||
525 | |||
526 | if(multiSourceFrame != null) |
||
527 | { |
||
528 | multiSourceFrame = null; |
||
529 | } |
||
530 | } |
||
531 | |||
532 | public bool PollBodyFrame (KinectInterop.SensorData sensorData, ref KinectInterop.BodyFrameData bodyFrame, |
||
533 | ref Matrix4x4 kinectToWorld, bool bIgnoreJointZ) |
||
534 | { |
||
535 | bool bNewFrame = false; |
||
536 | |||
537 | if((multiSourceFrameReader != null && multiSourceFrame != null) || |
||
538 | bodyFrameReader != null) |
||
539 | { |
||
540 | BodyFrame frame = multiSourceFrame != null ? msBodyFrame : |
||
541 | bodyFrameReader.AcquireLatestFrame(); |
||
542 | |||
543 | if(frame != null) |
||
544 | { |
||
545 | frame.GetAndRefreshBodyData(bodyData); |
||
546 | |||
547 | bodyFrame.liPreviousTime = bodyFrame.liRelativeTime; |
||
548 | bodyFrame.liRelativeTime = frame.RelativeTime.Ticks; |
||
549 | |||
550 | if(sensorData.hintHeightAngle) |
||
551 | { |
||
552 | // get the floor plane |
||
553 | Windows.Kinect.Vector4 vFloorPlane = frame.FloorClipPlane; |
||
554 | Vector3 floorPlane = new Vector3(vFloorPlane.X, vFloorPlane.Y, vFloorPlane.Z); |
||
555 | |||
556 | sensorData.sensorRotDetected = Quaternion.FromToRotation(floorPlane, Vector3.up); |
||
557 | sensorData.sensorHgtDetected = vFloorPlane.W; |
||
558 | } |
||
559 | |||
560 | frame.Dispose(); |
||
561 | frame = null; |
||
562 | |||
563 | for(int i = 0; i < sensorData.bodyCount; i++) |
||
564 | { |
||
565 | Body body = bodyData[i]; |
||
566 | |||
567 | if (body == null) |
||
568 | { |
||
569 | bodyFrame.bodyData[i].bIsTracked = 0; |
||
570 | continue; |
||
571 | } |
||
572 | |||
573 | bodyFrame.bodyData[i].bIsTracked = (short)(body.IsTracked ? 1 : 0); |
||
574 | |||
575 | if(body.IsTracked) |
||
576 | { |
||
577 | // transfer body and joints data |
||
578 | bodyFrame.bodyData[i].liTrackingID = (long)body.TrackingId; |
||
579 | |||
580 | // cache the body joints (following the advice of Brian Chasalow) |
||
581 | Dictionary<Windows.Kinect.JointType, Windows.Kinect.Joint> bodyJoints = body.Joints; |
||
582 | |||
583 | for(int j = 0; j < sensorData.jointCount; j++) |
||
584 | { |
||
585 | Windows.Kinect.Joint joint = bodyJoints[(Windows.Kinect.JointType)j]; |
||
586 | KinectInterop.JointData jointData = bodyFrame.bodyData[i].joint[j]; |
||
587 | |||
588 | //jointData.jointType = (KinectInterop.JointType)j; |
||
589 | jointData.trackingState = (KinectInterop.TrackingState)joint.TrackingState; |
||
590 | |||
591 | if((int)joint.TrackingState != (int)TrackingState.NotTracked) |
||
592 | { |
||
593 | float jPosZ = (bIgnoreJointZ && j > 0) ? bodyFrame.bodyData[i].joint[0].kinectPos.z : joint.Position.Z; |
||
594 | jointData.kinectPos = new Vector3(joint.Position.X, joint.Position.Y, joint.Position.Z); |
||
595 | jointData.position = kinectToWorld.MultiplyPoint3x4(new Vector3(joint.Position.X, joint.Position.Y, jPosZ)); |
||
596 | } |
||
597 | |||
598 | jointData.orientation = Quaternion.identity; |
||
599 | // Windows.Kinect.Vector4 vQ = body.JointOrientations[jointData.jointType].Orientation; |
||
600 | // jointData.orientation = new Quaternion(vQ.X, vQ.Y, vQ.Z, vQ.W); |
||
601 | |||
602 | if(j == 0) |
||
603 | { |
||
604 | bodyFrame.bodyData[i].position = jointData.position; |
||
605 | bodyFrame.bodyData[i].orientation = jointData.orientation; |
||
606 | } |
||
607 | |||
608 | bodyFrame.bodyData[i].joint[j] = jointData; |
||
609 | } |
||
610 | |||
611 | // tranfer hand states |
||
612 | bodyFrame.bodyData[i].leftHandState = (KinectInterop.HandState)body.HandLeftState; |
||
613 | bodyFrame.bodyData[i].leftHandConfidence = (KinectInterop.TrackingConfidence)body.HandLeftConfidence; |
||
614 | |||
615 | bodyFrame.bodyData[i].rightHandState = (KinectInterop.HandState)body.HandRightState; |
||
616 | bodyFrame.bodyData[i].rightHandConfidence = (KinectInterop.TrackingConfidence)body.HandRightConfidence; |
||
617 | } |
||
618 | } |
||
619 | |||
620 | bNewFrame = true; |
||
621 | } |
||
622 | } |
||
623 | |||
624 | return bNewFrame; |
||
625 | } |
||
626 | |||
627 | public bool PollColorFrame (KinectInterop.SensorData sensorData) |
||
628 | { |
||
629 | bool bNewFrame = false; |
||
630 | |||
631 | if((multiSourceFrameReader != null && multiSourceFrame != null) || |
||
632 | colorFrameReader != null) |
||
633 | { |
||
634 | ColorFrame colorFrame = multiSourceFrame != null ? msColorFrame : |
||
635 | colorFrameReader.AcquireLatestFrame(); |
||
636 | |||
637 | if(colorFrame != null) |
||
638 | { |
||
639 | var pColorData = GCHandle.Alloc(sensorData.colorImage, GCHandleType.Pinned); |
||
640 | colorFrame.CopyConvertedFrameDataToIntPtr(pColorData.AddrOfPinnedObject(), (uint)sensorData.colorImage.Length, ColorImageFormat.Rgba); |
||
641 | pColorData.Free(); |
||
642 | |||
643 | sensorData.lastColorFrameTime = colorFrame.RelativeTime.Ticks; |
||
644 | |||
645 | colorFrame.Dispose(); |
||
646 | colorFrame = null; |
||
647 | |||
648 | bNewFrame = true; |
||
649 | } |
||
650 | } |
||
651 | |||
652 | return bNewFrame; |
||
653 | } |
||
654 | |||
655 | public bool PollDepthFrame (KinectInterop.SensorData sensorData) |
||
656 | { |
||
657 | bool bNewFrame = false; |
||
658 | |||
659 | if((multiSourceFrameReader != null && multiSourceFrame != null) || |
||
660 | depthFrameReader != null) |
||
661 | { |
||
662 | DepthFrame depthFrame = multiSourceFrame != null ? msDepthFrame : |
||
663 | depthFrameReader.AcquireLatestFrame(); |
||
664 | |||
665 | if(depthFrame != null) |
||
666 | { |
||
667 | var pDepthData = GCHandle.Alloc(sensorData.depthImage, GCHandleType.Pinned); |
||
668 | depthFrame.CopyFrameDataToIntPtr(pDepthData.AddrOfPinnedObject(), (uint)sensorData.depthImage.Length * sizeof(ushort)); |
||
669 | pDepthData.Free(); |
||
670 | |||
671 | sensorData.lastDepthFrameTime = depthFrame.RelativeTime.Ticks; |
||
672 | |||
673 | depthFrame.Dispose(); |
||
674 | depthFrame = null; |
||
675 | |||
676 | bNewFrame = true; |
||
677 | } |
||
678 | |||
679 | if((multiSourceFrameReader != null && multiSourceFrame != null) || |
||
680 | bodyIndexFrameReader != null) |
||
681 | { |
||
682 | BodyIndexFrame bodyIndexFrame = multiSourceFrame != null ? msBodyIndexFrame : |
||
683 | bodyIndexFrameReader.AcquireLatestFrame(); |
||
684 | |||
685 | if(bodyIndexFrame != null) |
||
686 | { |
||
687 | var pBodyIndexData = GCHandle.Alloc(sensorData.bodyIndexImage, GCHandleType.Pinned); |
||
688 | bodyIndexFrame.CopyFrameDataToIntPtr(pBodyIndexData.AddrOfPinnedObject(), (uint)sensorData.bodyIndexImage.Length); |
||
689 | pBodyIndexData.Free(); |
||
690 | |||
691 | sensorData.lastBodyIndexFrameTime = bodyIndexFrame.RelativeTime.Ticks; |
||
692 | |||
693 | bodyIndexFrame.Dispose(); |
||
694 | bodyIndexFrame = null; |
||
695 | |||
696 | bNewFrame = true; |
||
697 | } |
||
698 | } |
||
699 | } |
||
700 | |||
701 | return bNewFrame; |
||
702 | } |
||
703 | |||
704 | public bool PollInfraredFrame (KinectInterop.SensorData sensorData) |
||
705 | { |
||
706 | bool bNewFrame = false; |
||
707 | |||
708 | if((multiSourceFrameReader != null && multiSourceFrame != null) || |
||
709 | infraredFrameReader != null) |
||
710 | { |
||
711 | InfraredFrame infraredFrame = multiSourceFrame != null ? msInfraredFrame : |
||
712 | infraredFrameReader.AcquireLatestFrame(); |
||
713 | |||
714 | if(infraredFrame != null) |
||
715 | { |
||
716 | var pInfraredData = GCHandle.Alloc(sensorData.infraredImage, GCHandleType.Pinned); |
||
717 | infraredFrame.CopyFrameDataToIntPtr(pInfraredData.AddrOfPinnedObject(), (uint)sensorData.infraredImage.Length * sizeof(ushort)); |
||
718 | pInfraredData.Free(); |
||
719 | |||
720 | sensorData.lastInfraredFrameTime = infraredFrame.RelativeTime.Ticks; |
||
721 | |||
722 | infraredFrame.Dispose(); |
||
723 | infraredFrame = null; |
||
724 | |||
725 | bNewFrame = true; |
||
726 | } |
||
727 | } |
||
728 | |||
729 | return bNewFrame; |
||
730 | } |
||
731 | |||
732 | public void FixJointOrientations(KinectInterop.SensorData sensorData, ref KinectInterop.BodyData bodyData) |
||
733 | { |
||
734 | // no fixes are needed |
||
735 | } |
||
736 | |||
737 | public bool IsBodyTurned(ref KinectInterop.BodyData bodyData) |
||
738 | { |
||
739 | //face = On: Face (357.0/1.0) |
||
740 | //face = Off |
||
741 | //| Head_px <= -0.02 |
||
742 | //| | Neck_dx <= 0.08: Face (46.0/1.0) |
||
743 | //| | Neck_dx > 0.08: Back (3.0) |
||
744 | //| Head_px > -0.02 |
||
745 | //| | SpineShoulder_px <= -0.02: Face (4.0) |
||
746 | //| | SpineShoulder_px > -0.02: Back (64.0/1.0) |
||
747 | |||
748 | bool bBodyTurned = false; |
||
749 | |||
750 | if(bFaceTrackingInited) |
||
751 | { |
||
752 | bool bFaceOn = IsFaceTracked(bodyData.liTrackingID); |
||
753 | |||
754 | if(bFaceOn) |
||
755 | { |
||
756 | bBodyTurned = false; |
||
757 | } |
||
758 | else |
||
759 | { |
||
760 | // face = Off |
||
761 | if(bodyData.joint[(int)KinectInterop.JointType.Head].posRel.x <= -0.02f) |
||
762 | { |
||
763 | bBodyTurned = (bodyData.joint[(int)KinectInterop.JointType.Neck].posVel.x > 0.08f); |
||
764 | } |
||
765 | else |
||
766 | { |
||
767 | // Head_px > -0.02 |
||
768 | bBodyTurned = (bodyData.joint[(int)KinectInterop.JointType.SpineShoulder].posRel.x > -0.02f); |
||
769 | } |
||
770 | } |
||
771 | } |
||
772 | |||
773 | return bBodyTurned; |
||
774 | } |
||
775 | |||
776 | public Vector2 MapSpacePointToDepthCoords (KinectInterop.SensorData sensorData, Vector3 spacePos) |
||
777 | { |
||
778 | Vector2 vPoint = Vector2.zero; |
||
779 | |||
780 | if(coordMapper != null) |
||
781 | { |
||
782 | CameraSpacePoint camPoint = new CameraSpacePoint(); |
||
783 | camPoint.X = spacePos.x; |
||
784 | camPoint.Y = spacePos.y; |
||
785 | camPoint.Z = spacePos.z; |
||
786 | |||
787 | CameraSpacePoint[] camPoints = new CameraSpacePoint[1]; |
||
788 | camPoints[0] = camPoint; |
||
789 | |||
790 | DepthSpacePoint[] depthPoints = new DepthSpacePoint[1]; |
||
791 | coordMapper.MapCameraPointsToDepthSpace(camPoints, depthPoints); |
||
792 | |||
793 | DepthSpacePoint depthPoint = depthPoints[0]; |
||
794 | |||
795 | if(depthPoint.X >= 0 && depthPoint.X < sensorData.depthImageWidth && |
||
796 | depthPoint.Y >= 0 && depthPoint.Y < sensorData.depthImageHeight) |
||
797 | { |
||
798 | vPoint.x = depthPoint.X; |
||
799 | vPoint.y = depthPoint.Y; |
||
800 | } |
||
801 | } |
||
802 | |||
803 | return vPoint; |
||
804 | } |
||
805 | |||
806 | public Vector3 MapDepthPointToSpaceCoords (KinectInterop.SensorData sensorData, Vector2 depthPos, ushort depthVal) |
||
807 | { |
||
808 | Vector3 vPoint = Vector3.zero; |
||
809 | |||
810 | if(coordMapper != null && depthPos != Vector2.zero) |
||
811 | { |
||
812 | DepthSpacePoint depthPoint = new DepthSpacePoint(); |
||
813 | depthPoint.X = depthPos.x; |
||
814 | depthPoint.Y = depthPos.y; |
||
815 | |||
816 | DepthSpacePoint[] depthPoints = new DepthSpacePoint[1]; |
||
817 | depthPoints[0] = depthPoint; |
||
818 | |||
819 | ushort[] depthVals = new ushort[1]; |
||
820 | depthVals[0] = depthVal; |
||
821 | |||
822 | CameraSpacePoint[] camPoints = new CameraSpacePoint[1]; |
||
823 | coordMapper.MapDepthPointsToCameraSpace(depthPoints, depthVals, camPoints); |
||
824 | |||
825 | CameraSpacePoint camPoint = camPoints[0]; |
||
826 | vPoint.x = camPoint.X; |
||
827 | vPoint.y = camPoint.Y; |
||
828 | vPoint.z = camPoint.Z; |
||
829 | } |
||
830 | |||
831 | return vPoint; |
||
832 | } |
||
833 | |||
834 | public bool MapDepthFrameToSpaceCoords (KinectInterop.SensorData sensorData, ref Vector3[] vSpaceCoords) |
||
835 | { |
||
836 | if(coordMapper != null && sensorData.depthImage != null) |
||
837 | { |
||
838 | var pDepthData = GCHandle.Alloc(sensorData.depthImage, GCHandleType.Pinned); |
||
839 | var pSpaceCoordsData = GCHandle.Alloc(vSpaceCoords, GCHandleType.Pinned); |
||
840 | |||
841 | coordMapper.MapDepthFrameToCameraSpaceUsingIntPtr( |
||
842 | pDepthData.AddrOfPinnedObject(), |
||
843 | sensorData.depthImage.Length * sizeof(ushort), |
||
844 | pSpaceCoordsData.AddrOfPinnedObject(), |
||
845 | (uint)vSpaceCoords.Length); |
||
846 | |||
847 | pSpaceCoordsData.Free(); |
||
848 | pDepthData.Free(); |
||
849 | |||
850 | return true; |
||
851 | } |
||
852 | |||
853 | return false; |
||
854 | } |
||
855 | |||
856 | public Vector2 MapDepthPointToColorCoords (KinectInterop.SensorData sensorData, Vector2 depthPos, ushort depthVal) |
||
857 | { |
||
858 | Vector2 vPoint = Vector2.zero; |
||
859 | |||
860 | if(coordMapper != null && depthPos != Vector2.zero) |
||
861 | { |
||
862 | DepthSpacePoint depthPoint = new DepthSpacePoint(); |
||
863 | depthPoint.X = depthPos.x; |
||
864 | depthPoint.Y = depthPos.y; |
||
865 | |||
866 | DepthSpacePoint[] depthPoints = new DepthSpacePoint[1]; |
||
867 | depthPoints[0] = depthPoint; |
||
868 | |||
869 | ushort[] depthVals = new ushort[1]; |
||
870 | depthVals[0] = depthVal; |
||
871 | |||
872 | ColorSpacePoint[] colPoints = new ColorSpacePoint[1]; |
||
873 | coordMapper.MapDepthPointsToColorSpace(depthPoints, depthVals, colPoints); |
||
874 | |||
875 | ColorSpacePoint colPoint = colPoints[0]; |
||
876 | vPoint.x = colPoint.X; |
||
877 | vPoint.y = colPoint.Y; |
||
878 | } |
||
879 | |||
880 | return vPoint; |
||
881 | } |
||
882 | |||
883 | public bool MapDepthFrameToColorCoords (KinectInterop.SensorData sensorData, ref Vector2[] vColorCoords) |
||
884 | { |
||
885 | if(coordMapper != null && sensorData.colorImage != null && sensorData.depthImage != null) |
||
886 | { |
||
887 | var pDepthData = GCHandle.Alloc(sensorData.depthImage, GCHandleType.Pinned); |
||
888 | var pColorCoordsData = GCHandle.Alloc(vColorCoords, GCHandleType.Pinned); |
||
889 | |||
890 | coordMapper.MapDepthFrameToColorSpaceUsingIntPtr( |
||
891 | pDepthData.AddrOfPinnedObject(), |
||
892 | sensorData.depthImage.Length * sizeof(ushort), |
||
893 | pColorCoordsData.AddrOfPinnedObject(), |
||
894 | (uint)vColorCoords.Length); |
||
895 | |||
896 | pColorCoordsData.Free(); |
||
897 | pDepthData.Free(); |
||
898 | |||
899 | return true; |
||
900 | } |
||
901 | |||
902 | return false; |
||
903 | } |
||
904 | |||
905 | public bool MapColorFrameToDepthCoords (KinectInterop.SensorData sensorData, ref Vector2[] vDepthCoords) |
||
906 | { |
||
907 | if(coordMapper != null && sensorData.colorImage != null && sensorData.depthImage != null) |
||
908 | { |
||
909 | var pDepthData = GCHandle.Alloc(sensorData.depthImage, GCHandleType.Pinned); |
||
910 | var pDepthCoordsData = GCHandle.Alloc(vDepthCoords, GCHandleType.Pinned); |
||
911 | |||
912 | coordMapper.MapColorFrameToDepthSpaceUsingIntPtr( |
||
913 | pDepthData.AddrOfPinnedObject(), |
||
914 | (uint)sensorData.depthImage.Length * sizeof(ushort), |
||
915 | pDepthCoordsData.AddrOfPinnedObject(), |
||
916 | (uint)vDepthCoords.Length); |
||
917 | |||
918 | pDepthCoordsData.Free(); |
||
919 | pDepthData.Free(); |
||
920 | |||
921 | return true; |
||
922 | } |
||
923 | |||
924 | return false; |
||
925 | } |
||
926 | |||
927 | // returns the index of the given joint in joint's array or -1 if joint is not applicable |
||
928 | public int GetJointIndex(KinectInterop.JointType joint) |
||
929 | { |
||
930 | return (int)joint; |
||
931 | } |
||
932 | |||
933 | // // returns the joint at given index |
||
934 | // public KinectInterop.JointType GetJointAtIndex(int index) |
||
935 | // { |
||
936 | // return (KinectInterop.JointType)(index); |
||
937 | // } |
||
938 | |||
939 | // returns the parent joint of the given joint |
||
940 | public KinectInterop.JointType GetParentJoint(KinectInterop.JointType joint) |
||
941 | { |
||
942 | switch(joint) |
||
943 | { |
||
944 | case KinectInterop.JointType.SpineBase: |
||
945 | return KinectInterop.JointType.SpineBase; |
||
946 | |||
947 | case KinectInterop.JointType.Neck: |
||
948 | return KinectInterop.JointType.SpineShoulder; |
||
949 | |||
950 | case KinectInterop.JointType.SpineShoulder: |
||
951 | return KinectInterop.JointType.SpineMid; |
||
952 | |||
953 | case KinectInterop.JointType.ShoulderLeft: |
||
954 | case KinectInterop.JointType.ShoulderRight: |
||
955 | return KinectInterop.JointType.SpineShoulder; |
||
956 | |||
957 | case KinectInterop.JointType.HipLeft: |
||
958 | case KinectInterop.JointType.HipRight: |
||
959 | return KinectInterop.JointType.SpineBase; |
||
960 | |||
961 | case KinectInterop.JointType.HandTipLeft: |
||
962 | return KinectInterop.JointType.HandLeft; |
||
963 | |||
964 | case KinectInterop.JointType.ThumbLeft: |
||
965 | return KinectInterop.JointType.WristLeft; |
||
966 | |||
967 | case KinectInterop.JointType.HandTipRight: |
||
968 | return KinectInterop.JointType.HandRight; |
||
969 | |||
970 | case KinectInterop.JointType.ThumbRight: |
||
971 | return KinectInterop.JointType.WristRight; |
||
972 | } |
||
973 | |||
974 | return (KinectInterop.JointType)((int)joint - 1); |
||
975 | } |
||
976 | |||
977 | // returns the next joint in the hierarchy, as to the given joint |
||
978 | public KinectInterop.JointType GetNextJoint(KinectInterop.JointType joint) |
||
979 | { |
||
980 | switch(joint) |
||
981 | { |
||
982 | case KinectInterop.JointType.SpineBase: |
||
983 | return KinectInterop.JointType.SpineMid; |
||
984 | case KinectInterop.JointType.SpineMid: |
||
985 | return KinectInterop.JointType.SpineShoulder; |
||
986 | case KinectInterop.JointType.SpineShoulder: |
||
987 | return KinectInterop.JointType.Neck; |
||
988 | case KinectInterop.JointType.Neck: |
||
989 | return KinectInterop.JointType.Head; |
||
990 | |||
991 | case KinectInterop.JointType.ShoulderLeft: |
||
992 | return KinectInterop.JointType.ElbowLeft; |
||
993 | case KinectInterop.JointType.ElbowLeft: |
||
994 | return KinectInterop.JointType.WristLeft; |
||
995 | case KinectInterop.JointType.WristLeft: |
||
996 | return KinectInterop.JointType.HandLeft; |
||
997 | case KinectInterop.JointType.HandLeft: |
||
998 | return KinectInterop.JointType.HandTipLeft; |
||
999 | |||
1000 | case KinectInterop.JointType.ShoulderRight: |
||
1001 | return KinectInterop.JointType.ElbowRight; |
||
1002 | case KinectInterop.JointType.ElbowRight: |
||
1003 | return KinectInterop.JointType.WristRight; |
||
1004 | case KinectInterop.JointType.WristRight: |
||
1005 | return KinectInterop.JointType.HandRight; |
||
1006 | case KinectInterop.JointType.HandRight: |
||
1007 | return KinectInterop.JointType.HandTipRight; |
||
1008 | |||
1009 | case KinectInterop.JointType.HipLeft: |
||
1010 | return KinectInterop.JointType.KneeLeft; |
||
1011 | case KinectInterop.JointType.KneeLeft: |
||
1012 | return KinectInterop.JointType.AnkleLeft; |
||
1013 | case KinectInterop.JointType.AnkleLeft: |
||
1014 | return KinectInterop.JointType.FootLeft; |
||
1015 | |||
1016 | case KinectInterop.JointType.HipRight: |
||
1017 | return KinectInterop.JointType.KneeRight; |
||
1018 | case KinectInterop.JointType.KneeRight: |
||
1019 | return KinectInterop.JointType.AnkleRight; |
||
1020 | case KinectInterop.JointType.AnkleRight: |
||
1021 | return KinectInterop.JointType.FootRight; |
||
1022 | } |
||
1023 | |||
1024 | return joint; // in case of end joint - Head, HandTipLeft, HandTipRight, FootLeft, FootRight |
||
1025 | } |
||
1026 | |||
1027 | public bool IsFaceTrackingAvailable(ref bool bNeedRestart) |
||
1028 | { |
||
1029 | bool bOneCopied = false, bAllCopied = true; |
||
1030 | string sTargetPath = "."; |
||
1031 | |||
1032 | if(!KinectInterop.Is64bitArchitecture()) |
||
1033 | { |
||
1034 | // 32 bit |
||
1035 | sTargetPath = KinectInterop.GetTargetDllPath(".", false) + "/"; |
||
1036 | |||
1037 | Dictionary<string, string> dictFilesToUnzip = new Dictionary<string, string>(); |
||
1038 | dictFilesToUnzip["Kinect20.Face.dll"] = sTargetPath + "Kinect20.Face.dll"; |
||
1039 | dictFilesToUnzip["KinectFaceUnityAddin.dll"] = sTargetPath + "KinectFaceUnityAddin.dll"; |
||
1040 | dictFilesToUnzip["msvcp110.dll"] = sTargetPath + "msvcp110.dll"; |
||
1041 | dictFilesToUnzip["msvcr110.dll"] = sTargetPath + "msvcr110.dll"; |
||
1042 | |||
1043 | KinectInterop.UnzipResourceFiles(dictFilesToUnzip, "KinectV2UnityAddin.x86.zip", ref bOneCopied, ref bAllCopied); |
||
1044 | } |
||
1045 | else |
||
1046 | { |
||
1047 | //Debug.Log("Face - x64-architecture."); |
||
1048 | sTargetPath = KinectInterop.GetTargetDllPath(".", true) + "/"; |
||
1049 | |||
1050 | Dictionary<string, string> dictFilesToUnzip = new Dictionary<string, string>(); |
||
1051 | dictFilesToUnzip["Kinect20.Face.dll"] = sTargetPath + "Kinect20.Face.dll"; |
||
1052 | dictFilesToUnzip["KinectFaceUnityAddin.dll"] = sTargetPath + "KinectFaceUnityAddin.dll"; |
||
1053 | dictFilesToUnzip["msvcp110.dll"] = sTargetPath + "msvcp110.dll"; |
||
1054 | dictFilesToUnzip["msvcr110.dll"] = sTargetPath + "msvcr110.dll"; |
||
1055 | |||
1056 | KinectInterop.UnzipResourceFiles(dictFilesToUnzip, "KinectV2UnityAddin.x64.zip", ref bOneCopied, ref bAllCopied); |
||
1057 | } |
||
1058 | |||
1059 | KinectInterop.UnzipResourceDirectory(sTargetPath, "NuiDatabase.zip", sTargetPath + "NuiDatabase"); |
||
1060 | |||
1061 | bNeedRestart = (bOneCopied && bAllCopied); |
||
1062 | |||
1063 | return true; |
||
1064 | } |
||
1065 | |||
1066 | public bool InitFaceTracking(bool bUseFaceModel, bool bDrawFaceRect) |
||
1067 | { |
||
1068 | isDrawFaceRect = bDrawFaceRect; |
||
1069 | |||
1070 | // // load the native dlls to make sure libraries are loaded (after previous finish-unload) |
||
1071 | // KinectInterop.LoadNativeLib("Kinect20.Face.dll"); |
||
1072 | // KinectInterop.LoadNativeLib("KinectFaceUnityAddin.dll"); |
||
1073 | |||
1074 | // specify the required face frame results |
||
1075 | FaceFrameFeatures faceFrameFeatures = |
||
1076 | FaceFrameFeatures.BoundingBoxInColorSpace |
||
1077 | //| FaceFrameFeatures.BoundingBoxInInfraredSpace |
||
1078 | | FaceFrameFeatures.PointsInColorSpace |
||
1079 | //| FaceFrameFeatures.PointsInInfraredSpace |
||
1080 | | FaceFrameFeatures.RotationOrientation |
||
1081 | | FaceFrameFeatures.FaceEngagement |
||
1082 | | FaceFrameFeatures.Glasses |
||
1083 | | FaceFrameFeatures.Happy |
||
1084 | | FaceFrameFeatures.LeftEyeClosed |
||
1085 | | FaceFrameFeatures.RightEyeClosed |
||
1086 | | FaceFrameFeatures.LookingAway |
||
1087 | | FaceFrameFeatures.MouthMoved |
||
1088 | | FaceFrameFeatures.MouthOpen |
||
1089 | ; |
||
1090 | |||
1091 | // create a face frame source + reader to track each face in the FOV |
||
1092 | faceFrameSources = new FaceFrameSource[this.bodyCount]; |
||
1093 | faceFrameReaders = new FaceFrameReader[this.bodyCount]; |
||
1094 | |||
1095 | if(bUseFaceModel) |
||
1096 | { |
||
1097 | hdFaceFrameSources = new HighDefinitionFaceFrameSource[this.bodyCount]; |
||
1098 | hdFaceFrameReaders = new HighDefinitionFaceFrameReader[this.bodyCount]; |
||
1099 | |||
1100 | hdFaceModels = new FaceModel[this.bodyCount]; |
||
1101 | hdFaceAlignments = new FaceAlignment[this.bodyCount]; |
||
1102 | } |
||
1103 | |||
1104 | for (int i = 0; i < bodyCount; i++) |
||
1105 | { |
||
1106 | // create the face frame source with the required face frame features and an initial tracking Id of 0 |
||
1107 | faceFrameSources[i] = FaceFrameSource.Create(this.kinectSensor, 0, faceFrameFeatures); |
||
1108 | |||
1109 | // open the corresponding reader |
||
1110 | faceFrameReaders[i] = faceFrameSources[i].OpenReader(); |
||
1111 | |||
1112 | if(bUseFaceModel) |
||
1113 | { |
||
1114 | ///////// HD Face |
||
1115 | hdFaceFrameSources[i] = HighDefinitionFaceFrameSource.Create(this.kinectSensor); |
||
1116 | hdFaceFrameReaders[i] = hdFaceFrameSources[i].OpenReader(); |
||
1117 | |||
1118 | hdFaceModels[i] = FaceModel.Create(); |
||
1119 | hdFaceAlignments[i] = FaceAlignment.Create(); |
||
1120 | } |
||
1121 | } |
||
1122 | |||
1123 | // allocate storage to store face frame results for each face in the FOV |
||
1124 | faceFrameResults = new FaceFrameResult[this.bodyCount]; |
||
1125 | |||
1126 | // FrameDescription frameDescription = this.kinectSensor.ColorFrameSource.FrameDescription; |
||
1127 | // faceDisplayWidth = frameDescription.Width; |
||
1128 | // faceDisplayHeight = frameDescription.Height; |
||
1129 | |||
1130 | bFaceTrackingInited = true; |
||
1131 | |||
1132 | return bFaceTrackingInited; |
||
1133 | } |
||
1134 | |||
1135 | public void FinishFaceTracking() |
||
1136 | { |
||
1137 | if(faceFrameReaders != null) |
||
1138 | { |
||
1139 | for (int i = 0; i < faceFrameReaders.Length; i++) |
||
1140 | { |
||
1141 | if (faceFrameReaders[i] != null) |
||
1142 | { |
||
1143 | faceFrameReaders[i].Dispose(); |
||
1144 | faceFrameReaders[i] = null; |
||
1145 | } |
||
1146 | } |
||
1147 | } |
||
1148 | |||
1149 | if(faceFrameSources != null) |
||
1150 | { |
||
1151 | for (int i = 0; i < faceFrameSources.Length; i++) |
||
1152 | { |
||
1153 | faceFrameSources[i] = null; |
||
1154 | } |
||
1155 | } |
||
1156 | |||
1157 | ///////// HD Face |
||
1158 | if(hdFaceFrameSources != null) |
||
1159 | { |
||
1160 | for (int i = 0; i < hdFaceAlignments.Length; i++) |
||
1161 | { |
||
1162 | hdFaceAlignments[i] = null; |
||
1163 | } |
||
1164 | |||
1165 | for (int i = 0; i < hdFaceModels.Length; i++) |
||
1166 | { |
||
1167 | if (hdFaceModels[i] != null) |
||
1168 | { |
||
1169 | hdFaceModels[i].Dispose(); |
||
1170 | hdFaceModels[i] = null; |
||
1171 | } |
||
1172 | } |
||
1173 | |||
1174 | for (int i = 0; i < hdFaceFrameReaders.Length; i++) |
||
1175 | { |
||
1176 | if (hdFaceFrameReaders[i] != null) |
||
1177 | { |
||
1178 | hdFaceFrameReaders[i].Dispose(); |
||
1179 | hdFaceFrameReaders[i] = null; |
||
1180 | } |
||
1181 | } |
||
1182 | |||
1183 | for (int i = 0; i < hdFaceFrameSources.Length; i++) |
||
1184 | { |
||
1185 | //hdFaceFrameSources[i].Dispose(true); |
||
1186 | hdFaceFrameSources[i] = null; |
||
1187 | } |
||
1188 | } |
||
1189 | |||
1190 | bFaceTrackingInited = false; |
||
1191 | |||
1192 | // // unload the native dlls to prevent hd-face-wrapper's memory leaks |
||
1193 | // KinectInterop.DeleteNativeLib("KinectFaceUnityAddin.dll", true); |
||
1194 | // KinectInterop.DeleteNativeLib("Kinect20.Face.dll", true); |
||
1195 | |||
1196 | } |
||
1197 | |||
1198 | public bool UpdateFaceTracking() |
||
1199 | { |
||
1200 | if(bodyData == null || faceFrameSources == null || faceFrameReaders == null) |
||
1201 | return false; |
||
1202 | |||
1203 | for(int i = 0; i < this.bodyCount; i++) |
||
1204 | { |
||
1205 | if(faceFrameSources[i] != null) |
||
1206 | { |
||
1207 | if(!faceFrameSources[i].IsTrackingIdValid) |
||
1208 | { |
||
1209 | faceFrameSources[i].TrackingId = 0; |
||
1210 | } |
||
1211 | |||
1212 | if(bodyData[i] != null && bodyData[i].IsTracked) |
||
1213 | { |
||
1214 | faceFrameSources[i].TrackingId = bodyData[i].TrackingId; |
||
1215 | } |
||
1216 | } |
||
1217 | |||
1218 | if (faceFrameReaders[i] != null) |
||
1219 | { |
||
1220 | FaceFrame faceFrame = faceFrameReaders[i].AcquireLatestFrame(); |
||
1221 | |||
1222 | if (faceFrame != null) |
||
1223 | { |
||
1224 | int index = GetFaceSourceIndex(faceFrame.FaceFrameSource); |
||
1225 | |||
1226 | if(ValidateFaceBox(faceFrame.FaceFrameResult)) |
||
1227 | { |
||
1228 | faceFrameResults[index] = faceFrame.FaceFrameResult; |
||
1229 | } |
||
1230 | else |
||
1231 | { |
||
1232 | faceFrameResults[index] = null; |
||
1233 | } |
||
1234 | |||
1235 | faceFrame.Dispose(); |
||
1236 | faceFrame = null; |
||
1237 | } |
||
1238 | } |
||
1239 | |||
1240 | ///////// HD Face |
||
1241 | if(hdFaceFrameSources != null && hdFaceFrameSources[i] != null) |
||
1242 | { |
||
1243 | if(!hdFaceFrameSources[i].IsTrackingIdValid) |
||
1244 | { |
||
1245 | hdFaceFrameSources[i].TrackingId = 0; |
||
1246 | } |
||
1247 | |||
1248 | if(bodyData[i] != null && bodyData[i].IsTracked) |
||
1249 | { |
||
1250 | hdFaceFrameSources[i].TrackingId = bodyData[i].TrackingId; |
||
1251 | } |
||
1252 | } |
||
1253 | |||
1254 | if(hdFaceFrameReaders != null && hdFaceFrameReaders[i] != null) |
||
1255 | { |
||
1256 | HighDefinitionFaceFrame hdFaceFrame = hdFaceFrameReaders[i].AcquireLatestFrame(); |
||
1257 | |||
1258 | if(hdFaceFrame != null) |
||
1259 | { |
||
1260 | if(hdFaceFrame.IsFaceTracked && (hdFaceAlignments[i] != null)) |
||
1261 | { |
||
1262 | hdFaceFrame.GetAndRefreshFaceAlignmentResult(hdFaceAlignments[i]); |
||
1263 | } |
||
1264 | |||
1265 | hdFaceFrame.Dispose(); |
||
1266 | hdFaceFrame = null; |
||
1267 | } |
||
1268 | } |
||
1269 | |||
1270 | } |
||
1271 | |||
1272 | return true; |
||
1273 | } |
||
1274 | |||
1275 | private int GetFaceSourceIndex(FaceFrameSource faceFrameSource) |
||
1276 | { |
||
1277 | int index = -1; |
||
1278 | |||
1279 | for (int i = 0; i < this.bodyCount; i++) |
||
1280 | { |
||
1281 | if (this.faceFrameSources[i] == faceFrameSource) |
||
1282 | { |
||
1283 | index = i; |
||
1284 | break; |
||
1285 | } |
||
1286 | } |
||
1287 | |||
1288 | return index; |
||
1289 | } |
||
1290 | |||
1291 | private bool ValidateFaceBox(FaceFrameResult faceResult) |
||
1292 | { |
||
1293 | bool isFaceValid = faceResult != null; |
||
1294 | |||
1295 | if (isFaceValid) |
||
1296 | { |
||
1297 | var faceBox = faceResult.FaceBoundingBoxInColorSpace; |
||
1298 | //if (faceBox != null) |
||
1299 | { |
||
1300 | // check if we have a valid rectangle within the bounds of the screen space |
||
1301 | isFaceValid = (faceBox.Right - faceBox.Left) > 0 && |
||
1302 | (faceBox.Bottom - faceBox.Top) > 0; // && |
||
1303 | //faceBox.Right <= this.faceDisplayWidth && |
||
1304 | //faceBox.Bottom <= this.faceDisplayHeight; |
||
1305 | } |
||
1306 | } |
||
1307 | |||
1308 | return isFaceValid; |
||
1309 | } |
||
1310 | |||
1311 | public bool IsFaceTrackingActive() |
||
1312 | { |
||
1313 | return bFaceTrackingInited; |
||
1314 | } |
||
1315 | |||
1316 | public bool IsDrawFaceRect() |
||
1317 | { |
||
1318 | return isDrawFaceRect; |
||
1319 | } |
||
1320 | |||
1321 | public bool IsFaceTracked(long userId) |
||
1322 | { |
||
1323 | for (int i = 0; i < this.bodyCount; i++) |
||
1324 | { |
||
1325 | if(faceFrameSources != null && faceFrameSources[i] != null && faceFrameSources[i].TrackingId == (ulong)userId) |
||
1326 | { |
||
1327 | if(faceFrameResults != null && faceFrameResults[i] != null) |
||
1328 | { |
||
1329 | return true; |
||
1330 | } |
||
1331 | } |
||
1332 | } |
||
1333 | |||
1334 | return false; |
||
1335 | } |
||
1336 | |||
1337 | public bool GetFaceRect(long userId, ref Rect faceRect) |
||
1338 | { |
||
1339 | for (int i = 0; i < this.bodyCount; i++) |
||
1340 | { |
||
1341 | if(faceFrameSources != null && faceFrameSources[i] != null && faceFrameSources[i].TrackingId == (ulong)userId) |
||
1342 | { |
||
1343 | if(faceFrameResults != null && faceFrameResults[i] != null) |
||
1344 | { |
||
1345 | var faceBox = faceFrameResults[i].FaceBoundingBoxInColorSpace; |
||
1346 | |||
1347 | //if (faceBox != null) |
||
1348 | { |
||
1349 | faceRect.x = faceBox.Left; |
||
1350 | faceRect.y = faceBox.Top; |
||
1351 | faceRect.width = faceBox.Right - faceBox.Left; |
||
1352 | faceRect.height = faceBox.Bottom - faceBox.Top; |
||
1353 | |||
1354 | return true; |
||
1355 | } |
||
1356 | } |
||
1357 | } |
||
1358 | } |
||
1359 | |||
1360 | return false; |
||
1361 | } |
||
1362 | |||
1363 | public void VisualizeFaceTrackerOnColorTex(Texture2D texColor) |
||
1364 | { |
||
1365 | if(bFaceTrackingInited) |
||
1366 | { |
||
1367 | for (int i = 0; i < this.bodyCount; i++) |
||
1368 | { |
||
1369 | if(faceFrameSources != null && faceFrameSources[i] != null && faceFrameSources[i].IsTrackingIdValid) |
||
1370 | { |
||
1371 | if(faceFrameResults != null && faceFrameResults[i] != null) |
||
1372 | { |
||
1373 | var faceBox = faceFrameResults[i].FaceBoundingBoxInColorSpace; |
||
1374 | |||
1375 | //if (faceBox != null) |
||
1376 | { |
||
1377 | UnityEngine.Color color = UnityEngine.Color.magenta; |
||
1378 | Vector2 pt1, pt2; |
||
1379 | |||
1380 | // bottom |
||
1381 | pt1.x = faceBox.Left; pt1.y = faceBox.Top; |
||
1382 | pt2.x = faceBox.Right; pt2.y = pt1.y; |
||
1383 | DrawLine(texColor, pt1, pt2, color); |
||
1384 | |||
1385 | // right |
||
1386 | pt1.x = pt2.x; pt1.y = pt2.y; |
||
1387 | pt2.x = pt1.x; pt2.y = faceBox.Bottom; |
||
1388 | DrawLine(texColor, pt1, pt2, color); |
||
1389 | |||
1390 | // top |
||
1391 | pt1.x = pt2.x; pt1.y = pt2.y; |
||
1392 | pt2.x = faceBox.Left; pt2.y = pt1.y; |
||
1393 | DrawLine(texColor, pt1, pt2, color); |
||
1394 | |||
1395 | // left |
||
1396 | pt1.x = pt2.x; pt1.y = pt2.y; |
||
1397 | pt2.x = pt1.x; pt2.y = faceBox.Top; |
||
1398 | DrawLine(texColor, pt1, pt2, color); |
||
1399 | } |
||
1400 | } |
||
1401 | } |
||
1402 | } |
||
1403 | } |
||
1404 | } |
||
1405 | |||
1406 | private void DrawLine(Texture2D a_Texture, Vector2 ptStart, Vector2 ptEnd, UnityEngine.Color a_Color) |
||
1407 | { |
||
1408 | KinectInterop.DrawLine(a_Texture, (int)ptStart.x, (int)ptStart.y, (int)ptEnd.x, (int)ptEnd.y, a_Color); |
||
1409 | } |
||
1410 | |||
1411 | public bool GetHeadPosition(long userId, ref Vector3 headPos) |
||
1412 | { |
||
1413 | for (int i = 0; i < this.bodyCount; i++) |
||
1414 | { |
||
1415 | if(bodyData[i].TrackingId == (ulong)userId && bodyData[i].IsTracked) |
||
1416 | { |
||
1417 | CameraSpacePoint vHeadPos = bodyData[i].Joints[Windows.Kinect.JointType.Head].Position; |
||
1418 | |||
1419 | if(vHeadPos.Z > 0f) |
||
1420 | { |
||
1421 | headPos.x = vHeadPos.X; |
||
1422 | headPos.y = vHeadPos.Y; |
||
1423 | headPos.z = vHeadPos.Z; |
||
1424 | |||
1425 | return true; |
||
1426 | } |
||
1427 | } |
||
1428 | } |
||
1429 | |||
1430 | return false; |
||
1431 | } |
||
1432 | |||
1433 | public bool GetHeadRotation(long userId, ref Quaternion headRot) |
||
1434 | { |
||
1435 | for (int i = 0; i < this.bodyCount; i++) |
||
1436 | { |
||
1437 | if(faceFrameSources != null && faceFrameSources[i] != null && faceFrameSources[i].TrackingId == (ulong)userId) |
||
1438 | { |
||
1439 | if(faceFrameResults != null && faceFrameResults[i] != null) |
||
1440 | { |
||
1441 | Windows.Kinect.Vector4 vHeadRot = faceFrameResults[i].FaceRotationQuaternion; |
||
1442 | |||
1443 | if(vHeadRot.W > 0f) |
||
1444 | { |
||
1445 | headRot = new Quaternion(vHeadRot.X, vHeadRot.Y, vHeadRot.Z, vHeadRot.W); |
||
1446 | return true; |
||
1447 | } |
||
1448 | // else |
||
1449 | // { |
||
1450 | // Debug.Log(string.Format("Bad rotation: ({0:F2}, {1:F2}, {2:F2}, {3:F2}})", vHeadRot.X, vHeadRot.Y, vHeadRot.Z, vHeadRot.W)); |
||
1451 | // return false; |
||
1452 | // } |
||
1453 | |||
1454 | } |
||
1455 | } |
||
1456 | } |
||
1457 | |||
1458 | return false; |
||
1459 | } |
||
1460 | |||
1461 | public bool GetAnimUnits(long userId, ref Dictionary<KinectInterop.FaceShapeAnimations, float> dictAU) |
||
1462 | { |
||
1463 | for (int i = 0; i < this.bodyCount; i++) |
||
1464 | { |
||
1465 | if(hdFaceFrameSources != null && hdFaceFrameSources[i] != null && hdFaceFrameSources[i].TrackingId == (ulong)userId) |
||
1466 | { |
||
1467 | if(hdFaceAlignments != null && hdFaceAlignments[i] != null) |
||
1468 | { |
||
1469 | foreach(Microsoft.Kinect.Face.FaceShapeAnimations akey in hdFaceAlignments[i].AnimationUnits.Keys) |
||
1470 | { |
||
1471 | dictAU[(KinectInterop.FaceShapeAnimations)akey] = hdFaceAlignments[i].AnimationUnits[akey]; |
||
1472 | } |
||
1473 | |||
1474 | return true; |
||
1475 | } |
||
1476 | } |
||
1477 | } |
||
1478 | |||
1479 | return false; |
||
1480 | } |
||
1481 | |||
1482 | public bool GetShapeUnits(long userId, ref Dictionary<KinectInterop.FaceShapeDeformations, float> dictSU) |
||
1483 | { |
||
1484 | for (int i = 0; i < this.bodyCount; i++) |
||
1485 | { |
||
1486 | if(hdFaceFrameSources != null && hdFaceFrameSources[i] != null && hdFaceFrameSources[i].TrackingId == (ulong)userId) |
||
1487 | { |
||
1488 | if(hdFaceModels != null && hdFaceModels[i] != null) |
||
1489 | { |
||
1490 | foreach(Microsoft.Kinect.Face.FaceShapeDeformations skey in hdFaceModels[i].FaceShapeDeformations.Keys) |
||
1491 | { |
||
1492 | dictSU[(KinectInterop.FaceShapeDeformations)skey] = hdFaceModels[i].FaceShapeDeformations[skey]; |
||
1493 | } |
||
1494 | |||
1495 | return true; |
||
1496 | } |
||
1497 | } |
||
1498 | } |
||
1499 | |||
1500 | return false; |
||
1501 | } |
||
1502 | |||
1503 | public int GetFaceModelVerticesCount(long userId) |
||
1504 | { |
||
1505 | for (int i = 0; i < this.bodyCount; i++) |
||
1506 | { |
||
1507 | if(hdFaceFrameSources != null && hdFaceFrameSources[i] != null && (hdFaceFrameSources[i].TrackingId == (ulong)userId || userId == 0)) |
||
1508 | { |
||
1509 | if(hdFaceModels != null && hdFaceModels[i] != null) |
||
1510 | { |
||
1511 | var vertices = hdFaceModels[i].CalculateVerticesForAlignment(hdFaceAlignments[i]); |
||
1512 | int verticesCount = vertices.Count; |
||
1513 | |||
1514 | return verticesCount; |
||
1515 | } |
||
1516 | } |
||
1517 | } |
||
1518 | |||
1519 | return 0; |
||
1520 | } |
||
1521 | |||
1522 | public bool GetFaceModelVertices(long userId, ref Vector3[] avVertices) |
||
1523 | { |
||
1524 | for (int i = 0; i < this.bodyCount; i++) |
||
1525 | { |
||
1526 | if(hdFaceFrameSources != null && hdFaceFrameSources[i] != null && (hdFaceFrameSources[i].TrackingId == (ulong)userId || userId == 0)) |
||
1527 | { |
||
1528 | if(hdFaceModels != null && hdFaceModels[i] != null) |
||
1529 | { |
||
1530 | var vertices = hdFaceModels[i].CalculateVerticesForAlignment(hdFaceAlignments[i]); |
||
1531 | int verticesCount = vertices.Count; |
||
1532 | |||
1533 | if(avVertices.Length == verticesCount) |
||
1534 | { |
||
1535 | for(int v = 0; v < verticesCount; v++) |
||
1536 | { |
||
1537 | avVertices[v].x = vertices[v].X; |
||
1538 | avVertices[v].y = vertices[v].Y; |
||
1539 | avVertices[v].z = vertices[v].Z; // -vertices[v].Z; |
||
1540 | } |
||
1541 | } |
||
1542 | |||
1543 | return true; |
||
1544 | } |
||
1545 | } |
||
1546 | } |
||
1547 | |||
1548 | return false; |
||
1549 | } |
||
1550 | |||
1551 | public int GetFaceModelTrianglesCount() |
||
1552 | { |
||
1553 | var triangleIndices = FaceModel.TriangleIndices; |
||
1554 | int triangleLength = triangleIndices.Count; |
||
1555 | |||
1556 | return triangleLength; |
||
1557 | } |
||
1558 | |||
1559 | public bool GetFaceModelTriangles(bool bMirrored, ref int[] avTriangles) |
||
1560 | { |
||
1561 | var triangleIndices = FaceModel.TriangleIndices; |
||
1562 | int triangleLength = triangleIndices.Count; |
||
1563 | |||
1564 | if(avTriangles.Length >= triangleLength) |
||
1565 | { |
||
1566 | for(int i = 0; i < triangleLength; i += 3) |
||
1567 | { |
||
1568 | //avTriangles[i] = (int)triangleIndices[i]; |
||
1569 | avTriangles[i] = (int)triangleIndices[i + 2]; |
||
1570 | avTriangles[i + 1] = (int)triangleIndices[i + 1]; |
||
1571 | avTriangles[i + 2] = (int)triangleIndices[i]; |
||
1572 | } |
||
1573 | |||
1574 | if(bMirrored) |
||
1575 | { |
||
1576 | Array.Reverse(avTriangles); |
||
1577 | } |
||
1578 | |||
1579 | return true; |
||
1580 | } |
||
1581 | |||
1582 | return false; |
||
1583 | } |
||
1584 | |||
1585 | public bool IsSpeechRecognitionAvailable(ref bool bNeedRestart) |
||
1586 | { |
||
1587 | bool bOneCopied = false, bAllCopied = true; |
||
1588 | |||
1589 | if(!KinectInterop.Is64bitArchitecture()) |
||
1590 | { |
||
1591 | //Debug.Log("Speech - x32-architecture."); |
||
1592 | string sTargetPath = KinectInterop.GetTargetDllPath(".", false) + "/"; |
||
1593 | |||
1594 | Dictionary<string, string> dictFilesToUnzip = new Dictionary<string, string>(); |
||
1595 | dictFilesToUnzip["Kinect2SpeechWrapper.dll"] = sTargetPath + "Kinect2SpeechWrapper.dll"; |
||
1596 | dictFilesToUnzip["msvcp110.dll"] = sTargetPath + "msvcp110.dll"; |
||
1597 | dictFilesToUnzip["msvcr110.dll"] = sTargetPath + "msvcr110.dll"; |
||
1598 | |||
1599 | KinectInterop.UnzipResourceFiles(dictFilesToUnzip, "KinectV2UnityAddin.x86.zip", ref bOneCopied, ref bAllCopied); |
||
1600 | } |
||
1601 | else |
||
1602 | { |
||
1603 | //Debug.Log("Face - x64-architecture."); |
||
1604 | string sTargetPath = KinectInterop.GetTargetDllPath(".", true) + "/"; |
||
1605 | |||
1606 | Dictionary<string, string> dictFilesToUnzip = new Dictionary<string, string>(); |
||
1607 | dictFilesToUnzip["Kinect2SpeechWrapper.dll"] = sTargetPath + "Kinect2SpeechWrapper.dll"; |
||
1608 | dictFilesToUnzip["msvcp110.dll"] = sTargetPath + "msvcp110.dll"; |
||
1609 | dictFilesToUnzip["msvcr110.dll"] = sTargetPath + "msvcr110.dll"; |
||
1610 | |||
1611 | KinectInterop.UnzipResourceFiles(dictFilesToUnzip, "KinectV2UnityAddin.x64.zip", ref bOneCopied, ref bAllCopied); |
||
1612 | } |
||
1613 | |||
1614 | bNeedRestart = (bOneCopied && bAllCopied); |
||
1615 | |||
1616 | return true; |
||
1617 | } |
||
1618 | |||
1619 | public int InitSpeechRecognition(string sRecoCriteria, bool bUseKinect, bool bAdaptationOff) |
||
1620 | { |
||
1621 | // if(kinectSensor != null) |
||
1622 | // { |
||
1623 | // float fWaitTime = Time.realtimeSinceStartup + 5f; |
||
1624 | // |
||
1625 | // while(!kinectSensor.IsAvailable && Time.realtimeSinceStartup < fWaitTime) |
||
1626 | // { |
||
1627 | // // wait |
||
1628 | // } |
||
1629 | // } |
||
1630 | |||
1631 | return InitSpeechRecognizerNative(sRecoCriteria, bUseKinect, bAdaptationOff); |
||
1632 | } |
||
1633 | |||
1634 | public void FinishSpeechRecognition() |
||
1635 | { |
||
1636 | FinishSpeechRecognizerNative(); |
||
1637 | } |
||
1638 | |||
1639 | public int UpdateSpeechRecognition() |
||
1640 | { |
||
1641 | return UpdateSpeechRecognizerNative(); |
||
1642 | } |
||
1643 | |||
1644 | public int LoadSpeechGrammar(string sFileName, short iLangCode, bool bDynamic) |
||
1645 | { |
||
1646 | return LoadSpeechGrammarNative(sFileName, iLangCode, bDynamic); |
||
1647 | |||
1648 | // int hr = AddSpeechGrammarNative(sFileName, iLangCode, bDynamic); |
||
1649 | // if(hr >= 0) |
||
1650 | // { |
||
1651 | // hr = SetGrammarStateNative(sFileName, true); |
||
1652 | // } |
||
1653 | // |
||
1654 | // return hr; |
||
1655 | } |
||
1656 | |||
1657 | public int AddGrammarPhrase(string sFromRule, string sToRule, string sPhrase, bool bClearRulePhrases, bool bCommitGrammar) |
||
1658 | { |
||
1659 | return AddGrammarPhraseNative(sFromRule, sToRule, sPhrase, bClearRulePhrases, bCommitGrammar); |
||
1660 | } |
||
1661 | |||
1662 | public void SetSpeechConfidence(float fConfidence) |
||
1663 | { |
||
1664 | SetSpeechConfidenceNative(fConfidence); |
||
1665 | } |
||
1666 | |||
1667 | public bool IsSpeechStarted() |
||
1668 | { |
||
1669 | return IsSpeechStartedNative(); |
||
1670 | } |
||
1671 | |||
1672 | public bool IsSpeechEnded() |
||
1673 | { |
||
1674 | return IsSpeechEndedNative(); |
||
1675 | } |
||
1676 | |||
1677 | public bool IsPhraseRecognized() |
||
1678 | { |
||
1679 | return IsPhraseRecognizedNative(); |
||
1680 | } |
||
1681 | |||
1682 | public float GetPhraseConfidence() |
||
1683 | { |
||
1684 | return GetPhraseConfidenceNative(); |
||
1685 | } |
||
1686 | |||
1687 | public string GetRecognizedPhraseTag() |
||
1688 | { |
||
1689 | IntPtr pPhraseTag = GetRecognizedPhraseTagNative(); |
||
1690 | string sPhraseTag = Marshal.PtrToStringUni(pPhraseTag); |
||
1691 | |||
1692 | return sPhraseTag; |
||
1693 | } |
||
1694 | |||
1695 | public void ClearRecognizedPhrase() |
||
1696 | { |
||
1697 | ClearRecognizedPhraseNative(); |
||
1698 | } |
||
1699 | |||
1700 | public bool IsBackgroundRemovalAvailable(ref bool bNeedRestart) |
||
1701 | { |
||
1702 | bBackgroundRemovalInited = KinectInterop.IsOpenCvAvailable(ref bNeedRestart); |
||
1703 | return bBackgroundRemovalInited; |
||
1704 | } |
||
1705 | |||
1706 | public bool InitBackgroundRemoval(KinectInterop.SensorData sensorData, bool isHiResPrefered) |
||
1707 | { |
||
1708 | return KinectInterop.InitBackgroundRemoval(sensorData, isHiResPrefered); |
||
1709 | } |
||
1710 | |||
1711 | public void FinishBackgroundRemoval(KinectInterop.SensorData sensorData) |
||
1712 | { |
||
1713 | KinectInterop.FinishBackgroundRemoval(sensorData); |
||
1714 | bBackgroundRemovalInited = false; |
||
1715 | } |
||
1716 | |||
1717 | public bool UpdateBackgroundRemoval(KinectInterop.SensorData sensorData, bool isHiResPrefered, Color32 defaultColor, bool bAlphaTexOnly) |
||
1718 | { |
||
1719 | return KinectInterop.UpdateBackgroundRemoval(sensorData, isHiResPrefered, defaultColor, bAlphaTexOnly); |
||
1720 | } |
||
1721 | |||
1722 | public bool IsBackgroundRemovalActive() |
||
1723 | { |
||
1724 | return bBackgroundRemovalInited; |
||
1725 | } |
||
1726 | |||
1727 | public bool IsBRHiResSupported() |
||
1728 | { |
||
1729 | return true; |
||
1730 | } |
||
1731 | |||
1732 | public Rect GetForegroundFrameRect(KinectInterop.SensorData sensorData, bool isHiResPrefered) |
||
1733 | { |
||
1734 | return KinectInterop.GetForegroundFrameRect(sensorData, isHiResPrefered); |
||
1735 | } |
||
1736 | |||
1737 | public int GetForegroundFrameLength(KinectInterop.SensorData sensorData, bool isHiResPrefered) |
||
1738 | { |
||
1739 | return KinectInterop.GetForegroundFrameLength(sensorData, isHiResPrefered); |
||
1740 | } |
||
1741 | |||
1742 | public bool PollForegroundFrame(KinectInterop.SensorData sensorData, bool isHiResPrefered, Color32 defaultColor, bool bLimitedUsers, ICollection<int> alTrackedIndexes, ref byte[] foregroundImage) |
||
1743 | { |
||
1744 | return KinectInterop.PollForegroundFrame(sensorData, isHiResPrefered, defaultColor, bLimitedUsers, alTrackedIndexes, ref foregroundImage); |
||
1745 | } |
||
1746 | |||
1747 | } |
||
1748 | #endif |