= PoseOSC + FaceOSC + HandOSC + OcrOSC + CatOSC + DogOSC
Send (almost) all Apple Vision Framework's detection results via OSC. (You can pick which one(s) to detect & send). Written in openFrameworks using Objective-C++. macOS 11+ only.
Inspired by PoseOSC, but faster, no more electron-bloat or horse neighing hacks. Compatible with the ARR
format of PoseOSC.
Do not attempt to re-build (with projectGenerator) unless absolutely necessary, in which case follow the following steps:
- File > Project Settings, Build System -> New Build System
- Left sidebar, click project name, General > Frameworks, Libraries,... Add Vision, AVKit, Foundation, AVFoundation, CoreML
- Build Phases, Link Binary with Libraries, Change "mac catalyst" to "always"
- Change deployment target to 11.3
- For each file in src folder, on right sidebar, change file "Type" to Objective C++ (not extension, just in the dropdown menu)
If you encounter Undefined symbol: __objc_msgSend$identifier
, you might need to set Excluded Architectures arm64
. See this issue for details.
If there're some complaints about ARC, you might need to remove all mentions of autorelease
in src
folder.
Settings in settings.xml
will be loaded upon start.
In the packaged app, the settings.xml
can be found in Contents/Resources
.
See demos/VisionOSCProcessingReceiver for a Processing demo receiving all the detection types.
This is the same as ARR
format of PoseOSC, copied below:
ARR will be sent to poses/arr
OSC Address as an array of values (OSC spec allows multiple values of different types for each address).
- The first value (int) is width of the frame.
- The second value (int) is height of the frame.
- The third value (int) is the number of poses. (When you read this value, you'll know how many more values to read, i.e.
nPoses*(1+17*3)
. So if this number is 0 it means no pose is detected, so you can stop reading). - The next 52 values are data for the first pose, and the 52 values after that are data for the second pose (if there is), and so on...
- For each pose, the first value (float) is the score for that pose, the rest 51 values (floats) can be divided into 17 groups of 3, with each group being (x,y,score) of a keypoint. For the ordering of keypoints, see PoseNet spec.
Similar to pose format (see above); sent to faces/arr
OSC Address:
- The first value (int) is width of the frame.
- The second value (int) is height of the frame.
- The third value (int) is the number of faces.
- The next 229 values are data for the first face, and the 229 values after that are data for the second face (if there is), and so on...
- For each face, the first value (float) is the score for that face, the rest 228 values (floats) can be divided into 76 groups of 3, with each group being (x,y,score) of a keypoint.
Similar to pose format (see above); sent to hands/arr
OSC Address:
- The first value (int) is width of the frame.
- The second value (int) is height of the frame.
- The third value (int) is the number of hands.
- The next 64 values are data for the first hand, and the 64 values after that are data for the second hand (if there is), and so on...
- For each hand, the first value (float) is the score for that hand, the rest 63 values (floats) can be divided into 21 groups of 3, with each group being (x,y,score) of a keypoint. For the ordering of the keypoints, see handpose spec
Sent to texts/arr
OSC Address:
- The first value (int) is width of the frame.
- The second value (int) is height of the frame.
- The third value (int) is the number of text regions.
- The next 6 values are data for the first text, and the 6 values after that are data for the second text (if there is), and so on...
- For each text, the first value (float) is the score for that text, the next four values (float) are the (left,top,width,height) of the bounding box. The last value is what the text says (string).
Currently only cats and dogs are supported, per Apple's documentation.
Similar to texts format (see above); sent to animals/arr
OSC Address:
- The first value (int) is width of the frame.
- The second value (int) is height of the frame.
- The third value (int) is the number of animals.
- The next 6 values are data for the first animal, and the 6 values after that are data for the second animal (if there is), and so on...
- For each animal, the first value (float) is the score for that animal, the next four values (float) are the (left,top,width,height) of the bounding box. The last value is what the animal is (string): "Cat"/"Dog".
The JSON
and XML
formats supported by PoseOSC are now excluded because I've since realized it's a silly idea to add this sort of parsing overhead. Let me know if you have a case against this decision.
I recommand Protokol for testing/inspecting OSC.
Tested on MacBook Pro (13-inch, M1, 2020) Memory 16 GB.
- Body: 60 FPS
- Hand: 60 FPS
- Face: 45 FPS
- Text: 10 FPS
- Animal: 60FPS
- Face, body, & hand: 25 FPS
- Everything all on: 5 FPS