Perception is a set of OpenCV routines made available in Objective-C. This framework mostly exists because Swift doesn't support calling to C++.
No particular attention is paid to maximising data marshalling performance – NSArrays of NSNumbers are used like an animal (mostly because of laziness and because the data marshalling is a small fraction of runtime of the methods made available here).
The following steps were roughly speaking involved in building the OpenCV macOS binary that's statically linked into this framework, not automated at the moment to any extent:
-
git clone [email protected]:opencv/opencv.git
- The commit built is d8425d88816ef7dbc9c8e5fa4cce407c7cf7a64e (this specific commit in the 3.2.0 development series adds some APIs I needed).
-
git clone [email protected]:opencv/opencv_contrib.git
- The tag built is 3.2.0.
-
Build opencv (and copy libraries and include directories in place after):
cd opencv
mkdir build
cmake -DBUILD_SHARED_LIBS=OFF -DWITH_IPP=OFF -DOPENCV_ENABLE_NONFREE=ON -DOPENCV_EXTRA_MODULES_PATH=/Users/mz2/Projects/opencv_contrib/modules ../