Simple project to detect objects and display 3D labels above them in AR. This serves as a basic Template for an ARKit project to use CoreML.
This simple project detects objects in Augmented Reality and displays 3D labels on top of them. This serves as a basic template for an ARKit project to use CoreML.
Model: Inception V3
Language: Swift 4.0
Written in: Xcode 9.0 GM (9A235) (Updated) XCode 9 beta 3 (9M174d)
Content Technology: SceneKit
Tested on iPhone 7 plus running iOS 11 beta 3 (15A5318g)
Note: SceneKit can achieve a 60 FPS on iPhone7+ - though when it gets hot, it’ll drop to 30 FPS.
You’ll have to download “Inceptionv3.mlmodel” from Apple’s Machine Learning page, and copy it into your XCode project. (As depicted in the following gif)
If you’re having issues, double check that the model is part of a target (source: stackoverflow).
SceneKit Text Labels are expensive to render. Too many polygons (too much text, smoothness, characters) - can cause crashes. In future, SpriteKit would be more efficient for text-labels.
Not entirely certain if the code is actually transferring RGB data to the Vision Model (as opposed to YUV). Proof-of-concept-wise, it appears to work just fine with the Inception V3 model for now.
Whilst ARKit’s FPS , is displayed - CoreML’s speed is not. However, it does appear sufficiently fast for real-time ARKit applications.
Placement of the label is simply determined by the raycast screen centre-point to a ARKit feature-point. This could be altered for more stable placement.
There are some good tutorials / sample projects for getting CoreML running. See: [ 1 ] [ 2 ] [ 3 ]
What we do differently here is we’re using ARKit’s ARFrame as the image to be fed into CoreML.
let pixbuff : CVPixelBuffer? = (sceneView.session.currentFrame?.capturedImage)
let dispatchQueueML = DispatchQueue(label: "com.hw.dispatchqueueml")
...
loopCoreMLUpdate() // on viewLoad
...
func loopCoreMLUpdate() {
dispatchQueueML.async {
// 1. Run Update.
self.updateCoreML()
// 2. Loop this function.
self.loopCoreMLUpdate()
}
}