Questions tagged [apple-vision]

Apple Vision is a high-level computer vision framework to identify faces, detect and track features, and classify images, video, tabular data, audio and motion sensors data.

Apple Vision framework performs face and face landmark detection on input images and video, barcode recognition, image registration, text detection, and, of course, feature tracking. Vision API allows the use of custom CoreML models for tasks like classification or object detection.

205 questions
2
votes
2 answers

Using Vision to scan images from photo library

Is there a way that I can use the Vision framework to scan an existing image from the user's photo library? As in, not taking a new picture using the camera, but just choosing an image that the user already has?
rice888
  • 51
  • 8
2
votes
1 answer

Enum for .mlmodels?

So I am currently working on a project that includes image recognition with CreateML, CoreML, and Vision. I am still trying datasets and improving the models with CreateML, but if I change from a model to another, I have to manually change a…
NicolasElPapu
  • 1,612
  • 2
  • 11
  • 26
2
votes
2 answers

Apple Vision Framework: LCD/LED digit recognition

I was developing on an iOS app and everything seemed to work pretty well until I tried capturing images of digital clock, calculators, blood pressure monitors, electronic thermometers, etc. For some reason Apple Vision Framework and…
Michael Samoylov
  • 2,933
  • 3
  • 25
  • 33
2
votes
0 answers

Vision – Face recognition performed but correct coordinates cannot be obtained

Swift 5, Xcode 11, iOS 13.0. In the code below, we will get the face from the image 'test', recognize the left eye, and display 'num' at its coordinates. However, the coordinates of the nose are not displayed at the correct position. I'm in trouble…
notane
  • 31
  • 3
2
votes
1 answer

Vision Framework vs CoreML Model Class

I have a CoreML model that I generated using CreateML. If I drag and drop that model into Xcode it will create a class for me automatically, which I can use to detect/classify the image. The generated class has prediction function which will return…
john doe
  • 9,220
  • 23
  • 91
  • 167
2
votes
2 answers

How to apply a 3D Model on detected face by Apple Vision "NO AR"

With iPhoneX True-Depth camera its possible to get the 3D Coordinates of any object and use that information to position and scale the object, but with older iPhones we don't have access to AR on front-face camera, what i've done so far was…
Shahriyar
  • 520
  • 7
  • 18
2
votes
0 answers

How calculate transform and position of SCNNode from .scn for area of ARSCNView

I dont undestand how I can calculate SCNNode position on my ARSCNView with using Vision object. I have a detected face (VNFaceObservation) in ARSCNView, load node from .scn file. And I want that node was moving with face position. I calculate 2D…
Stas Telnov
  • 305
  • 2
  • 13
2
votes
2 answers

How can I only allow entering my virtual scene from a portal?

I have an application which will render an augmented reality scene and a portal for which you can walk into the scene. The scene is occluded from view by a plane, but if you walk through that plane, you "bust" into the virtual environment. I'm not…
giraffesyo
  • 4,860
  • 1
  • 29
  • 39
2
votes
1 answer

How to translate X-axis correctly from VNFaceObservation boundingBox (Vision + ARKit)

I'm using both ARKit & Vision, following along Apple's sample project, "Using Vision in Real Time with ARKit". So I am not setting up my camera as ARKit handles that for me. Using Vision's VNDetectFaceRectanglesRequest, I'm able to get back a…
xta
  • 729
  • 2
  • 8
  • 29
2
votes
3 answers

How to fix IOAF code GPU errors while using ARKit2 & Vision (VNDetectFaceRectanglesRequest) on iPhone XS

While running ARKit on iPhone XS (with iOS 12.1.2 and Xcode 10.1), I'm getting errors and crashes/hangs while running Vision code to detect face bounds. Errors I'm getting are: 2019-01-04 03:03:03.155867-0800 ARKit Vision Demo[12969:3307770]…
xta
  • 729
  • 2
  • 8
  • 29
2
votes
1 answer

How to detect contours of object and describe it to compare on server with ARKit

I want to detect shape and then describe it (somehow) to compare it with server data. So the first question is, is it possible to detect shape like blob with ARKit? To be more specific, let's describe my usecase generally. I want to scan image by…
Boomerange
  • 616
  • 1
  • 10
  • 18
2
votes
1 answer

How does the minimum size work on detected rectangle using VNDetectRectanglesRequest?

Can you please explain how the minimumSize property of VNDetectRectanglesRequest works. I read the details but did not understand it. Description from the docs A float specifying the minimum size of the rectangle to be detected, as a proportion of…
Bhushan B
  • 2,470
  • 4
  • 26
  • 38
2
votes
2 answers

Vision and ARKit frameworks in Xcode project

I want to create an ARKit app using Xcode. I want it to recognize a generic rectangle without pressing a button and that subsequently the rectangle does a certain function. How to do it?
2
votes
0 answers

How to put CALayer on CMSampleBuffer of video data output?

Here's my code: func captureOutput(_ output: AVCaptureOutput, didOutput sampleBuffer: CMSampleBuffer, from connection: AVCaptureConnection) { guard CMSampleBufferDataIsReady(sampleBuffer) else { return } // This is the…
Tushar Katyal
  • 412
  • 5
  • 12
2
votes
0 answers

ARKit "pool of pixel buffers", why does it freeze the app?

In Apple's tutorial on performing computer vision tasks within ARKit they note: Important Making sure only one buffer is being processed at a time ensures good performance. The camera recycles a finite pool of pixel buffers, so retaining too…
freakTheMighty
  • 1,172
  • 1
  • 12
  • 27