Learn more >
Get the code
View the demo
Build on IBM Cloud
by Sanjeev Ghimire | Published April 4, 2018
Artificial intelligenceMobile developmentVirtual realityVisual recognitionGaming
Augmented reality provides an enhanced version of reality by superimposing virtual objects over a user’s view of the real world. ARKit blends digital objects and information with the environment around you, taking apps far beyond the screen and freeing them to interact with the real world in entirely new ways. This code pattern combines ARKit with Watson Visual Recognition and a Cloudant database to give you a complete augmented reality experience.
The easiest way to find and connect to people around the world is through social media apps like Facebook, Twitter, and LinkedIn. However, these only provide text-based search capabilities. With the recently announced release of the iOS ARKit toolkit, you can now do a search using facial recognition. Combining iOS face recognition using the Vision API, classification using IBM Watson Visual Recognition, and person identification using classified image and data, you can build an app to search faces and identify them. One use case is to build an augmented reality-based résumé using visual recognition.
This code pattern explains how to create this type of augmented reality-based résumés with Visual Recognition. The iOS app recognizes the face and presents you with the AR view that displays a résumé of the person in the camera view. The app uses Watson Visual Recognition and Core ML to classify the image and then uses that classification to get details about the person from data stored in an IBM Cloudant NoSQL database. The images are classified offline using a deep neural network that is trained by Visual Recognition.
After completing this code pattern, you should know how to:
Find the detailed steps for this pattern in the README. Those steps will show you how to:
March 11, 2019
Back to top