This guide shows you how to:
- Set up your development environment to work with Augmented Faces
- Apply 2D textures and attach 3D objects to detected faces
- Xcode version 13.0 or later
- Cocoapods 1.4.0 or later if using Cocoapods
- An ARKit-compatible Apple device running iOS 12.0 or later (deployment target of iOS 12.0 or later required)
Using Augmented Faces
The following steps use the Augmented Faces sample app to show you the critical tasks for configuring and building an app with Augmented Faces capabilities.
Get the Augmented Faces sample app
- Clone or download the latest ARCore SDK for iOS from GitHub to obtain the sample app code.
- Open a Terminal or Finder window and navigate to the folder where you cloned or downloaded the SDK.
- You can find the sample app code in
The AugmentedFacesSample app ships with a Podfile preconfigured with the ARCore SDK and iOS versions that you'll need. To install these dependencies:
- Open a Terminal window and run
pod installfrom the folder where the Xcode project exists. This generates an
*.xcworkspacefile that you'll use later to build and run the app.
Build and run the app
- Open the
.xcodeproj) file for the project in Xcode.
- Open the target’s general settings and change the bundle identifier.
- Hit Cmd+R or click Run. Use a physical device, not the simulator, to work with Augmented Faces. The device should be connected to your development machine via USB.
Try out the sample app
- Build and run the sample app from the
.xcworkspacefile to launch it on your device.
- If prompted, grant camera permissions to the app. ARCore then starts detecting faces visible in the camera. Currently, only once face is detected at a time.
- Make sure that the camera is facing the user.
- See the Augmented Faces Developer Guide for iOS to explore the sample app code and learn more about working with Augmented Faces in your own apps.
- Review the ARCore iOS API Reference.
- Watch the following video on how to get started with Augmented Faces on iOS: