PlayCanvasAR makes it easy to build lightning fast augmented reality applications. With it, you can construct web-based AR applications with zero programming using the PlayCanvas Editor. Just drag in your 3D models and watch them appear in AR!
But if you want to create something more interactive, you can call on the full power of the PlayCanvas scripting API.
Before you begin, print out the following markers: Hiro and Kanji.
Now click here for a demonstration. We recommend running it on mobile. If you use iOS, you'll need to ensure you have upgraded to iOS 11, which is required to access the camera from Safari.
Or check out a little YouTube video.
- Amazing Performance - PlayCanvasAR delivers 60 frames per second, even on mobile.
- Marker-based Tracking - PlayCanvasAR utilizes the amazing ARToolkit, an open source marker-based tracking library. Originally coded in C++, it has been ported to JavaScript using Emscripten.
- Incredible Visuals - PlayCanvas provides an advanced WebGL graphics engine. It supports the latest WebGL 2 graphics API and implements Physically Based Rendering to achieve incredible visuals for your AR applications. And yes, that's open sourced too.
- Clone this repository and navigate inside the
playcanvas-ar
directory you have just created. - Run
npm install
. This will download the requiredhttp-server
dependency so that you can easily run this on a local machine. - Run
npm run preinstall
to generate your SSL keys. These keys are not trusted by most browsers but this is okay for development. NOTE: In order for WebRTC to access your camera(s) on a mobile device, an SSL certificate must be installed. - Run
npm run start
and navigate tohttps://localhost:8081
in your browser.
The PlayCanvas Editor allows you to build your AR apps visually. But you may want to create AR powered entities programmatically. Or you may want to build AR apps by simply using the PlayCanvas Engine without the Editor. PlayCanvas AR exposes two script objects: 'arCamera' and 'arMarker'. 'Engine-only' scripting examples can be found in the examples folder.
This code creates an AR-enabled camera:
var camera = new pc.Entity("AR Camera");
camera.addComponent("camera", {
clearColor: new pc.Color(0, 0, 0, 0)
});
camera.addComponent("script");
camera.script.create("arCamera", {
attributes: {
cameraCalibration: asset,
detectionMode: 0, // Color Template
matrixCodeType: 0, // 3x3
labelingMode: 1, // Black Region
processingMode: 0, // Frame
thresholdMode: 0, // Manual
threshold: 100,
trackerResolution: 0, // Full
trackAlternateFrames: false,
debugOverlay: false,
videoTexture: false
}
});
app.root.addChild(camera);
Attribute | Type | Description |
---|---|---|
cameraCalibration | pc.Asset | Data file containing the calibration properties for the camera to be used. |
detectionMode | Number | The pattern detection determines the method by which ARToolKit matches detected squares in the video image to marker templates and/or IDs. ARToolKit can match against pictorial "template" markers, whose pattern files are created with the mk_patt utility, in either colour or mono, and additionally can match against 2D-barcode-type "matrix" markers, which have an embedded marker ID. Two different two-pass modes are also available, in which a matrix-detection pass is made first, followed by a template-matching pass. Defaults to 0 (Color Template). |
matrixCodeType | Number | Set the size and ECC algorithm to be used for matrix code (2D barcode) marker detection. When matrix-code (2D barcode) marker detection is enabled (see Detection Mode) then the size of the barcode pattern and the type of error checking and correction (ECC) with which the markers were produced can be set via this function. This setting is global to a given AR Camera; It is not possible to have two different matrix code types in use at once. Defaults to 0 (3x3). |
labelingMode | Number | Select between detection of black markers and white markers.\n\nARToolKit's labelling algorithm can work with both black-bordered markers on a white background ('Black Region') or white-bordered markers on a black background ('White Region'). This property allows you to specify the type of markers to look for. Note that this does not affect the pattern-detection algorithm which works on the interior of the marker. |
processingMode | Number | When the image processing mode is 'Frame', ARToolKit processes all pixels in each incoming image to locate markers. When the mode is 'Field', ARToolKit processes pixels in only every second pixel row and column. This is useful both for handling images from interlaced video sources (where alternate lines are assembled from alternate fields and thus have one field time-difference, resulting in a 'comb' effect) such as Digital Video cameras. The effective reduction by 75% in the pixels processed also has utility in accelerating tracking by effectively reducing the image size to one quarter size, at the cost of pose accuracy. |
thresholdMode | Number | The thresholding mode to use. The standard ARToolKit options are available: Manual (0), Median (1), Otsu (2), Adaptive (3). Defaults to 0 (Manual). |
threshold | Number | The binarization threshold is an 8-bit number that is in the range [0, 255], inclusive. The default value is 100, allowing ARToolKit to easily find markers in images that have good contrast. This value is only used when the mode is set to Manual. Defaults to 100. |
trackerResolution | Number | Controls the resolution of the tracker image. Each video frame is copied to the tracker image for marker detection. Reducing the tracker image resolution will speed up marker detection but will also make it less precise. For example, a video camera source may have a resolution of 640x480. The tracker image will have the following resolutions based on the selected option: 'Full': 640x480, 'Three Quarters': 480x360, 'Half': 320x240, 'Quarter': 160x120. |
trackAlternateFrames | Boolean | If selected, tracking is only performed on every other update. This can increase lag in tracking but will reduce CPU load. |
debugOverlay | Boolean | Enables or disables the debug overlay. When enabled, a black and white debug image is generated during marker detection. The debug image is useful for visualizing the binarization process and choosing a threshold value. The image is displayed as an overlay on top of the 3D scene. |
videoTexture | Boolean | Streams the camera feed to a video texture if enabled. Otherwise, a video DOM element is used. Defaults to false. |
This code creates an AR marker entity:
var hiro = new pc.Entity("Hiro Marker");
hiro.addComponent("script");
hiro.script.create("arMarker", {
attributes: {
pattern: asset,
matrixId: 0,
width: 1,
deactivationTime: 0.25,
shadow: true,
shadowStrength: 0.5
}
});
app.root.addChild(hiro);
Attribute | Type | Description |
---|---|---|
pattern | pc.Asset | The marker pattern to track. This can be the Hiro or Kanji markers or a marker you have generated yourself. |
matrixId | Number | The matrix ID. If no pattern template is set, the marker is a matrix. Defaults to 0. |
width | Number | The width of the marker. Defaults to 1. |
deactivationTime | Number | The time in seconds from when a marker is lost before its children are deactivated. Defaults to 0.25. |
shadow | Boolean | Enable this option to generate shadows in the plane of the marker that blend with the camera feed. Defaults to true. |
shadowStrength | Number | Control the strength of the shadow. 1 is full strength and 0 is disabled. Defaults to 0.5. |
- Android 4.0+ (Chrome 21+)
- iOS 11+ (Safari 11+, Chrome 21+, Firefox 17+)
- Windows XP+ (Chrome 21+, Firefox 17+, Opera 18+, Edge 12+)
- macOS (Safari 11+, Chrome 21+, Firefox 17+, Opera 18+)
- Linux (Chrome 21+, Firefox 17+, Opera 18+)
AR is fun, so why not get involved? Find a bug? Need a feature? Want to contribute something? Add an issue!