Human Segmentation and Environment Occlusion support
ARKit Human Segmentation
The plugin keeps evolving and I'm slowly but surely running out of unsupported features. Today, I added the ARKit Human Segmentation support. And it even works in Windows Unity Editor!
Please check that your iOS device has A12 chip or newer and iOS 13+.
A video is below, after the environment occlusion section.
Environment Occlusion (preview)
This feature was added in AR Foundation 4.1 (currently in preview). In ARKit this feature requires a LiDAR (iPad 2020) and iOS 14.
But, to my surprise, on Android, it just works with a regular camera! It works slower than LiDAR, but the quality is great, good job Google!
Camera video with iOS device + Windows Editor
It turned out that a big number of people are developing iOS apps on Windows PCs. So this update fixes the camera background video in Windows Unity Editor.
Universal Render Pipeline support
I've broken the support for URP in version 3.11.3, sorry for that. Thanks to [USER=1534701]@LeoThePhoenix[/USER] [USER=4967593]@Scherabon[/USER] [USER=1455931]@daapage0[/USER] for reporting the bug, this helped me to fix it
Get Unity: AR Foundation Remote
Unity: AR Foundation Remote
Test and Debug your AR project in Editor
Status | Released |
Category | Tool |
Author | Kyrylo Kuzyk |
Tags | ar, arcore, ar-foundation, arkit, ar-testing, Augmented Reality, remote, simulator, Unity |
More posts
- Version 4.11.0Apr 16, 2021
- ARKit 2D and 3D Body Tracking support.Oct 15, 2020
- Meshing, ARKit Mesh Classification, and Anchors supportSep 14, 2020
- Connection and touch input improvementsJul 28, 2020
- Camera Background support in EditorJul 05, 2020
- ARRaycastManager support addedJun 22, 2020
- Image Tracking support addedJun 16, 2020
- ARKit Face Tracking Blendshapes supportJun 08, 2020
- Face Tracking addedJun 08, 2020
Leave a comment
Log in with itch.io to leave a comment.