Pages: [1]   Go Down
  Print  
Author Topic: These Gestures Are How You Control Apple Vision Pro  (Read 188 times)
HCK
Global Moderator
Hero Member
*****
Posts: 79425



« on: June 11, 2023, 04:05:03 pm »

These Gestures Are How You Control Apple Vision Pro

Apple Vision Pro, Apple's new "spatial computing" device, does not have a hardware-based control mechanism. It relies on eye tracking and hand gestures to allow users to manipulate objects in the virtual space in front of them. In a recent developer session, Apple designers outlined the specific gestures that can be used with Vision Pro, and how some of the interactions will work.



" width="3000" height="1687" class="aligncenter size-full wp-image-899457

<ul>

    <li><strong>Tap</strong> - Tapping the thumb and the index finger together signals to the headset that you want to tap on a virtual element on the display that you're looking at. Users have also described this as a pinch, and it is the equivalent to tapping on the screen of an iPhone.</li>

    <li><strong>Double Tap</strong> - Tapping twice initiates a double tap gesture.</li>

    <li><strong>Pinch and Hold</strong> - A pinch and a hold is similar to a tap and hold gesture, and it does things like highlighting text.</li>

    <li><strong>Pinch and Drag</strong> - Pinching and dragging can be used to scroll and to move windows around. You can scroll horizontally or vertically, and if you move your hand faster, you'll be able to scroll faster.</li>

    <li><strong>Zoom</strong> - Zoom is one of two main two-handed gestures. You can pinch your fingers together and pull your hands apart to zoom in, and presumably zooming out will have a pushing sort of motion. Window sizes can also be adjusted by dragging at the corners.</li>

    <li><strong>Rotate</strong> - Rotate is the other two-handed gesture and based on Apple's chart, it will involve pinching the fingers together and rotating the hands to manipulate virtual objects.</li>

</ul>

Gestures will work in tandem with eye movements, and the many cameras in the Vision Pro will track where you are looking with great accuracy. Eye position will be a key factor in targeting what you want to interact with using hand gestures. As an example, looking at an app icon or on-screen element targets it and highlights it, and then you can follow up with a gesture.



Hand gestures do not need to be grand, and you can keep your hands in your lap. Apple is encouraging that, in fact, because it will keep your hands and arms from getting tired from being held in the air. You only need a tiny pinch gesture for the equivalent of a tap, because the cameras can track precise movements.



<div class="center-wrap"><blockquote class="tiktok-embed" cite="https://www.tiktok.com/@macrumors/video/7242130218226552110" data-video-id="7242130218226552110" style="max-width: 605px;min-width: 325px;" > <section> <a target="_blank" title="@macrumors" href="https://www.tiktok.com/@macrumors?refer=embed">@macrumors[/url] Eye tracking and pinching is the interface on the new <a title="applevisionpro" target="_blank" href="https://www.tiktok.com/tag/applevisionpro?refer=embed">#AppleVisionPro[/url] and we could to try jt out! <a title="apple" target="_blank" href="https://www.tiktok.com/tag/apple?refer=embed">#Apple[/url] <a title="vr" target="_blank" href="https://www.tiktok.com/tag/vr?refer=embed">#VR[/url] <a title="ar" target="_blank" href="https://www.tiktok.com/tag/ar?refer=embed">#AR[/url] <a target="_blank" title="♬ original sound - MacRumors" href="https://www.tiktok.com/music/original-sound-7242130214243142446?refer=embed">♬ original sound - MacRumors[/url] </section> <script async src="https://www.tiktok.com/embed.js"></script></div>

What you're looking at will let you select and manipulate objects that are both close to you and far from you, and Apple does anticipate scenarios where you might want to use larger gestures to control objects that are right in front of you. You can reach out and use your fingertips to interact with an object. For example, if you have a Safari window right in front of you, you can reach your hand out and scroll from there rather than using your fingers in your lap.



In addition to gestures, the headset will support hand movements such as air typing, though it doesn't seem like those who have received a demo have been able to try this feature as of yet. Gestures will work together, of course, and to do something like create a drawing, you'll look at a spot on the canvas, select a brush with your hand, and use a gesture in the air to draw. If you look elsewhere, you'll be able to move the cursor immediately to where you're looking.



While these are the six main system gestures that Apple has described, developers can create custom gestures for their apps that will perform other actions. Developers will need to make sure custom gestures are distinct from the system gestures or common hand movements that people might use, and that the gestures can be repeated frequently without hand strain.



To supplement hand and eye gestures, Bluetooth keyboards, trackpads, mice, and game controllers can be connected to the headset, and there are also voice-based search and dictation tools.



Multiple people who have been able to try the Vision Pro have had the same word to describe the control system - intuitive. Apple's designers seem to have created it to work similarly to multitouch gestures on the &zwnj;iPhone&zwnj; and the iPad, and so far, reactions have been positive.



MacRumors videographer Dan Barbera was able to try out the headset and he was impressed with the controls. You can see his full overview of his experience on our YouTube channel.<div class="linkback">Related Roundup: Apple Vision Pro</div><div class="linkback">Related Forum: Apple Vision Pro</div>
This article, &quot;These Gestures Are How You Control Apple Vision Pro&quot; first appeared on MacRumors.com

Discuss this article in our forums



Source: These Gestures Are How You Control Apple Vision Pro
Logged
Pages: [1]   Go Up
  Print  
 
Jump to: