Home > Article > Technology peripherals > Apple releases new Vision Pro headset, integrating gestures, eye movements and voice commands
News on June 7, one of Apple’s innovations is to provide a more intuitive interaction method, making the operation simple and easy to use, with almost no learning cost. However, the latest release of Vision The Pro headset integrates gestures, eye movements and voice commands, which increases the learning threshold and takes users some time to adapt.
According to the editor’s understanding, Apple’s recently launched Vision The Pro headset combines three interaction methods: gestures, eye movements and voice commands. Users can lock their viewing position with their eyes and use gestures to make selections. In addition, users can also use the microphone button to issue voice commands or record text.
The newly released Vision Pro headset has attracted widespread comments. Many media believe that users use these three interaction methods to operate Vision The Pro headset requires some adaptation time. In contrast, mainstream headsets reduce learning costs by holding controllers for the first time, while Apple's Vision Pro headset requires users to learn gestures to operate.
To complete the typing capabilities built into the Vision Pro headset, users can connect an iPhone or a Bluetooth keyboard. Additionally, Apple offers a virtual keyboard, as well as an alternative to transcribing text.
Overall, Apple’s Vision The Pro headset brings a new experience to users by integrating gestures, eye movements and voice commands. Although it takes some time to get used to it, this innovative interaction method is expected to further improve user convenience and experience.
The above is the detailed content of Apple releases new Vision Pro headset, integrating gestures, eye movements and voice commands. For more information, please follow other related articles on the PHP Chinese website!