3. What is “Spatial Computing” ?
Computing systems that augment ,extend or even replace our real world
experiences, providing information and functionalities that surround the user
in 3D space
4. • Virtual Reality (VR) aims to replace your real
world experience entirely with artificial
computer generated experiences
• “you are in the fantasy world ”
• Augmented or Mixed Reality (AR/MR) brings
the artificial computer generated experience in
to the real world
• “the fantasy world comes to your real world”
What is “Spatial Computing” ?
5. How does “Spatial Computing” work?
Spatial computing systems are not just a display technology, but complex and
computationally demanding “sensor fusion systems”
A lot of options for both input sources and output devices ,from headsets and
hand controllers to full coverage haptic suits and feedback systems
7. What is Vision pro
o This is an Augmented reality/Virtual reality
headset
o This has brunch of cameras and sensors that you
strap to your face
o This has M2 chip with R1 chip
o You don’t have to pair it with your iPhone or mac
in fact, it’s a standalone computer with its own
wi-fi
o Two hours battery life
8. Some features of Vision pro
o Apps have dimensions
o You are not isolated from other people..
(can see through the glasses with the help of cameras and they
can see you)
o Up-to 100ft wide screen
o Do facetime without letting them know that you are wearing VR
o Seamless App Integration
10. What is Vision OS
o Vision OS is the operating system designed specifically for Apple's mixed
reality headset, the Apple Vision Pro
o Apple’s first ever spatial operating system
o It works with only SwiftUI Framework
o The core of vision OS are Windows, Volumes and Spaces
Windows Volumes Spaces
11. Window is spacial version of window that we use in our
iPhone or Mac
12. Volume is where we can have 3d experience.
It defines width, height and depth
13. Spaces are where your windows and volumes can live side by
side (full space is where only one app is active either in window
or in volume or both in windows and volume)
14. o Apple Vison Pro headset allows the wearer to see the real or physical
world around them, unlike VR headsets that fully envelop the face and
limit visibility.
o Apple Vision Pro seamlessly blends digital content with your physical
space.
o You navigate simply by using your eyes, hands, and voice.
Tap to select Pinch to rotate Manipulate create custom
objects gestures
Vision Pro Gestures
16. A singular piece of three-dimensionally formed laminated glass flows into an aluminum
alloy frame that gently curves to wrap around your face.
ENCLOSURE CAMERAS & SENSORS CROWN & BUTTON DISPLAYS
17. A pair of high-resolution cameras transmit over one billion pixels per second to the
displays so you can see the world around you clearly. The system also helps
deliver precise head and hand tracking and real-time 3D mapping, all while
understanding your hand gestures from a wide range of positions.
ENCLOSURE CAMERAS & SENSORS CROWN & BUTTON DISPLAYS
18. Press the Digital Crown to bring up the Home View,
and turn it to control your level of immersion while
using Environments
Press the top button to take spatial videos and
spatial photos in the moment
ENCLOSURE CAMERAS & SENSORS CROWN & BUTTON DISPLAYS
19. A pair of custom micro - OLED displays deliver more pixels than a 4K TV to each eye - for stunning clarity.
ENCLOSURE CAMERAS & SENSORS CROWN & BUTTON DISPLAYS
20. Spatial Audio system
Ambient Spatial Audio makes sounds feel like they’re coming from your surroundings. And with audio raytracing,
Vision Pro analyzes your room’s acoustic properties - including the physical materials - to adapt and match sound
to your space.
21. Vision OS
In vision OS, apps can fill the space around you, beyond the
boundaries of a display.
They can be moved anywhere, scaled to the perfect size, react
to the lighting in your room, and even cast shadows.
22.
23. Navigate vision OS simply by
looking at apps, buttons, and text
fields. App icons and buttons
subtly come to life when you look
at them.
Simply look at the microphone
button in a search field and start
speaking to dictate text.
Tap your fingers together to make
a selection and gently flick to
scroll. It is designed to understand
hand gestures from comfortable
positions, like resting in your lap or
on the sofa.
Use Siri to quickly
open or close
apps, play media,
and more.
24. Responsive, precision eye tracking
A high-performance eye tracking system of LEDs and infrared cameras projects invisible light patterns onto
each eye.
This system provides input without your needing to hold any controllers, so you can select elements just by
looking at them.
26. o A unique dual-chip design enables the spatial experiences on Vision Pro.
o The powerful M2 chip simultaneously runs vision OS, executes advanced computer vision algorithms, and
delivers stunning graphics, all with incredible efficiency.
o The brand-new R1 chip is specifically dedicated to process input from the cameras, sensors, and
microphones, streaming images to the displays within 12 milliseconds — for a virtually lag-free, real-time
view of the world.
27. Apple Vision Pro covers a combination of deep learning and image
processing techniques to achieve its advanced visual recognition
capabilities.
•Deep Learning Models: At the core are deep learning models, particularly Convolutional Neural
Networks (CNNs) trained on massive datasets of images and videos. CNNs excel at extracting
features from visual data, enabling Apple Vision Pro to perform tasks like object recognition, image
classification, and object tracking with high accuracy.
•Image Processing Techniques: Image processing algorithms likely play a role in enhancing
the quality of the visual data. Techniques like noise reduction, contrast improvement, and edge
detection can be used to improve the accuracy of subsequent analysis by the deep learning models.
28. Convolutional Neural Network — CNN (Deep
Learning)
1.Preprocessing: The image captured by the device's camera goes through initial processing. This
might involve resizing, color normalization, or other techniques to ensure consistency for the CNN.
2.Convolution: The image is fed into the convolutional layers. These layers apply filters that scan
the image, extracting features like edges, shapes, and colors. Each filter detects specific patterns,
and multiple filters are used to capture various features. The resulting data is called feature maps.
3.Pooling: Pooling layers down sample the feature maps, reducing their dimensionality without
losing important information. This helps control computational complexity and makes the model
more robust to minor variations in the image.
29. 4.Activation: Activation functions are applied to the data after each convolution and pooling step.
These functions introduce non-linearity, allowing the network to learn complex relationships between
features.
5.Fully Connected Layers: After multiple convolutional and pooling stages, the extracted features
are fed into fully connected layers. These layers act like traditional artificial neural networks,
connecting all the neurons from the previous layer to every neuron in the next. Here, the high-level
features are combined, and classifications or detections are made based on the learned patterns.
6.Output: The final layer outputs the results, which could be class probabilities (identifying an object
type) or bounding boxes (localizing objects in the image) depending on the specific task Apple Vision
Pro is performing.
30.
31. Image processing
techniques
Apple utilizes a combination of image processing techniques to prepare the
visual data for their deep learning models in Apple Vision Pro
o Noise Reduction: Digital cameras can introduce noise into images. Noise
reduction techniques can help eliminate this unwanted noise, leading to
cleaner data for the CNNs to analyze.
o Contrast Enhancement: Adjusting image contrast can make features like
edges and objects more distinct. This can improve the accuracy of object
detection and recognition.
o Color Correction: Inconsistent lighting or white balance issues can affect
image quality. Color correction techniques can help normalize the color
balance, ensuring the CNNs process consistent color information.
32. o Sharpening: Sharpening techniques can enhance the definition of edges and
details in an image. This can be particularly helpful for tasks like object
recognition where precise feature extraction is crucial.
o Image Segmentation: This technique can separate objects from the
background in an image. This pre-processed data can be beneficial for the
CNNs to focus on specific objects of interest.
o Eye Tracking: Apple Vision Pro uses high-speed cameras to track eye
movements. This requires image processing techniques to identify and track
the user's pupils in real-time.
33. The LiDAR Scanner and TrueDepth camera work together to
create a fused 3D map of your surroundings
34. Vision Pro makes it easy to collaborate and
connect in facetime wherever you are.
Within FaceTime, you can also use apps to
collaborate with colleagues on the same
documents simultaneously.
Vision Pro is Apple’s first 3D camera.
You can capture magical spatial photos
and spatial videos in 3D, then relive
those cherished moments like never
before with immersive Spatial Audio.
36. Your favorite movies and TV
shows from Apple TV+ and
other streaming services are
available on Vision Pro.
Vision OS offers to Play your favorite
iPad games using Bluetooth game
controllers.
37. Advantag
es
Immersive Experience: Users feel like they are part of the action.
Innovative Design: It features a comfortable fit and intuitive controls for
effortless use.
Wide Range of Applications: valuable tool in various sectors, from
professional training and education to enhanced creativity and productivity.
Apple Ecosystem Integration.
High-Quality Display.
38.
39. Disadvantag
es
• High Cost
• Battery Life: The battery life might not meet everyone's needs.
• Discomfort and neck strain: The weight, especially the front-
heavy design, can cause fatigue and discomfort after a while.
• Limited usability: The weight makes it impractical for activities
that require a lot of movement, like VR workouts.
• Privacy Concerns
• Limited Ecosystem Integration
40. Conclusion
o In conclusion, Apple Vision Pro, powered by a combination of cutting-edge computer vision algorithms
and a dedicated operating system that leads to a new era of human-computer interaction.
o Its capabilities for real-time object recognition, spatial computing, and immersive experiences have the
potential to revolutionize various fields, from entertainment and gaming to education, design, and even
healthcare. As Apple continues to refine Vision Pro, we can expect even more innovative applications
that blur the lines between the physical and digital worlds.
41. “This is the prototype of something big in Future”