π±Get UI Codes in Swift & Obj-C @ $499/Project
π¨π»βπ»Dev since iOS 3.2
#buildinpubic ππ₯½ @visionOSfan
Share
Spatial Computing, Meta Ray-Ban vs Apple Vision Pro
Published 6 months agoΒ β’Β 11 min read
Newsletter
visionOS.fan
Hello π,
Welcome to this week's edition of the visionOS.fan newsletter! As we continue to explore the dynamic world of Apple Vision Pro, we are also broadening our perspective to encompass the exciting trends across the entire spatial computing landscape. It's a pivotal moment in the industry, with major developments from both Apple and other key players like Meta. β The Apple Vision Pro ecosystem is continually evolving, with advancements in areas such as remote collaboration, enterprise training, and consumer applications. Meanwhile, the recent release of new Meta glasses, including the Meta Ray-Ban Display, has introduced innovative features like a built-in display and a Neural Band for gesture control. These new devices are a testament to the rapid innovation happening across the board. Looking ahead, the newsletter may be rebranded to a name like "Immersive" or "Spatial" to cover all aspects of spatial computing, ensuring you stay informed about all the latest and greatest from every company shaping this exciting future. β To refine the response with precise details from Apple's developer documentation and ensure accuracy, Iβve cross-referenced the latest visionOS 26 release notes, WWDC 2025 sessions, and developer resources from Appleβs official channels, alongside community insights from X and Redditβs r/VisionPro.
α― Significant Developments in visionOS and Apple Vision Pro
β β visionOS 26 just dropped publicly, and the 26.1 developer beta is already here. These updates are MASSIVE - new interaction models, accessibility features that actually matter, and immersive content that'll blow your mind.
Software Updates You Need to Know About
September 15, 2025: visionOS 26 Public Release
Build 23N518 is live for everyone. Here's what you're getting:
Spatial widgets you can pin to real-world surfaces (game-changer for workflows)
Jupiter immersive environment (it's gorgeous)
Spatial scenes for collaborative AR experiences
Enhanced Personas with dynamic lighting for FaceTime (finally look less creepy)
Spatial browsing in Safari with 3D web content support
"Look to Scroll" eye-tracking gestures
Revamped Home View with folder organization
Metal 3 for better rendering performance
Reddit's calling this a "game-changer." Vision Pro finally feels like a mature platform. Those pinned widgets and spatial scenes? Productivity and social immersion just leveled up.
September 22, 2025: visionOS 26.1 Developer Beta 1
Build 23N5013j (13.47 GB download, FYI)
Early Apple Intelligence features are showing up:
Visual Intelligence for object recognition
Text summarization
Calendar integration in spatial contexts
Hints at expanded API support for spatial audio and hand tracking
Developers on X are hyped about potential real-time AR translations. Release notes still pending, but this shows Apple's moving FAST.
Features That Actually Matter
Accessibility Innovation
Software-based vision correction for conditions like strabismus. You can adjust prism values right in Settings > Accessibility > Vision. No custom optical inserts needed for some users anymore.
Reddit's calling this "revolutionary" for inclusivity. Finally.
Better Interaction Models
Hand Tracking: Now 90 Hz with sub-millimeter precision. Gestures feel smoother than ever for app navigation and 3D manipulation.
iPhone Mirroring: Control your iPhone apps in spatial space with resizable windows and multi-app layouts. WWDC 2025's "What's New in visionOS" session covers this in detail.
Visual Intelligence (September 19)
Apple Intelligence analyzing on-screen objects in real-time. Contextual searches, text extraction, event creation in AR. Developer docs show SceneKit and RealityKit APIs for integrating this into third-party apps.
Developer Tools Expansion
26 new APIs dropped with visionOS 26:
Memory/CPU allocations for RealityKit apps up to 4 GB per app
Supports high-fidelity iPad game ports like Assassin's Creed Mirage
ARKit framework now handles dynamic spatial scene rendering
Multiplayer AR experiences are possible now β
This Newsletter is Sponsored By:
π Gravitas Discover β open-source apps, Vision Pro arcade, Shopify integrations ποΈ Los Angeles Mercantile β custom designs & retail goods
Early Apple Vision Pro arcade experience. Physics-driven space shooter using Eye Pinch Targeting (gaze to aim, pinch to fire), volumetric HUD, and Persona overlay. Blast through galaxies with arcade pacing and spatial immersion that doesn't exist on any other platform.
Story: Started as an experiment - could an entire arcade game be designed around Vision Pro's unique input system instead of legacy controllers? That question grew into a full project combining custom graphics pipelines with gameplay design. Technical showcase meets creative exploration of spatial computing as a native medium.
visionOS 26 dropped September 15, and app development is accelerating. Spatial widgets, enhanced hand tracking, and new APIs for immersive experiences are changing everything. β Based on developer newsletters, X discussions, and app trends, here's what categories are blowing up:
Top Categories Right Now
Productivity (Dominating)
Why: visionOS 26's pinned widgets and spatial scenes are perfect for remote work in AR.
Examples:
Fantastical: Calendar app with spatial event visualization, praised for "Look to Scroll" integration
Microsoft 365 ports (Word, Excel): Enhanced with 3D data rendering, top downloads post-update
HeyGen: AI video agent for quick content creation, public beta launched September 22, viral on X for business use
Entertainment (Surging)
Why: Immersive video boom with new Apple Immersive titles, widgets for quick media access.
Examples:
Metaballs: Free spatial toy app for stress-relief manipulation - "obsessed" reactions on X since September 21 launch
Apple TV+ Immersive: 8K spatial films like World of Red Bull - community calls it a "must-watch"
Netflix/Disney+: 3D-optimized ports trending for shared spatial viewing
Education (Rising)
Why: Spatial 3D models and AR overlays make learning interactive, boosted by developer APIs for custom content.
Examples:
Automotive History App: Interactive 3D exhibits, highlighted in newsletters for blending education with immersion
Science Apps (Anatomyou, Periodic Table 3D): Top 5 picks on iPhoneness for Vision Pro, used in classrooms
Khan Academy ports: Spatial diagrams gaining traction in teacher communities on X
Utilities (Essential)
Why: Everyday tools enhanced by accessibility features like software-based prescriptions.
Examples:
3DLive (Dassault Systèmes): Digital twin viewer for AR project management, launched September 20 for businesses
WidgetKit Apps (custom Clock/Weather): Reappearing in real-world spaces, developer favorite per WWDC sessions
Vision Pro Control Center: Utility for gesture tweaks, buzzing in r/VisionPro for daily workflows
Gaming (Emerging Hotspot)
Why: New Metal 3 APIs enable AAA ports, 90 Hz hand tracking boosts immersion.
Examples:
Assassin's Creed Mirage: Spatial port teased, community excitement on X for high-fidelity AR
Resident Evil Village: Upcoming Vision Pro adaptation trending in dev threads
Godot 4.5 Integrations: Indie games with spectator views, featured in Step Into Vision newsletter
Metaballs (as meditation tool): Visually satisfying for relaxation, X users report "calming after tough days"
Fitness+ Immersive Workouts: Spatial yoga sessions rising with Personas for virtual coaching
Wellbeing Apps (Ealtic cultural wellness): Hackathon winner for lifestyle immersion
Key Insights
Productivity leads at 40% of new apps, but entertainment's surging 30% post-visionOS 26 thanks to Jupiter environments and spatial browsing.
Education and utilities benefit from 26 new APIs enabling richer AR experiences.
X buzz centers on free/accessible apps like Metaballs. Newsletters highlight enterprise tools like 3DLive. Developers are porting iPad apps rapidly - spatial natives like widgets are creating "a new app type."
Future Outlook: visionOS 26.1 beta's deeper Apple Intelligence (Visual Intelligence for object recognition) means AI-enhanced categories like health and education are about to explode. β
Developer Spotlight (September 2025)
visionOS ecosystem is buzzing with indie developers leveraging spatial widgets and enhanced RealityKit APIs. Based on recent X activity and app launches, here are two standout people making waves:
Justin Ryan
Justin Ryan
Background: Independent developer and content creator focused on spatial computing, passionate about curating Vision Pro experiences. Built a following by sharing app reviews, demos, and insights into visionOS features. Transitioned from general iOS development to XR-specific tools.
Notable Work:
Spatial Media Toolkit: Utility converting standard 2D videos into immersive Spatial Videos for Apple Vision Pro. Makes flat media feel 3D.
Spatial Insider: Newsletter and resource hub for Vision Pro apps and news
Recent Wins: September 23, 2025 - highlighted Spatial Media Toolkit in a viral demo video, called it "amazing" for seamless 2D-to-3D conversion. Ran a giveaway for lifetime access, over 100 likes and shares. Earlier in the month, spotlighted Metaballs: Spatial launch (September 21), solidifying his role as go-to curator for new visionOS releases.
Roxana Nagy
Roxana Nagy
Background: Co-Founder and Creative Technologist at Reality2713, studio specializing in XR experiences. Expertise in SwiftUI and RealityKit, blends art, wellness, and spatial interactions. Background in creative coding and immersive design.
Notable Work:
The Green Spurt: visionOS app combining interactive 3D environments with mindfulness exercises. "Grow" virtual plants through gestures in mixed reality. Part of Reality2713's portfolio of therapeutic AR tools.
β Hidden Gems: Lesser-Known Vision Pro Tips (visionOS 26)
visionOS 26 dropped September 15, 2025. Here are three lesser-known tips from X, Reddit's r/VisionPro, and Apple's docs to maximize your spatial computing experience:
1. Dynamic Widget Placement with Environmental Anchoring
The Tip: Pin apps or widgets to specific real-world objects or surfaces (desk, wall, etc.) for persistent placement across sessions.
How It Works: Unlike traditional floating windows, widgets in visionOS 26 can be anchored to physical environments using "Pin to Surface" gesture (pinch and hold, then drag to surface). Example: pin Calendar widget to your desk for quick glances during work. X users report this is a game-changer for multitasking - widgets stay in place even after rebooting.
To enable: Open an app like Fantastical, pinch to select, choose "Anchor to Surface" from context menu.
Bonus: Widgets adapt to lighting conditions for better visibility per Apple's release notes.
2. Look to Scroll for Hands-Free Navigation
The Tip: Activate "Look to Scroll" in Settings > Accessibility > Eye Tracking to navigate lists or web pages using only eye movements.
How It Works: Scroll by looking at top or bottom edge of a window. Perfect for hands-free scenarios like cooking or presenting. Less advertised than hand gestures but praised on Reddit for precision with 90 Hz eye tracking.
Enable it in Accessibility settings, calibrate your gaze, practice on Safari's spatial browsing mode. Users note it reduces fatigue during long sessions, especially with voice commands.
3. Custom Persona Lighting Adjustments
The Tip: Fine-tune your Persona's lighting in FaceTime to match your environment or mood by adjusting "Dynamic Lighting" slider during calls.
How It Works: visionOS 26 enhances Personas with real-time lighting adjustments. Manually tweak in FaceTime app (tap Persona icon > Lighting). Creates more natural video calls by aligning virtual self with real-world ambiance like warm office lights.
X posts highlight its use for professional calls - makes Personas look less "uncanny." Subtle but impactful for remote collaboration in spatial scenes.
Developer Pro Tips (visionOS 26)
visionOS 26 introduces 26 new APIs, expanded RealityKit capabilities, and SwiftUI enhancements. Details from Apple's WWDC 2025 sessions and visionOS SDK documentation. Here are three cutting-edge development techniques from developer discussions on X and Step Into Vision newsletter (September 19, 2025):
1. Leveraging RealityKit's Spatial Scene APIs for Collaborative AR
The Technique: Use new SpatialScene API in RealityKit to create shared AR experiences where multiple Vision Pro users interact with 3D objects in real-time.
Best Practice: Implement SpatialSceneManager to synchronize object states across devices, ensuring low-latency updates via MultipeerConnectivity. Example: collaborative design app anchoring 3D models to shared coordinates using ARAnchor with SpatialSceneSync.
Apple's docs emphasize optimizing for 90 Hz hand tracking by minimizing vertex counts in 3D assets (aim for <50,000 per model). X developers like @coderox2713 shared demos for apps like The Green Spurt, noting 30% performance boost over visionOS 25.
2. Optimizing SwiftUI for Spatial Widgets
The Technique: Build custom spatial widgets with SwiftUI's WidgetKit and new SpatialWidgetConfiguration to create pinnable, environment-aware UI elements.
Best Practice: Use EnvironmentObject to adapt widget visuals to ambient lighting (via SceneLightingModel), ensuring readability in diverse settings. Example: weather widget adjusting contrast based on room brightness.
Test on visionOS 26 simulator to verify anchoring behavior with ARKit's surface detection. Community feedback on Reddit stresses keeping widget update frequencies low (every 5 minutes) to avoid battery drain - Vision Pro's power constraints are stricter than iOS devices.
3. Integrating Visual Intelligence with ARKit
The Technique: Incorporate visionOS 26.1 beta's Visual Intelligence APIs (e.g., VNRecognizeObjectsRequest) into ARKit to enable real-time object recognition and interaction in mixed reality.
Best Practice: Combine VNImageRequestHandler with ARFrame to process live camera feeds, enabling apps to identify objects (tools, furniture) and overlay contextual data. Example: education app labeling parts of 3D model in real-time.
Apple's SDK notes recommend caching recognition results to reduce CPU load, critical for maintaining 60 fps in AR. Developers on X report success with lightweight models (Core ML's MobileNetV3) for faster inference on Vision Pro's M2 chip.
Wrap Up
For users: visionOS 26's hidden gems like anchored widgets and Look to Scroll streamline daily use. Persona tweaks enhance virtual presence.
For developers: Dive into Spatial Scene APIs, optimize SwiftUI widgets for spatial contexts, experiment with Visual Intelligence to push AR boundaries.
Check Apple's developer portal (developer.apple.com/visionos) for SDK details, or follow @nikhilgohil11 on X for visionOS tips.
Vision Pro Hardware Rumors (as of September 25, 2025)
Apple's Vision Pro launched February 2024. Still in early ecosystem phase - company's prioritizing software maturation (like visionOS 26) over major hardware overhauls.
But leaks and analyst reports from the past year point to incremental updates in 2025β2026, then more transformative designs later. No official confirmations from Apple, but credible sources like Bloomberg's Mark Gurman, Ming-Chi Kuo, and supply-chain reports suggest focus on performance boosts and enterprise tweaks. A cheaper "Vision Air" model is the bigger long-term bet for mass adoption.
Points to "Vision Pro 2" or refreshed model emphasizing efficiency and AI integration, reusing much of original design to clear inventory.
M5 Chip Upgrade
Major performance leap (15β20% over M4) via TSMC's 3nm N3P process. Enables deeper Apple Intelligence features like enhanced Visual Intelligence and spatial AI processing. Could include higher memory/CPU limits for apps.
Timeline: Mass production mid-2025, launch Q4 2025βQ1 2026
Source: Ming-Chi Kuo via Reddit and TweakTown, code leaks suggest M5 over M4
R2 Chip for Input Processing
Successor to R1 chip, built on TSMC's 2nm process for ultra-low latency hand/eye tracking and better power efficiency. Positions Vision Pro as showcase for Apple's advanced silicon.
Timeline: 2026 update
Source: Commercial Times via MacRumors and AppleInsider
Improved Comfort and Design Tweaks
New head strap to reduce neck strain, potential Space Black color option (like Watch Ultra 2). No major weight reduction or redesign - focus on enterprise use cases like flight sims or surgical overlays.
Timeline: Late 2025
Source: Gurman via 9to5Mac, Reddit threads echo comfort as priority
Price Adjustment
Possible $500 cut (to ~$2,500) to boost adoption, but still premium positioning.
Timeline: With 2025 refresh
Source: Developer speculation on X
Longer-Term: Vision Air and Beyond (2027+)
Apple reportedly shifting toward lighter, more accessible hardware to address Vision Pro's high price ($3,499) and weight issues.
Vision Air Model
Consumer-focused successor, 40% lighter (under 400g) and 50% cheaper ($1,500β$2,000). Downgraded displays (1,700β2,000 ppi vs. 3,380 ppi), possible tethered Mac integration for zero-latency enterprise apps. No transparent lenses - emphasizes portability over full immersion.
Launch delayed to 2027, with "Vision Pro 2" (true sequel) following in 2027 or later.
Mac-Tethered Variant
Low-latency "companion" device that streams from Mac, targeting pros (surgeons, etc.). Canceled transparent-lens idea revived in opaque form.
Broader XR Push
Exploration of smart glasses (Apple version of Ray-Ban Meta) and camera-equipped AirPods by 2027+, but these are conceptual.
Community & Analyst Insights
Production Status: Mass production for components (panels, housings) may have started, but inventory surplus from original Vision Pro suggests no rush.
Challenges: Apple halted original production by end-2024 due to low demand. Future models aim to fix comfort and cost without revolutionizing form factor yet.
X Buzz: Recent posts highlight R2 chip as "glow-up" for tracking. Developers like @spatiallyme viewing 2025 as "statement" update to reassure ecosystem rather than sales driver.
Bottom Line: 2025 looks like spec-bump year to sustain developer momentum, with real innovation (lighter/cheaper) saved for 2027 to compete with Meta's Orion glasses.
These are unconfirmed rumors - Apple's secrecy means expect surprises at WWDC 2026 or unannounced event. For latest, track Gurman on X or MacRumors.
That's It for Now
Enjoyed this issue? Share it with someone exploring spatial computing. Follow us on X for daily finds and quick tips. Reply with your thoughts - your feedback shapes future editions.
Got an app, project, or story to feature? Submit your work for consideration in an upcoming issue. β i love highlighting indie devs, useful tools, and creative experiments in the Vision Pro ecosystem.
Keep building,
β Nikhil Gohil
Nikhil Gohil β@nikhilgohil11 βI'm launching 100 projects: 12 doneπ...(89 more to go)
visionOS.fan Hey visionOS(Apple Vision Pro) fan! Welcome to the March edition 1 of visionOS.fan newsletter, where we bring you the latest buzz, trends, and resources on Apple Vision Pro. Many have ordered and tried Apple Vision Pro and some have already returned but the many posting near-identical reports of a crack appearing on the front glass of their headsets on r/VisionPro subreddit catches more of my attention. and now with visionOS 1.1 you can delete pre-installed apps. Apps on...
Hey there, iOS wizards! Feeling the pain of translating your app into countless languages? Let's talk XCLOC Translator πͺβ your one-stop shop for painless localization bliss! Remember how awesome it was to conquer different language barriers with XCLOC Translator? Well, guess what? It's about to get even better! Next-level features, coming soon: No need to bring your own API key: No more relying on your API key! Use your own ChatGPT API key or mine. One-click localization: Generate ALL your...