AI meets AR: How Rokid Glasses bring multilingual, real-time intelligence to smart eyewear globally
Updated
March 3, 2026 3:50 PM

Rokid's smart glasses. PHOTO: ROKID
Rokid, a Chinese company specializing in AI-powered smart eyewear and human–computer interaction, has rolled out a major software update for the international version of its Rokid Glasses. This update makes it the first smart glasses manufacturer to natively support Google’s Gemini, alongside three other leading large language models: OpenAI’s ChatGPT, Alibaba’s Qwen and DeepSeek.
The integration is powered by Rokid’s device-to-cloud architecture, which enables users to switch between AI models on the fly. In practice, this means a traveler can receive a real-time translation in Japanese using one AI model, then quickly switch to ChatGPT to answer a technical query—without noticeable delay. The system also supports multi-modal inputs like voice and gestures, making interactions more intuitive for everyday use.
This is more than a routine software update. By combining AI models from both U.S. and Chinese developers, Rokid is making its smart glasses relevant to global users, with features that adapt to local languages and preferences while maintaining high performance.
These technological advancements have directly fueled Rokid’s international growth. Between November 2024 and October 2025, Shangpu Group data shows Rokid Glasses ranked No.1 in global sales for AI glasses with display functionality. Crowdfunding milestones further reflect this momentum: the product became the fastest smart glasses to raise over 100 million Japanese Yen on Japan’s MAKUAKE platform and broke Kickstarter records for smart eyewear.
Taken together, Rokid’s update highlights a shift in the smart glasses space: success increasingly comes from openness, flexibility and localized AI experiences rather than closed, single-platform ecosystems. By giving users choice, integrating global AI capabilities and bridging cultural and linguistic gaps, Rokid is positioning itself as a serious contender in the international AR and AI wearable market.
Keep Reading
The focus is no longer just AI-generated worlds, but how those worlds become structured digital products
Updated
February 20, 2026 6:50 PM

The inside of a pair of HTC VR goggles. PHOTO: UNSPLASH
As AI tools improve, creating 3D content is becoming faster and easier. However, building that content into interactive experiences still requires time, structure and technical work. That difference between generation and execution is where HTC VIVERSE and World Labs are focusing their new collaboration.
HTC VIVERSE is a 3D content platform developed by HTC. It provides creators with tools to build, refine and publish interactive virtual environments. Meanwhile, World Labs is an AI startup founded by researcher Fei-Fei Li and a team of machine learning specialists. The company recently introduced Marble, a tool that generates full 3D environments from simple text, image or video prompts.
While Marble can quickly create a digital world, that world on its own is not yet a finished experience. It still needs structure, navigation and interaction. This is where VIVERSE fits in. By combining Marble’s world generation with VIVERSE’s building tools, creators can move from an AI-generated scene to a usable, interactive product.
In practice, the workflow works in two steps. First, Marble produces the base 3D environment. Then, creators bring that environment into VIVERSE, where they add game mechanics, scenes and interactive elements. In this model, AI handles the early visual creation, while the human creator defines how users explore and interact with the world.
To demonstrate this process, the companies developed three example projects. Whiskerhill turns a Marble-generated world into a simple quest-based experience. Whiskerport connects multiple AI-generated scenes into a multi-level environment that users navigate through portals. Clockwork Conspiracy, built by VIVERSE, uses Marble’s generation system to create a more structured, multi-scene game. These projects are not just demos. They serve as proof that AI-generated worlds can evolve beyond static visuals and become interactive environments.
This matters because generative AI is often judged by how quickly it produces content. However, speed alone does not create usable products. Digital experiences still require sequencing, design decisions and user interaction. As a result, the real challenge is not generation, but integration — connecting AI output to tools that make it functional.
Seen in this context, the collaboration is less about a single product and more about workflow. VIVERSE provides a system that allows AI-generated environments to be edited and structured. World Labs provides the engine that creates those environments in the first place. Together, they are testing whether AI can fit directly into a full production pipeline rather than remain a standalone tool.
Ultimately, the collaboration reflects a broader change in creative technology. AI is no longer only producing isolated assets. It is beginning to plug into the larger process of building complete experiences. The key question is no longer how quickly a world can be generated, but how easily that world can be turned into something people can actually use and explore.