Vizrt shows how live video can be produced anywhere, without complex studio setups
Vizrt, a media technology company, has introduced a new AI-powered tool to simplify the creation of virtual scenes in live production. Its latest release, the AI Keyer, is built around a simple idea: remove the need for green screens and make virtual production possible in almost any environment.
Traditionally, creating virtual backgrounds or augmented reality (AR) scenes requires controlled studio setups, green screens, precise lighting and skilled operators. That makes high-end visual production expensive and difficult to scale, especially for smaller teams or live, on-the-ground reporting.
The AI Keyer is designed to address that gap. It uses AI trained on real-world footage to identify people in a frame and separate them from the background in real time. This allows production teams to replace backgrounds, insert AR graphics or place presenters into virtual environments—whether they are indoors, outdoors or on location.
"Creating XR environments typically demands large infrastructure investments and requires specialized skills for daily operations. The Vizrt AI Keyer removes all these constraints, so high-quality virtual scenes and AR graphics become a reality for live productions of every size", says Edouard Griveaud, Senior Product Manager at Vizrt.
In practical terms, this means a presenter can appear in a different location without moving, a remote speaker can be placed inside a virtual event space or branded graphics can be added to live interviews without a complex setup. The system works without chroma keying, reducing both preparation time and production overhead.
This shift also reflects how the company is approaching AI more broadly. Instead of treating it as a background feature, Vizrt is positioning AI as a core part of the content creation and delivery process.
"AI is transforming the world, and the creative industries are no exception. At Vizrt, we have been on this journey for years, embedding intelligence into our solutions, empowering storytellers and delivering real, measurable impact for our customers", says Rohit Nagarajan, CEO of Vizrt. "That is not a vision for tomorrow. That is happening today. The Vizrt AI Keyer is the latest proof point of our relentless commitment to innovation. Putting breakthrough technology in the hands of every creative, at every level, everywhere in the world".
Beyond the product itself, the direction is clear. By removing the need for green screens and complex setups, tools like the AI Keyer make it easier to produce high-quality visual content in more flexible settings. The result is a production model that is less tied to physical studios and more adaptable to real-world environments, where content can be created and adjusted in real time.
A new approach examines how individual cells respond to drugs, aiming to identify risks earlier in development.
DeepCyte, a startup in the drug development space, is focusing on a long-standing problem: why drugs that appear safe in early testing still fail in clinical trials or are withdrawn later due to toxicity. DeepCyte has launched with US$1.5 million in seed funding to build tools that detect and explain the harmful effects of drugs at much earlier stages.
The startup’s approach focuses on how individual cells respond to a drug. Instead of analysing cells in bulk, it studies them one by one. This helps capture differences in how cells react, which are often missed in traditional testing methods.
Drug toxicity remains one of the main reasons for failure in drug development. Methods such as animal testing and bulk cell analysis do not always reflect how human cells behave. This gap has pushed the industry to look for more reliable and human-relevant ways to test drug safety.
DeepCyte combines cell-level data with artificial intelligence. Its platform, MetaCore, studies what is happening inside individual cells by capturing detailed molecular information. This data is used to build large datasets that can train AI models.
Additionally, the company has developed an AI system called DeeImmuno. It is designed to predict whether a drug could be toxic and identify the biological reasons behind it. In internal testing on 100 drugs, the system identified different types of toxicity and their underlying mechanisms with a reported accuracy of 94 percent.
The focus on explaining why a drug is toxic, not just whether it is, reflects a broader shift in the industry. Regulators such as the U.S. Food and Drug Administration and the European Medicines Agency have been encouraging methods that rely more on human cell data and clearer biological evidence. The seed funding will be used to develop and scale these tools. The company aims to help drug developers make earlier decisions, which could reduce costly failures in later stages. Whether tools like this become widely used will depend on how they perform in real-world settings. For now, DeepCyte’s approach highlights a growing effort to make drug testing more precise by focusing on how drugs affect cells at the most detailed level.
A planned city explores how real-time data and automation can shape everyday urban systems
A newly built district in northern China is being used to test how cities function when infrastructure, data and automation are integrated from the ground up. In Xiong'an New Area, traffic systems, public monitoring and urban services are designed to respond in real time rather than operate on fixed rules.
At the centre of this is a traffic management system powered by more than 20,000 roadside sensors. These track traffic flow, vehicle types and congestion levels, feeding data into an AI system that adjusts signals in milliseconds. Official figures show this has reduced the average number of stops per vehicle by half. The system also detects equipment faults, sends alerts and generates maintenance requests without manual input.
Automation extends beyond roads. Drones are deployed across the city for routine monitoring. In the Rongdong district, roadside units release drones that follow fixed patrol routes of around 1.27 kilometres, completing each run in about five minutes. They are used to monitor traffic, detect illegal parking and inspect public spaces. Similar systems operate in parks to track water levels and issue flood alerts, while in some work zones, drones transport packages of up to five kilograms between buildings.
These applications reflect a broader approach: integrating multiple systems into a single, connected urban framework. Unlike older cities where infrastructure evolves in layers, Xiong’an has been built with coordinated digital systems from the outset. This allows transport, maintenance and public services to operate through shared data systems rather than in isolation.
Alongside this, the area is being developed as a technology and innovation hub. Since its establishment in 2017, it has attracted more than 400 branches of state-owned enterprises and over 200 companies working in sectors such as artificial intelligence, aerospace information and digital technology.
This ecosystem supports projects like the “Xiong’an-1” satellite, which completed research, design, production and testing within eight months of regulatory approval in 2025. The satellite is currently undergoing testing, with a planned launch expected in the second quarter of 2026. It forms part of a broader push to build an aerospace information industry in the region.
The area is also structured to bring companies, research and production closer together. At the Zhongguancun Science Park in Xiong’an, which spans 207,000 square metres, 269 technology companies operate across sectors including AI, robotics and biotechnology. The park hosts more than 2,700 researchers and industry professionals, with companies organised into sector-specific clusters.
Policy support continues to shape this development. In early 2026, the State Council approved the upgrade of Xiong’an’s high-tech industrial development zone to national level status, with a focus on attracting high-end research and strengthening links between scientific development and industrial output.
Xiong’an is positioned as a testing ground for how smart city systems can be deployed at scale. The model depends on coordinated planning, integrated infrastructure and sustained policy support. Whether these systems can be adapted to existing cities, where infrastructure and governance are more fragmented, remains an open question.
Backed by Menlo Ventures, BrainGrid tackles planning gaps as AI makes software building accessible to more founders.
As artificial intelligence makes it easier to write code, a different problem is starting to surface. Building software is no longer limited by technical skill alone. Increasingly, the challenge lies in deciding what to build, how to structure it, and how to turn an idea into something that actually works.
That shift sits at the centre of BrainGrid, a startup that has raised $1 million in pre-seed funding led by Menlo Ventures, with participation from Next Tier Ventures and Brainstorm Ventures. The company is building what it describes as an AI-powered planning layer for people who want to create software but may not have a technical background.
The timing reflects a broader change in how products are being built. Tools like Claude Code and Cursor have made it possible to generate working code through simple prompts. For many first-time founders, this has lowered the barrier to entry. But writing code is only one part of the process. Turning that code into a reliable product requires structure, sequencing and clarity—areas where many projects begin to fall apart.
In traditional teams, this responsibility sits with product managers who define what needs to be built and in what order. Without that layer, even well-written code can lead to products that feel disjointed or incomplete. Features may not work together, integrations can break and the final product often does not match the original idea.
BrainGrid is designed to address that gap. Instead of focusing on generating code, it helps users map out the structure of a product before development begins. The aim is to give builders a clearer starting point so that the tools they use—whether human or AI—can produce more consistent results.
The company says more than 500 builders have already used it to create software products across areas like fitness, healthcare and productivity. These range from first-time founders experimenting with new ideas to experienced developers working independently. In many cases, the products are already live and generating revenue, suggesting that the demand is not just for experimentation but for building something that can scale.
For investors, the appeal lies in the evolving role of software development. As AI takes on more of the technical work, the value shifts toward defining the problem and structuring the solution. In that sense, planning becomes less of a background task and more of a core capability.
The US$1 million raise is relatively modest, but it points to a larger trend. As more people gain access to AI tools, the number of potential builders expands. What remains limited is the ability to organise ideas into products that work in the real world. If that shift continues, the next wave of software may not be defined by who can code, but by who can plan.
HSUHK’s award-winning system shows how AI, drones and AR can cut training time, reduce errors and reshape warehouse operations
As global tech ecosystems become more interconnected, the ability to move innovation across borders is becoming just as important as building it. A new partnership between MTR Lab, the investment arm of MTR Corporation and ZGC Science City Ltd, a government-backed technology ecosystem based in Beijing’s Haidian district, reflects this shift.
At its core, the collaboration is designed to connect high-potential Chinese startups with global capital, real-world deployment opportunities and international markets. It focuses on sectors like AI, robotics, smart mobility and sustainable urban development—areas where China already has strong technical depth but where scaling beyond domestic markets can be more complex.
This is where the partnership begins to matter. ZGC Science City sits at the center of one of China’s most concentrated innovation clusters, with thousands of AI companies and a growing base of specialised and high-growth firms. MTR Lab, on the other hand, brings access to international markets, industry networks and practical deployment environments tied to infrastructure, transport and urban systems. Together, they are attempting to bridge a familiar gap: turning local innovation into globally relevant products.
In practice, the model is straightforward. ZGC Science City will introduce MTR Lab to startups working in priority sectors, creating a pipeline for potential investment and collaboration. From there, MTR Lab can support these companies through funding, pilot projects and access to overseas markets. The idea is not just to invest, but to help startups test and apply their technologies in real-world settings, particularly in complex urban environments.
The timing is notable. China’s AI and deep tech ecosystem has expanded rapidly, with thousands of companies contributing to advancements in automation, smart infrastructure and sustainability. At the same time, global demand for these technologies is rising, especially as cities look for more efficient and scalable solutions. Yet, moving from innovation to adoption often requires cross-border coordination—something individual startups may struggle to navigate alone.
This partnership also builds on a broader pattern. Corporate venture arms like MTR Lab are increasingly positioning themselves not just as investors, but as connectors between markets. By combining capital with access to infrastructure and deployment scenarios, they offer startups a way to move faster from development to real-world use. For ZGC Science City, the collaboration adds an international layer to its ecosystem, helping local companies extend beyond domestic growth.
What emerges is a model that goes beyond a typical investment announcement. It reflects a growing recognition that innovation today is rarely confined to one geography. Technologies may be developed in one ecosystem, refined in another and scaled globally through partnerships like this.
As cross-border collaboration becomes more central to how startups grow, partnerships like the one between MTR Lab and ZGC Science City point to a more connected innovation landscape—one where access, not just invention, defines success.
AI actor Tilly Norwood releases a musical video arguing that artificial intelligence can expand creativity in film
As Hollywood prepares for this weekend’s Oscars, a different kind of performer is stepping into the spotlight — one that doesn’t physically exist.
Tilly Norwood, described as the world’s first AI actor, has released her debut musical comedy video, Take the Lead. The project arrives at a moment when artificial intelligence has become one of the most contentious topics in the film industry.
The message of the song is simple. AI should not be seen as a threat to actors. Instead, it can become another creative tool. The release also offers a first look at what Norwood’s creators call the “Tillyverse”. It is envisioned as a cloud-based entertainment world where AI characters can live, interact and perform.
Behind the character is actor and producer Eline van der Velden. She is the CEO of production company Particle6 and AI talent studio Xicoia. Van der Velden created Tilly as a way to experiment with how artificial intelligence could be used in storytelling.
The timing is not accidental. The entertainment industry has spent the past few years debating the role AI should play in filmmaking and acting. Questions about digital replicas, automated performances and creative ownership continue to divide artists and studios.
Norwood’s musical video enters that debate with a different tone. Instead of warning about AI replacing actors, the project suggests that the technology could expand what performers are able to do.
The video itself also serves as a technical experiment. The song Take the Lead was generated using the AI music platform Suno. The video was then produced using a combination of widely available AI tools and Particle6’s own creative process.
One of the newer techniques used in the project is performance capture. Van der Velden physically acted out Tilly’s movements and expressions so the digital character could mirror a human performance. But the production was far from automated. According to Particle6, a team of 18 people worked on the video. The group included a director, editor, production designer, costume designer, comedy writer and creative technologist. In other words, the project still relied heavily on human creativity.
“Tilly has always been a vehicle to test the creative capabilities and boundaries of AI,” van der Velden said. “It’s not about taking anyone’s job”. She added that even with powerful tools, good AI content still takes time, taste and creative direction.
The project also reflects how quickly production technology is evolving. Tools that once required large studios are now accessible to smaller creative teams experimenting with AI-driven storytelling.
For Particle6, the character of Tilly Norwood acts as a testing ground. Each project explores how AI performers might be developed, directed and integrated into entertainment. Whether audiences embrace digital actors remains an open question. Many in the industry are still wary of how AI could reshape creative work.
But projects like Take the Lead show another possibility. Instead of replacing performers, artificial intelligence could become part of the creative process itself. In that sense, Tilly Norwood may represent something more than a virtual performer. She is also an experiment in how humans and machines might collaborate in the future of entertainment.
A wearable ring, conversational AI and US$23M in funding. Sandbar wants to rethink how we interact with technology
Sandbar, a New York–based interface startup, has raised US$23 million in Series A funding to develop a wearable device that lets people interact with artificial intelligence via voice rather than screens.
Adjacent and Kindred Ventures led the round; both venture firms focused on early-stage technology startups. The investment brings Sandbar’s total funding to us$36 million. Earlier backing included a US$10 million seed round led by True Ventures, a venture capital firm, as well as a US$3 million pre-seed round supported by Upfront Ventures, a venture firm and Betaworks, a startup studio and investment firm.
Sandbar was founded by Mina Fahmi and Kirak Hong, who previously worked together at CTRL-labs, a neural interface startup acquired by Meta in 2019. Their earlier work explored how computers could respond more directly to human intent — an idea that continues to shape Sandbar’s approach to AI interfaces.
The new funding will help the company expand its team across machine learning, interaction design and software engineering as it prepares to launch its first product. That product, called Stream, combines a wearable ring with a conversational AI interface. The system allows users to speak to an AI assistant without unlocking a phone or opening an app.
The concept is simple. Instead of typing into a screen, users press a button on the ring and talk. The system can capture notes, organize ideas, retrieve information from the web or trigger actions through connected applications.
The ring includes a microphone, a touchpad and subtle haptic feedback. These elements allow the device to respond through gentle vibrations rather than visual alerts. According to the company, the ring only listens when the user presses the button — a design meant to address common concerns around always-on microphones.
That design reflects a larger shift Sandbar believes is underway. As AI assistants become more capable, many startups are experimenting with new ways to interact with them. The focus is moving away from screens and keyboards toward interfaces that feel more natural and immediate.
Stream uses multiple AI models working together to process requests, search the web and structure information in real time. The company says users remain in control of their data and can choose whether to share information with other apps.
Sandbar is also developing a feature called Inner Voice, which responds using a voice customized to the user. The feature will debut during a closed beta planned for this spring, giving the company time to refine how the software behaves in everyday use.
The startup currently employs a team of 15 people. Many have worked on well-known consumer devices including the iPhone, Fitbit, Kindle and Vision Pro. Recent hires include Sam Bowen, formerly of Amazon and Fitbit, who joined as vice president of hardware and Brooke Travis, previously at Equinox, Dior and Gap, who now leads marketing.
Sandbar plans to begin shipping Stream in summer 2026 after completing early testing. As artificial intelligence tools become more integrated into daily life, the company is betting that the next shift in computing will not come from another app — but from new ways for people to interact with AI itself.
Structured AI interviews and human judgment combine to address the global talent shortage
As hiring pressures mount across global markets, ManpowerGroup is turning to technology to strengthen how it connects people to work. The workforce solutions major has announced a global partnership with Hubert, a startup focused on AI-driven structured interviews. The aim is simple: make hiring faster and fairer, without removing the human touch.
ManpowerGroup has spent decades operating at the center of the global labor market. The company works with employers across industries to fill roles, manage workforce planning and build talent pipelines. With millions of placements each year, it has a clear view of how strained hiring has become. A large share of employers today report difficulty finding skilled talent. At the same time, candidates expect more transparency, quicker feedback and flexibility in how they engage with employers.
Hubert enters this picture as a specialist in structured digital interviewing. The startup has built tools that allow candidates to complete interviews online, at any time, while being assessed against consistent criteria. Instead of relying on informal screening calls or resume filters, its system focuses on standardized questions tied directly to job requirements. The idea is to bring more consistency to early-stage hiring.
The partnership brings these capabilities into ManpowerGroup’s global operations. AI-powered interviews will now support the first stage of screening, helping recruiters identify qualified candidates earlier in the process. This does not replace recruiters. Final decisions and contextual judgment remain with experienced hiring professionals. What changes is the speed and structure of the initial assessment.
For employers, this could mean earlier visibility into job-ready talent and less time spent on manual screening. For candidates, it offers more flexibility. A significant portion of interviews on Hubert’s platform are completed outside regular office hours, allowing applicants to engage when it suits them. That flexibility can make a difference in competitive labor markets where timing matters.
The collaboration is also positioned as a step toward reducing bias. By evaluating each candidate against the same transparent standards, the process becomes more consistent. While no system can remove bias entirely, structured assessments can reduce the variability that often comes with unstructured interviews.
At its core, the partnership addresses a gap many large organizations are facing. They need scale and speed, but they cannot afford to lose the human judgment that good hiring depends on. Manual processes are too slow. Fully automated systems can feel impersonal and risky. ManpowerGroup’s approach suggests a middle path, where technology handles repetition and structure and recruiters focus on potential and fit.
The move also reflects a broader shift in the workforce industry. AI is no longer being tested on the sidelines. It is being built into the foundation of hiring operations. For established players like ManpowerGroup, the challenge is not whether to adopt AI, but how to do so responsibly and at scale.
By working with Hubert, the company is signaling that the future of recruitment will likely blend structured digital tools with human expertise. In a market defined by talent shortages and rising expectations, that balance may prove critical.
AI meets AR: How Rokid Glasses bring multilingual, real-time intelligence to smart eyewear globally
Rokid, a Chinese company specializing in AI-powered smart eyewear and human–computer interaction, has rolled out a major software update for the international version of its Rokid Glasses. This update makes it the first smart glasses manufacturer to natively support Google’s Gemini, alongside three other leading large language models: OpenAI’s ChatGPT, Alibaba’s Qwen and DeepSeek.
The integration is powered by Rokid’s device-to-cloud architecture, which enables users to switch between AI models on the fly. In practice, this means a traveler can receive a real-time translation in Japanese using one AI model, then quickly switch to ChatGPT to answer a technical query—without noticeable delay. The system also supports multi-modal inputs like voice and gestures, making interactions more intuitive for everyday use.
This is more than a routine software update. By combining AI models from both U.S. and Chinese developers, Rokid is making its smart glasses relevant to global users, with features that adapt to local languages and preferences while maintaining high performance.
These technological advancements have directly fueled Rokid’s international growth. Between November 2024 and October 2025, Shangpu Group data shows Rokid Glasses ranked No.1 in global sales for AI glasses with display functionality. Crowdfunding milestones further reflect this momentum: the product became the fastest smart glasses to raise over 100 million Japanese Yen on Japan’s MAKUAKE platform and broke Kickstarter records for smart eyewear.
Taken together, Rokid’s update highlights a shift in the smart glasses space: success increasingly comes from openness, flexibility and localized AI experiences rather than closed, single-platform ecosystems. By giving users choice, integrating global AI capabilities and bridging cultural and linguistic gaps, Rokid is positioning itself as a serious contender in the international AR and AI wearable market.
The focus is no longer just AI-generated worlds, but how those worlds become structured digital products
As AI tools improve, creating 3D content is becoming faster and easier. However, building that content into interactive experiences still requires time, structure and technical work. That difference between generation and execution is where HTC VIVERSE and World Labs are focusing their new collaboration.
HTC VIVERSE is a 3D content platform developed by HTC. It provides creators with tools to build, refine and publish interactive virtual environments. Meanwhile, World Labs is an AI startup founded by researcher Fei-Fei Li and a team of machine learning specialists. The company recently introduced Marble, a tool that generates full 3D environments from simple text, image or video prompts.
While Marble can quickly create a digital world, that world on its own is not yet a finished experience. It still needs structure, navigation and interaction. This is where VIVERSE fits in. By combining Marble’s world generation with VIVERSE’s building tools, creators can move from an AI-generated scene to a usable, interactive product.
In practice, the workflow works in two steps. First, Marble produces the base 3D environment. Then, creators bring that environment into VIVERSE, where they add game mechanics, scenes and interactive elements. In this model, AI handles the early visual creation, while the human creator defines how users explore and interact with the world.
To demonstrate this process, the companies developed three example projects. Whiskerhill turns a Marble-generated world into a simple quest-based experience. Whiskerport connects multiple AI-generated scenes into a multi-level environment that users navigate through portals. Clockwork Conspiracy, built by VIVERSE, uses Marble’s generation system to create a more structured, multi-scene game. These projects are not just demos. They serve as proof that AI-generated worlds can evolve beyond static visuals and become interactive environments.
This matters because generative AI is often judged by how quickly it produces content. However, speed alone does not create usable products. Digital experiences still require sequencing, design decisions and user interaction. As a result, the real challenge is not generation, but integration — connecting AI output to tools that make it functional.
Seen in this context, the collaboration is less about a single product and more about workflow. VIVERSE provides a system that allows AI-generated environments to be edited and structured. World Labs provides the engine that creates those environments in the first place. Together, they are testing whether AI can fit directly into a full production pipeline rather than remain a standalone tool.
Ultimately, the collaboration reflects a broader change in creative technology. AI is no longer only producing isolated assets. It is beginning to plug into the larger process of building complete experiences. The key question is no longer how quickly a world can be generated, but how easily that world can be turned into something people can actually use and explore.