Publications

Showing 14 publications


Orchid: A Creative Approach for Authoring LLM-Driven Interactive Narratives
Orchid: A Creative Approach for Authoring LLM-Driven Interactive Narratives
Zhen Wu, Serkan Kumyol, Shing Yin Wong, Xiaozhu Hu, Xin Tong, Tristan Braud
In To be presented at Creativity and Cognition (C&C '25), June 23--25, 2025, Virtual, United Kingdom. 2025
@article{wu2025orchid, author = {Zhen Wu and Serkan Kumyol and Shing Yin Wong and Xiaozhu Hu and Xin Tong and Tristan Braud}, booktitle = {To be presented at Creativity and Cognition (C&C '25), June 23--25, 2025, Virtual, United Kingdom}, month = {jun}, title = {Orchid: A Creative Approach for Authoring LLM-Driven Interactive Narratives}, year = {2025} }
Integrating Large Language Models (LLMs) into Interactive Digital Narratives (IDNs) enables dynamic storytelling where user interactions shape the narrative in real time, challenging traditional authoring methods. This paper presents the design study of Orchid, a creative approach for authoring LLM-driven IDNs. Orchid allows users to structure the hierarchy of narrative stages and define the rules governing LLM narrative generation and transitions between stages. The development of Orchid consists of three phases. 1) Formulating Orchid through desk research on existing IDN practices. 2) Implementation of a technology probe based on Orchid. 3) Evaluating how IDN authors use Orchid to design IDNs, verify Orchid's hypotheses, and explore user needs for future authoring tools. This study confirms that authors are open to LLM-driven IDNs but desire strong authorial agency in narrative structures, highlighted in accuracy in branching transitions and story details. Future design implications for Orchid include introducing deterministic variable handling, support for trans-media applications, and narrative consistency across branches.
AIR-HLoc: Adaptive Image Retrieval for Efficient Visual Localisation
AIR-HLoc: Adaptive Image Retrieval for Efficient Visual Localisation
Changkun Liu, Jianhao Jiao, Huajian Huang, Zhengyang Ma, Dimitrios Kanoulas, Tristan Braud
In Proceedings of 2025 IEEE International Conference on Robotics and Automation (ICRA). 2025
@inproceedings{liu2024air, author = {Changkun Liu and Jianhao Jiao and Huajian Huang and Zhengyang Ma and Dimitrios Kanoulas and Tristan Braud}, booktitle = {2025 IEEE International Conference on Robotics and Automation (ICRA)}, month = {may}, title = {AIR-HLoc: Adaptive Image Retrieval for Efficient Visual Localisation}, year = {2025} }
State-of-the-art hierarchical localisation pipelines (HLoc) employ image retrieval (IR) to establish 2D-3D correspondences by selecting the top-k most similar images from a reference database. While increasing k improves localisation robustness, it also linearly increases computational cost and runtime, creating a significant bottleneck. This paper investigates the relationship between global and local descriptors, showing that greater similarity between the global descriptors of query and database images increases the proportion of feature matches. Low similarity queries significantly benefit from increasing k, while high similarity queries rapidly experience diminishing returns. Building on these observations, we propose an adaptive strategy that adjusts k based on the similarity between the query's global descriptor and those in the database, effectively mitigating the feature-matching bottleneck. Our approach optimizes processing time without sacrificing accuracy. Experiments on three indoor and outdoor datasets show that AIR-HLoc reduces feature matching time by up to 30\%, while preserving state-of-the-art accuracy. The results demonstrate that AIR-HLoc facilitates a latency-sensitive localisation system.
LiteVLoc: Map-Lite Visual Localization for Image Goal Navigation
LiteVLoc: Map-Lite Visual Localization for Image Goal Navigation
Jianhao Jiao, Jinhao He, Changkun Liu, Sebastian Aegidius, Xiangcheng Hu, Tristan Braud, Dimitrios Kanoulas
In Proceedings of 2025 IEEE International Conference on Robotics and Automation (ICRA). 2025
@inproceedings{jiao2024litevloc, author = {Jianhao Jiao and Jinhao He and Changkun Liu and Sebastian Aegidius and Xiangcheng Hu and Tristan Braud and Dimitrios Kanoulas}, booktitle = {2025 IEEE International Conference on Robotics and Automation (ICRA)}, month = {may}, title = {LiteVLoc: Map-Lite Visual Localization for Image Goal Navigation}, year = {2025} }
This paper presents LiteVLoc, a hierarchical visual localization framework that uses a lightweight topo-metric map to represent the environment. The method consists of three sequential modules that estimate camera poses in a coarse-to-fine manner. Unlike mainstream approaches relying on detailed 3D representations, LiteVLoc reduces storage overhead by leveraging learning-based feature matching and geometric solvers for metric pose estimation. A novel dataset for the map-free relocalization task is also introduced. Extensive experiments including localization and navigation in both simulated and real-world scenarios have validate the system's performance and demonstrated its precision and efficiency for large-scale deployment. Code and data will be made publicly available.
GS-CPR: Efficient Camera Pose Refinement via 3D Gaussian Splatting
GS-CPR: Efficient Camera Pose Refinement via 3D Gaussian Splatting
Changkun Liu, Shuai Chen, Yash Sanjay Bhalgat, Siyan HU, Ming Cheng, Zirui Wang, Victor Adrian Prisacariu, Tristan Braud
In Proceedings of The Thirteenth International Conference on Learning Representations (ICLR). 2025
@inproceedings{liu2025gscpr, author = {Changkun Liu and Shuai Chen and Yash Sanjay Bhalgat and Siyan HU and Ming Cheng and Zirui Wang and Victor Adrian Prisacariu and Tristan Braud}, booktitle = {The Thirteenth International Conference on Learning Representations (ICLR)}, month = {apr}, title = {{GS}-{CPR}: Efficient Camera Pose Refinement via 3D Gaussian Splatting}, url = {https://openreview.net/forum?id=mP7uV59iJM}, year = {2025} }
We leverage 3D Gaussian Splatting (3DGS) as a scene representation and propose a novel test-time camera pose refinement (CPR) framework, GS-CPR. This framework enhances the localization accuracy of state-of-the-art absolute pose regression and scene coordinate regression methods. The 3DGS model renders high-quality synthetic images and depth maps to facilitate the establishment of 2D-3D correspondences. GS-CPR obviates the need for training feature extractors or descriptors by operating directly on RGB images, utilizing the 3D foundation model, MASt3R, for precise 2D matching. To improve the robustness of our model in challenging outdoor environments, we incorporate an exposure-adaptive module within the 3DGS framework. Consequently, GS-CPR enables efficient one-shot pose refinement given a single RGB query and a coarse initial pose estimation. Our proposed approach surpasses leading NeRF-based optimization methods in both accuracy and runtime across indoor and outdoor visual localization benchmarks, achieving new state-of-the-art accuracy on two indoor datasets.
SC-OmniGS: Self-Calibrating Omnidirectional Gaussian Splatting
SC-OmniGS: Self-Calibrating Omnidirectional Gaussian Splatting
Huajian Huang, Yingshu Chen, Longwei Li, Hui Cheng, Tristan Braud, Yajie Zhao, Sai-Kit Yeung
In Proceedings of The Thirteenth International Conference on Learning Representations (ICLR). 2025
@inproceedings{huang2025scomnigs, author = {Huajian Huang and Yingshu Chen and Longwei Li and Hui Cheng and Tristan Braud and Yajie Zhao and Sai-Kit Yeung}, booktitle = {The Thirteenth International Conference on Learning Representations (ICLR)}, month = {apr}, title = {{SC}-Omni{GS}: Self-Calibrating Omnidirectional Gaussian Splatting}, url = {https://openreview.net/forum?id=7idCpuEAiR}, year = {2025} }
360-degree cameras streamline data collection for radiance field 3D reconstruction by capturing comprehensive scene data. However, traditional radiance field methods do not address the specific challenges inherent to 360-degree images. We present SC-OmniGS, a novel self-calibrating omnidirectional Gaussian splatting system for fast and accurate omnidirectional radiance field reconstruction using 360-degree images. Rather than converting 360-degree images to cube maps and performing perspective image calibration, we treat 360-degree images as a whole sphere and derive a mathematical framework that enables direct omnidirectional camera pose calibration accompanied by 3D Gaussians optimization. Furthermore, we introduce a differentiable omnidirectional camera model in order to rectify the distortion of real-world data for performance enhancement. Overall, the omnidirectional camera intrinsic model, extrinsic poses, and 3D Gaussians are jointly optimized by minimizing weighted spherical photometric loss. Extensive experiments have demonstrated that our proposed SC-OmniGS is able to recover a high-quality radiance field from noisy camera poses or even no pose prior in challenging scenarios characterized by wide baselines and non-object-centric configurations. The noticeable performance gain in the real-world dataset captured by consumer-grade omnidirectional cameras verifies the effectiveness of our general omnidirectional camera model in reducing the distortion of 360-degree images.
I Light U Up: Exploring a New Emergent Narrative Paradigm through Physical Data Participation in AI Generative Experiences
I Light U Up: Exploring a New Emergent Narrative Paradigm through Physical Data Participation in AI Generative Experiences
Zhen Wu, Ruoyu Wen, Marcel Zaes Sagesser, Sum Yi Ma, Jiashuo Xian, Wenda Wu, Tristan Braud
In SIGGRAPH Asia 2024 Art Papers. 2024
@incollection{wu2024light, author = {Zhen Wu and Ruoyu Wen and Marcel Zaes Sagesser and Sum Yi Ma and Jiashuo Xian and Wenda Wu and Tristan Braud}, booktitle = {SIGGRAPH Asia 2024 Art Papers}, month = {dec}, pages = {1--9}, title = {I Light U Up: Exploring a New Emergent Narrative Paradigm through Physical Data Participation in AI Generative Experiences}, year = {2024} }
This paper introduces "I Light U Up", an interactive installation that explores the use of light as input for emergent narrative powered by LLM. Through an installation with light sensors, the audience uses light to communicate with the parallel digital beings, generating a narrative experience that emerges from the interaction. This work serves as a proof-of-concept, supporting a conceptual framework that aims to open new perspectives on the participation of physical data in emergent narrative experiences. It seeks to develop a new dimension in contrast to the text-centric one-way relationship between reader and story by including agency beyond human interactivity.
Nature: Metaphysics+ Metaphor (N: M+ M): Exploring Persistence, Feedback, and Visualisation in Mixed Reality Performance Arts.
Nature: Metaphysics+ Metaphor (N: M+ M): Exploring Persistence, Feedback, and Visualisation in Mixed Reality Performance Arts.
Tristan Braud, Brian Lau, Dominie Hoi Lam Chan, Chun Ming Wu, Zhen Wu, Vi Jun Sean Yong, Kirill Shatilov
In SIGGRAPH Asia 2024 Art Papers. 2024
@incollection{braud2024nature, author = {Tristan Braud and Brian Lau and Dominie Hoi Lam Chan and Chun Ming Wu and Zhen Wu and Vi Jun Sean Yong and Kirill Shatilov}, booktitle = {SIGGRAPH Asia 2024 Art Papers}, month = {dec}, pages = {1--7}, title = {Nature: Metaphysics+ Metaphor (N: M+ M): Exploring Persistence, Feedback, and Visualisation in Mixed Reality Performance Arts.}, year = {2024} }
Nature: Metaphysics + Metaphor (N:M+M) is an art experience and exhibition that explores the intersection of Nature, Art, Man, and Technology. This immersive experience aims to bridge physical and digital, tangible and intangible, and ephemeral and persistent in mixed reality arts through new interaction and visualization methods. The performance was presented in front of an audience of 50 people, during which the artists constructed a physical-digital artwork that became part of a subsequent exhibition. This paper describes the rationale, creative approach, and technical contribution behind this work while reflecting on audience feedback and future directions.
SoundMorphTPU: Exploring Gesture Mapping in Deformable Interfaces for Music Interaction
SoundMorphTPU: Exploring Gesture Mapping in Deformable Interfaces for Music Interaction
Zhen Wu, Ze Gao, Hua Xu, Xingxing Yang, Tristan Braud
In Proceedings of Proceedings of the International Conference on New Interfaces for Musical Expression. 2024
@inproceedings{wu2024soundmorphtpu, author = {Zhen Wu and Ze Gao and Hua Xu and Xingxing Yang and Tristan Braud}, booktitle = {Proceedings of the International Conference on New Interfaces for Musical Expression}, month = {sep}, pages = {395--406}, title = {SoundMorphTPU: Exploring Gesture Mapping in Deformable Interfaces for Music Interaction}, year = {2024} }
Deformable interface is an emerging field with significant potential for use in computing applications, particularly in the design of Digital Music Instruments (DMIs). While prior works have investigated the design of gestural input for deformable interfaces and developed novel musical interactions, there remains a gap in understanding the tangible gestures as input and their corresponding output from the user's perspectives. This study explores the relationship between gestural input and the output of a deformable interface for multi-gestural music interaction. Following a pilot study to explore materials and their corresponding intuitive gestures with participants, we develop a TPU fabric interface as a probe to investigate this question in the context of musical interaction. Through user engagement with the probe as a sound control, we discovered that the input-output relationship between gestures and the sound can have meaningful implications for users' embodied interaction with the system. Our research deepens the understanding of designing deformable interfaces and their capacity to enhance embodied experiences in music interaction scenarios.
360Loc: A Dataset and Benchmark for Omnidirectional Visual Localization with Cross-device Queries
360Loc: A Dataset and Benchmark for Omnidirectional Visual Localization with Cross-device Queries
Huajian Huang, Changkun Liu, Yipeng Zhu, Hui Cheng, Tristan Braud, Sai-Kit Yeung
In Proceedings of Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2024
@inproceedings{huang2024360loc, author = {Huajian Huang and Changkun Liu and Yipeng Zhu and Hui Cheng and Tristan Braud and Sai-Kit Yeung}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)}, month = {jun}, pages = {22314--22324}, title = {360Loc: A Dataset and Benchmark for Omnidirectional Visual Localization with Cross-device Queries}, year = {2024} }
Portable 360∘ cameras are becoming a cheap and efficient tool to establish large visual databases. By capturing omnidirectional views of a scene, these cameras could expedite building environment models that are essential for visual localization. However, such an advantage is often overlooked due to the lack of valuable datasets. This paper introduces a new benchmark dataset, 360Loc, composed of 360∘ images with ground truth poses for visual localization. We present a practical implementation of 360∘ mapping combining 360∘ images with lidar data to generate the ground truth 6DoF poses. 360Loc is the first dataset and benchmark that explores the challenge of cross-device visual positioning, involving 360∘ reference frames, and query frames from pinhole, ultra-wide FoV fisheye, and 360∘ cameras. We propose a virtual camera approach to generate lower-FoV query frames from 360∘ images, which ensures a fair comparison of performance among different query types in visual localization tasks. We also extend this virtual camera approach to feature matching-based and pose regression-based methods to alleviate the performance loss caused by the cross-device domain gap, and evaluate its effectiveness against state-of-the-art baselines. We demonstrate that omnidirectional visual localization is more robust in challenging large-scale scenes with symmetries and repetitive structures. These results provide new insights into 360-camera mapping and omnidirectional visual localization with cross-device queries.
Hr-apr: Apr-agnostic framework with uncertainty estimation and hierarchical refinement for camera relocalisation
Hr-apr: Apr-agnostic framework with uncertainty estimation and hierarchical refinement for camera relocalisation
Changkun Liu, Shuai Chen, Yukun Zhao, Huajian Huang, Victor Prisacariu, Tristan Braud
In Proceedings of 2024 IEEE International Conference on Robotics and Automation (ICRA). 2024
@inproceedings{liu2024hr, author = {Changkun Liu and Shuai Chen and Yukun Zhao and Huajian Huang and Victor Prisacariu and Tristan Braud}, booktitle = {2024 IEEE International Conference on Robotics and Automation (ICRA)}, month = {may}, organization = {IEEE}, pages = {8544--8550}, title = {Hr-apr: Apr-agnostic framework with uncertainty estimation and hierarchical refinement for camera relocalisation}, year = {2024} }
Absolute Pose Regressors (APRs) directly estimate camera poses from monocular images, but their accuracy is unstable for different queries. Uncertainty-aware APRs provide uncertainty information on the estimated pose, alleviating the impact of these unreliable predictions. However, existing uncertainty modelling techniques are often coupled with a specific APR architecture, resulting in suboptimal performance compared to state-of-the-art (SOTA) APR methods. This work introduces a novel APR-agnostic framework, HR-APR, that formulates uncertainty estimation as cosine similarity estimation between the query and database features. It does not rely on or affect APR network architecture, which is flexible and computationally efficient. In addition, we take advantage of the uncertainty for pose refinement to enhance the performance of APR. The extensive experiments demonstrate the effectiveness of our framework, reducing 27.4% and 15.2% of computational overhead on the 7Scenes and Cambridge Landmarks datasets while maintaining the SOTA accuracy in single-image APRs.
MARViN: Mobile AR Dataset with Visual-Inertial Data
MARViN: Mobile AR Dataset with Visual-Inertial Data
Changkun Liu, Yukun Zhao, Tristan Braud
In Proceedings of 2024 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW). 2024
@inproceedings{liu2024marvin, author = {Changkun Liu and Yukun Zhao and Tristan Braud}, booktitle = {2024 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)}, month = {mar}, organization = {IEEE}, pages = {532--538}, title = {MARViN: Mobile AR Dataset with Visual-Inertial Data}, year = {2024} }
Accurate camera relocalisation is a fundamental technology for extended reality (XR), facilitating the seamless integration and persistence of digital content within the real world. Benchmark datasets that measure camera pose accuracy have driven progress in visual re-localisation research. Despite notable progress in this field, there is a limited availability of datasets incorporating Visual Inertial Odometry (VIO) data from typical mobile AR frameworks such as ARKit or ARCore. This paper presents a new dataset, MARViN, comprising diverse indoor and outdoor scenes captured using het- erogeneous mobile consumer devices. The dataset includes camera images, ARCore or ARKit VIO data, and raw sensor data for several mobile devices, together with the corresponding ground-truth poses. MARViN allows us to demonstrate the capability of ARKit and ARCore to provide relative pose estimates that closely approximate ground truth within a short timeframe. Subsequently, we evaluate the performance of mobile VIO data in enhancing absolute pose estimations in both desktop simulation and user study.
Mobilearloc: On-device robust absolute localisation for pervasive markerless mobile ar
Mobilearloc: On-device robust absolute localisation for pervasive markerless mobile ar
Changkun Liu, Yukun Zhao, Tristan Braud
In Proceedings of 2024 IEEE International Conference on Pervasive Computing and Communications Workshops and other Affiliated Events (PerCom Workshops). 2024
@inproceedings{liu2024mobilearloc, author = {Changkun Liu and Yukun Zhao and Tristan Braud}, booktitle = {2024 IEEE International Conference on Pervasive Computing and Communications Workshops and other Affiliated Events (PerCom Workshops)}, month = {Mar}, organization = {IEEE}, pages = {544--549}, title = {Mobilearloc: On-device robust absolute localisation for pervasive markerless mobile ar}, year = {2024} }
Recent years have seen significant improvement in absolute camera pose estimation, paving the way for pervasive markerless Augmented Reality (AR). However, accurate absolute pose estimation techniques are computation- and storage-heavy, requiring computation offloading. As such, AR systems rely on visual-inertial odometry (VIO) to track the device’s relative pose between requests to the server. However, VIO suffers from drift, requiring frequent absolute repositioning. This paper introduces MobileARLoc, a new framework for on-device large-scale markerless mobile AR that combines an absolute pose regressor (APR) with a local VIO tracking system. Absolute pose regressors (APRs) provide fast on-device pose estimation at the cost of reduced accuracy. To address APR accuracy and reduce VIO drift, MobileARLoc creates a feedback loop where VIO pose estimations refine the APR predictions. The VIO system identifies reliable predictions of APR, which are then used to compensate for the VIO drift. We comprehensively evaluate MobileARLoc through dataset simulations. MobileARLoc halves the error compared to the underlying APR and achieves fast (80ms) on-device inference speed.
Vr prem+: An immersive pre-learning branching visualization system for museum tours
Vr prem+: An immersive pre-learning branching visualization system for museum tours
Ze Gao, Xiang Li, Changkun Liu, Xian Wang, Anqi Wang, Liang Yang, Yuyang Wang, Pan Hui, Tristan Braud
In Proceedings of Proceedings of the Eleventh International Symposium of Chinese CHI. 2023
@inproceedings{gao2023vr, author = {Ze Gao and Xiang Li and Changkun Liu and Xian Wang and Anqi Wang and Liang Yang and Yuyang Wang and Pan Hui and Tristan Braud}, booktitle = {Proceedings of the Eleventh International Symposium of Chinese CHI}, month = {nov}, pages = {374--385}, title = {Vr prem+: An immersive pre-learning branching visualization system for museum tours}, year = {2023} }
We present VR PreM+, an innovative VR system designed to enhance web exploration beyond traditional computer screens. Unlike static 2D displays, VR PreM+ leverages 3D environments to create an immersive pre-learning experience. Using keyword-based information retrieval allows users to manage and connect various content sources in a dynamic 3D space, improving communication and data comparison. We conducted preliminary and user studies that demonstrated efficient information retrieval, increased user engagement, and a greater sense of presence. These findings yielded three design guidelines for future VR information systems: display, interaction, and user-centric design. VR PreM+ bridges the gap between traditional web browsing and immersive VR, offering an interactive and comprehensive approach to information acquisition. It holds promise for research, education, and beyond.
Unlogical instrument: Material-driven gesture-controlled sound installation
Unlogical instrument: Material-driven gesture-controlled sound installation
Zhen Wu, Ze Gao, Tristan Braud
In Proceedings of Companion Publication of the 2023 ACM Designing Interactive Systems Conference (Art Gallery). 2023
@inproceedings{wu2023unlogical, author = {Zhen Wu and Ze Gao and Tristan Braud}, booktitle = {Companion Publication of the 2023 ACM Designing Interactive Systems Conference (Art Gallery)}, month = {jul}, pages = {86--89}, title = {Unlogical instrument: Material-driven gesture-controlled sound installation}, year = {2023} }
This proposal describes the design and demonstration of the Unlogical Instrument. This textile-based gesture-controlled sound interface invites the audience to generate sound by interacting with the textile using different gestures, such as poking, flicking, and patting. Unlogical is a statement about breaking the dogmatic approach of commonly used musical instruments. It brings resilience to the form of musical expression and interaction for the audience. It introduces an interactive mechanism mediated by intuitive human gestures and the curiosity to explore tangible material. Unlogical Instrument locates the focus on the textile. This artwork investigates the relationship between the gestures and the sound perception that originates from the textile through a material-driven approach. Inkjet printing is applied to transfer the original textile into a sensorial interface. The audience will listen to the sound while interacting with the textile surface, exploring and understanding this instrument through free-form playing.