Publications
Permanent URI for this collectionhttps://hdl.handle.net/10217/239510
Browse
Browsing Publications by Author "Batmaz, Anil Ufuk, author"
Now showing 1 - 2 of 2
- Results Per Page
- Sort Options
Item Open Access An artists' perspectives on natural interactions for virtual reality 3D sketching(Colorado State University. Libraries, 2024-05-11) Rodriguez, Richard, author; Sullivan, Brian T., author; Machuca, Mayra Donaji Barrera, author; Batmaz, Anil Ufuk, author; Tornatzky, Cyane, author; Ortega, Francisco R., author; ACM, publisherVirtual Reality (VR) applications like OpenBrush offer artists access to 3D sketching tools within the digital 3D virtual space. These 3D sketching tools allow users to "paint" using virtual digital strokes that emulate real-world mark-making. Yet, users paint these strokes through (unimodal) VR controllers. Given that sketching in VR is a relatively nascent field, this paper investigates ways to expand our understanding of sketching in virtual space, taking full advantage of what an immersive digital canvas offers. Through a study conducted with the participation of artists, we identify potential methods for natural multimodal and unimodal interaction techniques in 3D sketching. These methods demonstrate ways to incrementally improve existing interaction techniques and incorporate artistic feedback into the design.Item Unknown Lights, headset, tablet, action: exploring the use of hybrid user interfaces for immersive situated analytics(Colorado State University. Libraries, 2024-10-24) Zhou, Xiaoyan, author; Lee, Benjamin, author; Ortega, Francisco R., author; Batmaz, Anil Ufuk, author; Yang, Yalong, author; ACM, publisherWhile augmented reality (AR) headsets provide entirely new ways of seeing and interacting with data, traditional computing devices can play a symbiotic role when used in conjunction with AR as a hybrid user interface. A promising use case for this setup is situated analytics. AR can provide embedded views that are integrated with their physical referents, and a separate device such as a tablet can provide a familiar situated overview of the entire dataset being examined. While prior work has explored similar setups, we sought to understand how people perceive and make use of visualizations presented on both embedded visualizations (in AR) and situated visualizations (on a tablet) to achieve their own goals. To this end, we conducted an exploratory study using a scenario and task familiar to most: adjusting light levels in a smart home based on personal preference and energy usage. In a prototype that simulates AR in virtual reality, embedded visualizations are positioned next to lights distributed across an apartment, and situated visualizations are provided on a handheld tablet. We observed and interviewed 19 participants using the prototype. Participants were easily able to perform the task, though the extent the visualizations were used during the task varied, with some making decisions based on the data and others only on their own preferences. Our findings also suggest the two distinct roles that situated and embedded visualizations can have, and how this clear separation might improve user satisfaction and minimize attention-switching overheads in this hybrid user interface setup. We conclude by discussing the importance of considering the user's needs, goals, and the physical environment for designing and evaluating effective situated analytics applications.