MobileHCI 2017: Workshop on Object Recognition for Input and Mobile Interaction
This workshop is on Sept 4th in Vienna, Austria in conjunction with MobileHCI 2017.
Submission deadline: May 26th (extended)
Acceptance notification: June 9th
Camera Ready papers due: July 21st
Papers available online: August 4th
Workshop date: September 4th
The full proceedings of the workshop will be available and linked from here on August 4th, 2017.
Today, we are seeing an emergence of devices that incorporate sensing capabilities that go beyond the traditional suite of hardware (e.g., touch sensing or proximity). These devices offer more fine-grained level of contextual information, such as object recognition, and they often vary in their size, portability, embeddability, and form factor. Despite this diversity, the manifestations of these new-generation sensing approaches will inevitably unlock many of the ubiquitous, tangible, mobile, and wearable computing ecosystems that promise to improve people’s lives.
These system are brought together by a variety of technologies, including computer vision, radar (e.g., Google ATAP’s Project Soli), acoustic sensing, fiducial tagging, and in general, IoT devices embedded with computational capabilities. Such systems open up a wide-range of applications spaces and novel forms of interaction. For instance, object-based interactions offer rich, contextual information that can power a wide range of user-centric applications (e.g., factory line optimization and safety, automatic grocery checkout, new forms of tangible interactions). Where, and how these interactions are applied also adds a new dimension to these applications (e.g., if a mobile device can detect which part of your body it is tapped into, it can launch the food app when tapped to your stomach).
Although the last few years have seen an increasing amount of research in this area, knowledge about this subject remains under explored, fragmented, and cuts across a set of related but heterogeneous issues. This workshop brings together researchers and practitioners interested in the challenges posed by “Object Recognition for Input and Mobile Interaction”.
This workshop aims to bring together active and interested researchers in sensing techniques, particularly those that advance input and interaction using novel capabilities (e.g., object recognition, radar sensing) across a range of modalities (e.g., mobile devices, wearables). This workshop will foster a scholarly environment for sharing approaches and experiences, identifying research and deployment challenges, and envision the next generation of applications that rely on widely deployed sensor systems, paying close attention to not just one, but an ecosystem of touchless, interaction devices.
There are many challenges with building the underlying system infrastructures for object recognition for mobile interaction. For example, what are the standards and operating system requirements? How can vision, radar or acoustic sensing or tagging systems which exist in multiple mobile and wearable elements act in a coordinated manner to reliably determine object interaction? What types of software libraries, simulators and IDEs are required for development? How can we ensure inter-device interoperability in the face of heterogeneous device configurations and varying approaches to object recognition?
In addition, there are significant challenges in the design and deployment of these kinds of object based interfaces. For example, what are the implications of interfaces that rely on sensing the movement and type of passive, physical objects? How might multiple points of sensing on, and around the body open up new design considerations? Which elements of the interface are best distributed in outputs? What new interaction techniques are necessary in these environments? What are the performance, comfort, and preference consequences of relying of object-interaction as a mobile interface?
We invite short (upto 2 pages) and longer (upto 6 pages) contributions from researchers and practitioners working in the area of object recognition and interaction. Some examples of topics of interest can be found below. Our goal is to have a group who can share approaches and experiences, identify research and deployment challenges, and envision the next generation of applications that rely on object recognition and interaction. Such systems can be embedded in personal, wearable and mobile devices i.e. physically decoupled in different ways yet are virtually coupled due to the interactions they support. We expect researchers who attend the workshop to be working on different views of the problem (e.g., at the interaction technique, application, middleware or hardware level), with a wide range of sensing systems and interaction technologies (e.g., Project Soli, computer vision, alternative hardware, …), and in a wide range of object interaction applications (e.g., education, on-the-go interaction, medicine, smart sensing).
Send submissions to firstname.lastname@example.org before the end of May 19th (friday, 24:00 AOE, using ACM SIGCHI format. Submissions will be peer reviewed by an international program committee. Submissions do not have to be anonymous.
Topics of interest
Broadly speaking this workshop is interested in object recognition work based on computer vision, radar (e.g. Project Soli by the Google ATAP (Advanced Technology and Projects)), acoustic sensing, tagging, smart objects etc. Participants are encouraged to bring their mobile and desktop based systems suitable for object recognition for mobile interaction. The following are example topics of interest but we are not limited to these:
- Understanding the design space and identifying factors that influence user interactions in this space
- Developing evaluation strategies to cope with the complex nature of computer vision or radar-based object interaction
- Ethnography and user studies of tangible and object interaction
- Examples of applications of object interaction
- Social factors that influence the design of suitable interaction techniques for object interaction
- Exploring interaction techniques that facilitate multi-person object interaction
- Novel input mechanisms for single and multi-point sensing systems (eg. in mobile, tablet and wrist worn devices)
- SDK/APIs, IDEs, and hardware platforms for the development of object recognition.
International Program Committee
- Andrea Bianchi, KAIST, Korea
- Nicholas Gillian, ATAP Google, USA
- Chris Harrison, Carnegie Mellon University, USA
- Niels Henze, University of Stuttgart, Germany
- Gierad Laput, Carnegie Mellon University, USA
- Rong-Hao Liang, Eindhoven University of Technology, Netherlands
- Aaron Quigley, University of St Andrews, UK
- Kaisa Vaananen, Tampere University of Technology, Finland
- Hui-Shyong Yeo, University of St Andrews, UK
- Liwei Chan, National Chiao Tung University, Taiwan
- Cheng Zhang, Georgia Tech, USA
- Chouchang (Jack) Yang, Disney Research
Last updated April 20th.