publications([{ "lang": "en", "type_publi": "icolcomlec", "doi": "http://dx.doi.org/10.1145/2909132.2909260", "title": "Designing 3D Gesture Guidance: Visual Feedback and Feedforward Design Options", "abstract": "Dynamic symbolic in-air hand gestures are an increasingly popular means of interaction with smart environments. However, novices need to know what commands are available and which gesture to execute in order to trigger these commands. We propose to adapt OctoPocus, a 2D gesture guiding system, to the case of 3D. The OctoPocus3D guidance system displays a set of 3D gestures as 3D pipes and allows users to understand how the system processes gesture input. Several feedback and feedforward visual alternatives are proposed in the literature. However, their impact on guidance remains to be evaluated. We report the results of two user experiments that aim at designing OctoPocus3D by exploring these alternatives. The results show that a concurrent feedback, which visually simplifies the 3D scene during the execution of the gesture, increases the recognition rate, but only during the first two repetitions. After the first two repetitions, users achieve the same recognition rate with a terminal feedback (after the execution of the gesture), a concurrent feedback, both or neither. With respect to feedforward, the overall stability of the 3D scene explored through the origin of the pipes during the execution of the gestures does not influence the recognition rate or the execution time. Finally, the results also show that displaying upcoming portions of the gestures allows 8% faster completion times than displaying the complete remaining portions. This indicates that preventing visual clutter of the 3D scene prevails over gesture anticipation.", "authors": { "1": { "first_name": "William", "last_name": "Delamare" }, "2": { "first_name": "Céline", "last_name": "Coutrix" }, "3": { "first_name": "Laurence", "last_name": "Nigay" } }, "year": 2016, "uri": "http://iihm.imag.fr/publication/DCN16a/", "pages": "152-159", "bibtype": "inproceedings", "id": 768, "abbr": "DCN16a", "address": "Bari, Italy", "date": "2016-06-07", "document": "http://iihm.imag.fr/publs/2016/DelamareetalAVI2016.pdf", "type": "Conférences internationales de large diffusion avec comité de lecture sur texte complet", "booktitle": "Proceedings of the International Working Conference on Advanced Visual Interfaces (AVI'16), Bari, Italy, 7-10 June, 2016" }, { "lang": "en", "type_publi": "icolcomlec", "doi": "http://dx.doi.org/10.1145/2774225.2774847", "title": "Designing Guiding Systems for Gesture-Based Interaction", "abstract": "2D or 3D gesture commands are still not routinely adopted, despite the technological advances for tracking gestures. The fact that gesture commands are not self-revealing is a bottleneck for this adoption. Guiding novice users is therefore crucial in order to reveal what commands are available and how to trigger them. However guiding systems are mainly designed in an ad hoc manner. Even if isolated design characteristics exist, they concentrate on a limited number of guidance aspects. We hence present a design space that unifies and completes these studies by providing a coherent set of issues for designing the behavior of a guiding system. We distinguish Feedback and Feedforward and consider four questions: When, What, How and Where. In order to leverage efficient use of our design space, we provide an online tool and illustrate with scenarios how practitioners can use it.", "authors": { "1": { "first_name": "William", "last_name": "Delamare" }, "2": { "first_name": "Céline", "last_name": "Coutrix" }, "3": { "first_name": "Laurence", "last_name": "Nigay" } }, "year": 2015, "uri": "http://iihm.imag.fr/publication/DCN15a/", "pages": "44-53 ", "bibtype": "inproceedings", "id": 725, "abbr": "DCN15a", "address": "Duisburg, Germany", "date": "2015-04-09", "document": "http://iihm.imag.fr/publs/2015/taxoguide_finalDOI2.pdf", "type": "Conférences internationales de large diffusion avec comité de lecture sur texte complet", "booktitle": "Proceedings of the 7th ACM SIGCHI Symposium on Engineering Interactive Computing Systems (EICS 2015)" }, { "lang": "fr", "type_publi": "these", "doi": "https://tel.archives-ouvertes.fr/tel-01270557", "title": "Interaction à distance en environnement physique augmenté", "abstract": "Nous nous intéressons à l’interaction dans le contexte d’environnements physiques augmentés, plus précisément avec les objets physiques qui les composent. Bien que l’augmentation de ces objets offre de nouvelles possibilités d’interaction, notamment celle d’interagir à distance, le monde physique possède des caractéristiques propres rendant difficile l’adaptation de techniques d’interaction existantes en environnements virtuels. Il convient alors d’identifier ces caractéristiques afin de concevoir des techniques d’interaction à la fois efficaces et plaisantes dédiées à ces environnements physiques augmentés. Dans nos travaux, nous décomposons cette interaction à distance avec des objets physiques augmentés en deux étapes complémentaires : la sélection et le contrôle. Nous apportons deux contributions à chacun de ces champs de recherche. Ces contributions sont à la fois conceptuelles, avec la création d’espaces de conception, et pratiques, avec la conception, la réalisation logicielle et l’évaluation expérimentale de techniques d’interaction :\r\n\r\n- Pour l’étape de sélection, nous explorons la désambiguïsation potentielle après un geste de pointage à distance définissant un volume de sélection comme avec une télécommande infrarouge par exemple. En effet, bien que ce type de pointage sollicite moins de précision de la part de l’utilisateur, il peut néanmoins impliquer la sélection de plusieurs objets dans le volume de sélection et donc nécessiter une phase de désambiguïsation. Nous définissons et utilisons un espace de conception afin de concevoir et évaluer expérimentalement deux techniques de désambiguïsation visant à maintenir l’attention visuelle de l’utilisateur sur les objets physiques.\r\n- Pour l’étape de contrôle, nous explorons le guidage de gestes 3D lors d’une interaction gestuelle afin de spécifier des commandes à distance. Ce guidage est nécessaire afin d’indiquer à l’utilisateur les commandes disponibles ainsi que les gestes associés. Nous définissons un espace de conception capturant les caractéristiques comportementales d’un large ensemble de guides ainsi qu’un outil en ligne facilitant son utilisation. Nous explorons ensuite plusieurs options de conception afin d’étudier expérimentalement leurs impacts sur la qualité du guidage de gestes 3D.", "year": 2015, "uri": "http://iihm.imag.fr/publication/D15a/", "id": 752, "bibtype": "phdthesis", "abbr": "D15a", "authors": { "1": { "first_name": "William", "last_name": "Delamare" } }, "date": "2015-12-10", "document": "http://iihm.imag.fr/publs/2015/thesis-delamare-valid.pdf", "type": "Thèses et habilitations", "pages": "189" }, { "lang": "en", "type_publi": "autre", "doi": "http://dx.doi.org/10.1145/2774225.2775440", "title": "A Tool for Optimizing the Use of a Large Design Space for Gesture Guiding Systems", "abstract": "We present a tool to help practitioners to characterize, compare and design gesture guiding systems. The tool can be used to find an example system meeting specific requirements or to start exploring an original research area based on unexplored design options. The motivation for the online tool is the large underlying design space including 35 design axes: the tool therefore helps explore and combine the various design options. Moreover the tool currently includes the description of 46 gesture guiding systems: the tool is thus also a repository of existing gesture guiding systems.", "authors": { "1": { "first_name": "William", "last_name": "Delamare" }, "2": { "first_name": "Céline", "last_name": "Coutrix" }, "3": { "first_name": "Laurence", "last_name": "Nigay" } }, "year": 2015, "uri": "http://iihm.imag.fr/publication/DCN15b/", "pages": "238-241", "bibtype": "unpublished", "id": 728, "abbr": "DCN15b", "address": "Duisburg, Germany", "date": "2015-05-18", "document": "http://iihm.imag.fr/publs/2015/DemoEICS2015-delamare.pdf", "type": "Autres publications", "booktitle": "Proceedings of the 7th ACM SIGCHI Symposium on Engineering Interactive Computing Systems (EICS 2015)" }, { "lang": "fr", "type_publi": "colloque", "title": "Techniques de Pointage à Distance : Cibles Numériques et Cibles Physique", "url": "https://ubimob2014.sciencesconf.org/42778", "abstract": "Au sein d’un environnement ubiquitaire, l’ordinateur devient évanescent : nos objets quotidiens sont augmentés d’électronique, les environnements deviennent perceptifs déconfinant l’interaction homme-machine de l’ancien ordinateur «boîte grise» à des espaces pervasifs. Désormais, l’utilisateur évolue dans un monde physico-numérique ou espace interactif mixte. Au sein de cet espace interactif, un besoin est alors d’interagir à distance que ce soit pour manipuler des objets numériques sur un écran distant ou des objets physiques. Cet article est dédié aux techniques de pointage à distance pour désigner un objet numérique ou physique. Nous décrivons six techniques de pointage pour interagir dans un environnement ubiquitaire, la première pour pointer à distance sur des cibles numériques, les cinq autres pour pointer sur des objets physiques avec et sans un dispositif mobile.", "authors": { "1": { "first_name": "Céline", "last_name": "Coutrix" }, "2": { "first_name": "William", "last_name": "Delamare" }, "3": { "first_name": "Maxime", "last_name": "Guillon" }, "4": { "first_name": "Takeshi", "last_name": "Kurata" }, "5": { "first_name": "François", "last_name": "Leitner" }, "6": { "first_name": "Laurence", "last_name": "Nigay" }, "7": { "first_name": "Thomas", "last_name": "Vincent" } }, "year": 2014, "uri": "http://iihm.imag.fr/publication/CDG+14a/", "pages": "5", "bibtype": "inproceedings", "id": 691, "abbr": "CDG+14a", "address": "Nice, France", "date": "2014-06-05", "document": "http://iihm.imag.fr/publs/2014/UBIMOB2014-DistantPointing.pdf", "type": "Autres conférences et colloques avec actes", "booktitle": "UbiMob2014 : 10èmes journées francophones Mobilité et Ubiquité" }, { "lang": "en", "type_publi": "icolcomlec", "doi": "http://doi.acm.org/10.1145/2480296.2480309", "title": "Designing Disambiguation Techniques for Pointing in the Physical World", "abstract": "Several ways for selecting physical objects exist, including touching and pointing at them. Allowing the user to interact at a distance by pointing at physical objects can be challenging when the environment contains a large number of interactive physical objects, possibly occluded by other everyday items. Previous pointing techniques highlighted the need for disambiguation techniques. Addressing this challenge, this paper contributes a design space that organizes along groups and axes a set of options for designers to relevantly (1) describe, (2) classify, and (3) design disambiguation techniques. First, we have not found techniques in the literature yet that our design space could not describe. Second, all the techniques show a different path along the axes of our design space. Third, it allows defining of several new paths/solutions that have not yet been explored. We illustrate this generative power with the example of such a designed technique, Physical Pointing Roll (P2Roll).", "authors": { "1": { "first_name": "William", "last_name": "Delamare" }, "2": { "first_name": "Céline", "last_name": "Coutrix" }, "3": { "first_name": "Laurence", "last_name": "Nigay" } }, "year": 2013, "uri": "http://iihm.imag.fr/publication/DCN13a/", "pages": "197-206", "bibtype": "inproceedings", "id": 628, "abbr": "DCN13a", "address": "London, UK", "date": "2013-04-22", "document": "http://iihm.imag.fr/publs/2013/finalEICS2013.pdf", "type": "Conférences internationales de large diffusion avec comité de lecture sur texte complet", "booktitle": "Proceedings of the 5th ACM SIGCHI Symposium on Engineering Interactive Computing Systems (EICS 2013)" }, { "lang": "en", "type_publi": "icolcomlec", "doi": "http://dx.doi.org/10.1145/2493190.2493232", "title": "Mobile Pointing Task in the Physical World: Balancing Focus and Performance while Disambiguating", "abstract": "We address the problem of mobile distal selection of physical objects when pointing at them in augmented environments. We focus on the disambiguation step needed when several objects are selected with a rough pointing gesture. A usual disambiguation technique forces the users to switch their focus from the physical world to a list displayed on a handheld device’s screen. In this paper, we explore the balance between change of users’ focus and performance. We present two novel interaction techniques allowing the users to maintain their focus in the physical world. Both use a cycling mechanism, respectively performed with a wrist rolling gesture for P2Roll or with a finger sliding gesture for P2Slide. A user experiment showed that keeping users’ focus in the physical world outperforms techniques that require the users to switch their focus to a digital representation distant from the physical objects, when disambiguating up to 8 objects. ", "authors": { "1": { "first_name": "William", "last_name": "Delamare" }, "2": { "first_name": "Céline", "last_name": "Coutrix" }, "3": { "first_name": "Laurence", "last_name": "Nigay" } }, "year": 2013, "uri": "http://iihm.imag.fr/publication/DCN13b/", "pages": "89-98", "bibtype": "inproceedings", "id": 631, "abbr": "DCN13b", "address": "Munich, Germany", "date": "2013-07-01", "document": "http://iihm.imag.fr/publs/2013/FinalMobileHCI.pdf", "type": "Conférences internationales de large diffusion avec comité de lecture sur texte complet", "booktitle": "Proceedings of the 15th International Conference on Human-Computer Interaction with Mobile Devices and Services (MobileHCI 2013)" }, { "lang": "en", "type_publi": "colloque", "title": "Pointing in the Physical World for Light Source Selection", "abstract": "We focus on the selection of light sources in the physical world. Their selection is challenging for the user, since numerous Light-Emitting Diodes (LEDs) can be embedded into various materials as well as environments, thus creating high densities of interactive objects. In this paper, we describe an innovative technique for light sources selection based on the pointing paradigm, that is, allowing interaction at a distance. To address the limitations of the pointing paradigm (e.g., aiming at distant and/or small targets), we design a two-step pointing technique: a rough aiming with an arm pointing gesture and a disambiguation mechanism with a wrist rolling gesture. Feedbacks lean on the various capacities of LED lights. We expect that our technique is well suited for the selection task in dense environments, no matter how small and how distant the targeted light sources are. We also expect that the technique supports an efficient interaction based on proprioception and muscular memory properties for expert users, who may perform the two interaction steps by a single combined gesture for better performance.", "address": "Newcastle, UK", "year": 2012, "uri": "http://iihm.imag.fr/publication/DCN12a/", "id": 588, "bibtype": "inproceedings", "abbr": "DCN12a", "authors": { "1": { "first_name": "William", "last_name": "Delamare" }, "2": { "first_name": "Céline", "last_name": "Coutrix" }, "3": { "first_name": "Laurence", "last_name": "Nigay" } }, "date": "2012-05-03", "document": "http://iihm.imag.fr/publs/2012/workshopDIS12.pdf", "type": "Autres conférences et colloques avec actes", "booktitle": "Proceedings of Designing Interactive Lighting workshop at DIS 2012" }]);