Illusionary Surface Interfaces is a research project within the Scalable Interaction Paradigms for Pervasive Computing Environments (SPP2199). The project aims to investigate how haptic sensations can be induced with less hardware as possible, as future technologies might rely on interfaces where haptic illusions are necessary to provide users with haptic feedback. Traditional computers typically refer (through visual affordances) to both the perceived and actual properties of the interface – suggesting not only fundamental functionalities, but also determining and communicating how humans possibly use the system. Such rich information visualization may, however, not suit the way we want pervasive computers and computational everyday environments to look and feel. We aim to create novel interactive experiences, that exploit multisensory illusions in order to extend the range of interface properties that can be displayed, using only everyday object surfaces as interfaces. In a manner similar to the “rubber hand illusion”, in which people can be induced to perceive a physical touch based purely on what they see, we will support visual and haptic feedback induced by augmented vision and sound. Instead of changing the objects’ physicality, we will visually and auditory augment them using “smart glasses” and projectors, while at the same time augmenting them haptically by inducing multisensory illusion. Technically, this includes sensing user interaction using machine learning tools and multimodal presentation of information.