As a part of the SmartKom project, subjects are recorded while using a self-explanatory, user-adaptive man-computer interface. This interface is supposed to interpret voice and gesture input, as well as the user's mimic. Within the man-machine dialogues the subjects are supposed to handle various tasks in various scenarios, like asking for the way or today's movie. As a result we get on the one hand video-data and on the other hand multi-channel data of spontaneous speech, which serves as a base for research and development in the field of speech recognition, speech synthesis and user-interface models.