umu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Turn your mobile into the football: rendering live football game by vibration
Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
Umeå University, Faculty of Science and Technology, Department of Applied Physics and Electronics.
2008 (English)In: IEEE transactions on multimedia, ISSN 1520-9210, E-ISSN 1941-0077, Vol. 10, no 6, 1022-1033 p.Article in journal (Refereed) Published
Abstract [en]

Vibration offers many potential benefits for the use of mobile phones. In this paper, we propose a new method of rendering live football game on mobile phones using vibration. A mobile phone is “synchronized” with the ball in the real field. By holding the phone, users are able to experience dynamic movements of the ball, to know attacking directions and which team is leading the attack. The usability test of our system shows that vibrotactile display is suitable for rendering live football information on mobile phones by adopting designed coding schemes with a right training process.

Place, publisher, year, edition, pages
2008. Vol. 10, no 6, 1022-1033 p.
National Category
Physical Sciences
Research subject
Physics
Identifiers
URN: urn:nbn:se:umu:diva-19256DOI: 10.1109/TMM.2008.2001352OAI: oai:DiVA.org:umu-19256DiVA: diva2:201627
Available from: 2009-03-05 Created: 2009-03-05 Last updated: 2017-12-13Bibliographically approved
In thesis
1. Expressing emotions through vibration for perception and control
Open this publication in new window or tab >>Expressing emotions through vibration for perception and control
2010 (English)Doctoral thesis, comprehensive summary (Other academic)
Alternative title[en]
Expressing emotions through vibration
Abstract [en]

This thesis addresses a challenging problem: “how to let the visually impaired ‘see’ others emotions”. We, human beings, are heavily dependent on facial expressions to express ourselves. A smile shows that the person you are talking to is pleased, amused, relieved etc. People use emotional information from facial expressions to switch between conversation topics and to determine attitudes of individuals. Missing emotional information from facial expressions and head gestures makes the visually impaired extremely difficult to interact with others in social events. To enhance the visually impaired’s social interactive ability, in this thesis we have been working on the scientific topic of ‘expressing human emotions through vibrotactile patterns’.

It is quite challenging to deliver human emotions through touch since our touch channel is very limited. We first investigated how to render emotions through a vibrator. We developed a real time “lipless” tracking system to extract dynamic emotions from the mouth and employed mobile phones as a platform for the visually impaired to perceive primary emotion types. Later on, we extended the system to render more general dynamic media signals: for example, render live football games through vibration in the mobile for improving mobile user communication and entertainment experience. To display more natural emotions (i.e. emotion type plus emotion intensity), we developed the technology to enable the visually impaired to directly interpret human emotions. This was achieved by use of machine vision techniques and vibrotactile display. The display is comprised of a ‘vibration actuators matrix’ mounted on the back of a chair and the actuators are sequentially activated to provide dynamic emotional information. The research focus has been on finding a global, analytical, and semantic representation for facial expressions to replace state of the art facial action coding systems (FACS) approach. We proposed to use the manifold of facial expressions to characterize dynamic emotions. The basic emotional expressions with increasing intensity become curves on the manifold extended from the center. The blends of emotions lie between those curves, which could be defined analytically by the positions of the main curves. The manifold is the “Braille Code” of emotions.

The developed methodology and technology has been extended for building assistive wheelchair systems to aid a specific group of disabled people, cerebral palsy or stroke patients (i.e. lacking fine motor control skills), who don’t have ability to access and control the wheelchair with conventional means, such as joystick or chin stick. The solution is to extract the manifold of the head or the tongue gestures for controlling the wheelchair. The manifold is rendered by a 2D vibration array to provide user of the wheelchair with action information from gestures and system status information, which is very important in enhancing usability of such an assistive system. Current research work not only provides a foundation stone for vibrotactile rendering system based on object localization but also a concrete step to a new dimension of human-machine interaction.

Place, publisher, year, edition, pages
Umeå: Umeå universitet, Institutionen för tillämpad fysik och elektronik, 2010. 159 p.
Series
Digital Media Lab, ISSN 1652-6295 ; 12
Keyword
Multimodal Signal Processing, Mobile Communication, Vibrotactile Rendering, Locally Linear Embedding, Object Detection, Human Facial Expression Analysis, Lip Tracking, Object Tracking, HCI, Expectation-Maximization Algorithm, Lipless Tracking, Image Analysis, Visually Impaired.
National Category
Signal Processing Computer Vision and Robotics (Autonomous Systems) Computer Science Telecommunications Information Science
Research subject
Computerized Image Analysis; Computing Science; Electronics; Systems Analysis
Identifiers
urn:nbn:se:umu:diva-32990 (URN)978-91-7264-978-1 (ISBN)
Public defence
2010-04-28, Naturvetarhuset, N300, Umeå universitet, Umeå, Sweden, 09:00 (English)
Opponent
Supervisors
Projects
Taktil Video
Available from: 2010-04-07 Created: 2010-04-06 Last updated: 2010-04-20Bibliographically approved

Open Access in DiVA

No full text

Other links

Publisher's full texthttp://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=4637879

Search in DiVA

By author/editor
ur Réhman, ShafiqLi, Haibo
By organisation
Department of Applied Physics and Electronics
In the same journal
IEEE transactions on multimedia
Physical Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 230 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf