This project, Visual Media Adaptation
for AToM, aims to develop the active adaptation
technology for visual media such as image & video, and also
makes users consume multimedia contents seamlessly according
to user preference, network characteristic, and terminal capability.
¡¡
The goal of this project is the followings
¡¡
1. A general solution
for universal multimedia access for the near future development
Visual Media Adaptation
system for AToM
¡¡
2. An initial testbed
for demonstration of multimedia content adaptation
Active Color Information
adaptation for active media
Video structure and
Visual Information adaptation according to user characteristics
Building Usage-case
scenario for AToM
¡¡
Fig.
1. Use case scenario for Active Room
¡¡
Fig. 1 shows
Active Room in which user can seamlessly consume active media
according to environments characteristics.In the active room,
there are many different terminals, HDTVs, PCs, notebooks, and
PDAs located at different places. The terminals are equipped
with visual and audio sensors to recognize users and allow user-machine
interaction.
¡¡
Application Scenario
In the active room, there are many different terminals, HDTVs,
PCs, notebooks, and PDAs located at different places. The terminals
are equipped with visual and audio sensors to recognize users
and allow user-machine interaction via speeches and gestures.The
user comes to an active media and communicates with the active
media about his need of certain content, e.g. a song or a movie.
The active media recognizes the user by his face and voice or
using any kinds of sensing machines, without asking for a user
name and password. Then the active media are being adapted according
to the user.
Later, the user moves to a bigger terminal, e.g. the HDTV. Even
the user does not say anything, the active media can easily
realize that the user is approaching and it activates a display
job. It quickly realizes his last session state. Then the HDTV
automatically continues to present the session to user, in a
seamless way. Of course at this time the user can enjoy the
active media with the high quality of visual information.
When the user leaves the active room, i.e., he/she cannot support
LAN environment, the user can access adaptation server by mobile
or PDA phone and then he/she can continue to consume the same
active media according to current environment. For example,
a movie can be adapted into a storyboard consisting of images
and text. Moreover, even if the user leaves the active room
and returns some day later, the active media can continue to
serve him without any annoying authentication and authorization.
¡¡
Fig. 2 shows
the adaptation system architecture for Active Room. In this
system, adaptation server consists of three modules like Constraint
Detection and Exchanger, Adaptation Engine, and Repository,
in general.
¡¡
Fig.
2. Agents Working Flow in each Platform
¡¡
The function of the three modules mentioned above is the followings;
¡¡
Constraint Detection
and Exchanger :
Collect the network characteristics,
device capabilities, user preferences and session
information for the correct and efficient adaptation
Session information can contain usage
history for the specific user
¡¡
Adaptation Engine :
Select the appropriate
transcodings or content versions
Provide transcoding algorithms
for resources or contents based on active media
¡¡
Repository :
Store resources, descriptions,
and a variety of user constraints
¡¡
¡¡
Jaeil Song, Seungji Yang,
Cheonseog Kim, Jeho Nam, Jin Woo Hong and Yong Man Ro, "Digital
Item Adaptation for Color Vision Variations", Proceeding
on SPIE Electronic Imaging, Jan, 2003
Yong Ju Jung, Truong Cong
Thang, Jeawook Lee, and Yong Man Ro. "Visual Media Adaptation
system for active media", The 2003 International Conference
on Imaging Science, Systems, and Technology (CISST¡¯03), June,
2003
T. C. Thang, Y. J. Jung,
Y. M. Ro, ¡°Modality conversion in content adaptation for Universal
multimedia access¡±, Int. Conf. on Imaging Science, Systems,
and Technology, Nevada, US, 2003