Sunday
Registration
Sunday 7:30 AM - 1:30 PM, 4:00 PM - 7:00 PM
Breakfast
Sunday 7:30 AM - 8:30 AM
Morning Tutorials
Sunday 8:30 AM - 12:00 PM
(Break from 10:00 AM - 10:30 AM)
Lunch
Sunday 12:00 PM - 1:00 PM
Afternoon Tutorials/CREST Workshop
Sunday 1:00 PM - 4:30 PM
(Break from 2:30 PM - 3:00 PM)
Design Contest
Sunday 6:00 PM - 8:00 PM
Monday
Registration
Monday 7:30 AM - 3:30 PM
Breakfast
Monday 7:30 AM - 8:30 AM
Opening Remarks
Monday 8:30 AM - 8:45 AM
Keynote
Monday 8:45 AM - 9:30 AM
- Appropriate Dress Required: Wearable Computing
and Context
- Bill Buxton
Session 1: Wellness
Monday 9:30 AM - 10:15 AM
- Less Contact: Heart-Rate Detection
Without Even Touching the User
- Florian Michahelles, Ramon Wicki, Bernt
Schiele
- Wearable Sensors for Auto-Event-Recording on
Medical Nursing - User Study of Ergonomic Design
- Haruo Noma, Aki Ohmura, Noriaki Kuwahara,
Kiyoshi Kogure
Break
Monday 10:15 AM - 10:45 AM
Session 2: Gesture
Monday 10:45 AM - 11:50 AM
- FreeDigiter: A Contact-Free Device for Gesture
Control
- Christian Metzger, Matt Anderson, Thad
Starner
- A Robust Hand Tracking for Gesture-Based
Interaction of Wearable Computer
- Yang Liu, Yunde Jia
- Improving Scalability of Sign Language
Recognition Systems: A Hybrid Approach
- Van Culver
Design Contest Award Presentation
Monday 11:50 AM - 12:00 PM
Lunch
Monday 12:00 PM - 1:30 PM
Session 3: Textiles
Monday 1:30 PM - 3:00 PM
- Intelligent Life Clothing - FICC (Floatable
Intelligent and Communicative Clothing) Project
- Vladan Koncar, Bohwon Kim, Emmanuel Bilala
Nebor, Xavier Joppin
- Fibre-Meshed Transducers Based a Real Time
Wearable Physiological Information Monitoring System
- Ravindra Wijesiriwardana
- E-Textiles for Autonomous Location Awareness
- Madhup Chandra, Mark T. Jones, Thomas L.
Martin
- The Design and Deployment of a Wearable
Vibrotactile Feedback System
- Robert Lindeman, John Sibert, Corinna Lathan,
Jack Vice
Break
Monday 3:00 PM - 3:30 PM
Session 4: System
Monday 3:30 PM - 5:10 PM
- Remote Collaboration using a Shoulder-Worn
Active Camera/Laser
- Takeshi Kurata, Nobuchika Sakata, Masakatsu
Kourogi, Hideki Kuzuoka, Mark Billinghurst
- An Event-driven Wearable System for Supporting
Motorbike Races
- Masakazu Miyamae, Tsutomu Terada, Masahiko
Tsukamoto, Keisuke Hiraoka, Takahito Fukuda, Shojiro Nishio
- Tracking of User Position and Orientation by
Stereo Measurement of Infrared Markers and Orientation Sensing
- Masaki Maeda, Takefumi Ogawa, Kiyoshi
Kiyokawa, Haruo Takemura
- Personal Mobile Hub
- Dirk Husemann, Chandra Narayanaswami, Michael
Nidd
Posters and Demo Session
Monday 5:30 PM - 7:30 PM
Reception
Monday 7:00 PM - 10:00 PM
Tuesday
Registration
Tuesday 7:30 AM - 3:30 PM
Breakfast
Monday 7:30 AM - 8:30 AM
Session 1: HCI1
Tuesday 8:45 AM - 10:00 AM
- Expert Chording Text Entry on the Twiddler One-Handed
Keyboard
- Kent Lyons, Daniel Plaisted, Thad Starner
- Analysis of Wearable Interface Factors for
Appropriate Information Notification
- Vlaho Kostov, Jun Ozawa, Satoshi Matsuura
- A Comparative Investigation into Two Pointing
Systems for use with Wearable Computers While Mobile
- Alan Chamberlain, Roy Kalawsky
Break
Tuesday 10:00 AM - 10:30 AM
Session 2: HCI2
Tuesday 10:30 AM - 11:30 AM
- An Interaction System for Watch Computers Using
Tactile Guidance and Bidirectional Segmented Strokes
- Gabor Blasko, Steven Feiner
- Evaluating Techniques for Interaction at a
Distance
- Jason Wither, Tobias Höllerer
- My Own Private Kiosk: Privacy-Preserving Public
Displays
- Marc Eaddy, Gabor Blasko, Jason Babcock, Steven
Feiner
- Visual Memory Augmentation: Using Eye Gaze as
an Attention Filter
- Deb Roy, Yair Ghitza, Jeff Bartelma, Charlie
Kehoe
Lunch
Tuesday 12:00 PM - 1:30 PM
Session 3: Context
Tuesday 1:30 PM - 3:05 PM
- Implementation and Evaluation of a Low-Power
Sound-Based User Activity Recognition System
- Mathias Staeger, Paul Lukowicz, Gerhard
Troester
- Pine versus Porcupine: a Study in
Distributed Wearable Activity Recognition
- Kristof Van Laerhoven, Hans-Werner
Gellersen
- Methods for Interrupting a Wearable Computer User
- Mikael Drugge, Marcus Nilsson, Urban Liljedahl,
Kåre Synnes, Peter Parnes
- A Model for Human Interruptability:
Experimental Evaluation and Automatic Estimation from Wearable
Sensors
- Nicky Kern, Stavros Antifakos, Bernt Schiele,
Adrian Schwaninger
Break
Tuesday 3:05 PM - 3:30 PM
Gadget Show and One-Minute Madness
Tuesday 3:30 PM - 4:30 PM
Business Meeting
Tuesday 4:30 PM - 5:00 PM
Student Colloquium
Tuesday 6:00 PM - 8:00 PM
Wednesday
Registration
Wednesday 8:00 AM - 4:00 PM
Breakfast
Wednesday 8:00 AM - 8:45 AM
ISWC/ISMAR Opening Remarks
Wednesday 8:45 AM - 9:00 AM
Joint ISWC/ISMAR Keynote
Wednesday 9:00 AM - 10:00 AM
Break
Wednesday 10:00 AM - 10:30 AM
Panel Session: Opportunities in Wearable
Computing and Augmented Reality
Wednesday 10:30 AM - 12:00 PM
The European Union has committed almost a quarter billion
dollars to research in fields related to mobile and ubiquitous
computing over the next few years. Countries in Asia are
beginning to devote considerable resources as well. With the
mobile phone now the dominant consumer computer platform in
the world, wearable computing and augmented reality are
becoming business domains as well as research fields.
This panel of program managers, invited from various funding
agencies, will discuss how wearable computing and augmented
reality relates to their fields of interest. A mixture of
scientific, military, government, and private agencies will be
represented. Attendees are invited to ask questions and help
explore how wearable computing and augmented reality research
can be promoted in upcoming initiatives.
ISWC Closing Remarks
Wednesday 12:00 PM - 12:15 PM
Lunch
Wednesday 12:15 PM - 1:30 PM
Keynote: "Appropriate Dress Required: Wearable
Computing and Context"
Bill Buxton
Bill Buxton is a Canadian designer, researcher, teacher and
writer. However, he began his career as a musician. In the late
60's and early 70's he became deeply involved in making music with
synthesizers and computers, which introduced him to the world of
technology, with all of its potential and all of its pitfalls.
Frustration being the mother of invention, he increasingly turned
his attention to understanding better ways to design, and interact
with computers. His work caught the eye of researcher's at Xerox
PARC, which led him increasingly into the world of research, as
well as extending the scope of his interests beyond music.
From 1987-1994 he was a research scientist at RankXerox EuroPARC
Cambridge, and XEROX's Palo Alto Research Center (PARC). From
1994 until December 2002, he was Chief Scientist of
Alias|Wavefront, and from 1995, its parent company SGI Inc. He is
currently Principal of his own boutique design and consulting
firm, Buxton Design, where his time is split between working for
clients, writing and lecturing. He also has an association with
Bruce Mau Design of Toronto, where he acts as Chief Scientist.
Buxton is an Associate Professor in the Department of Computer
Science at the University of Toronto and Visiting Professor at the
Knowledge Media Design Institute. In the fall of 2004, he will
also be a lecturer in the Department of Industrial Design at the
Ontario College of Art and Design.
In 1995, Buxton became the third recipient of the Canadian
Human-Computer Communications Society Award for contributions to
research in computer graphics and human-computer interaction, and
was given the New Media
Visionary of the Year Award at the 2000 Canadian New Media Awards.
In 2002, he was elected to the CHI Academy, and Time Magazine
named him one of the top 5 designers in Canada. In 2001, The
Hollywood Reporter named him one of the 10 most influential
innovators in Hollywood.
More information on Buxton and his work can be found at:
http://www.billbuxton.com
Keynote: "Augmenting This...Augmented That:
Maximizing Human Performance"
Through the DARPA sponsored efforts in the field of Augmented
Cognition, substantial progress has been made on the development
of technologies for sensing brain activity in real time.
Further refinement has focused on ruggedizing the sensors and
improving their wearability for the operational enviroment. As
efforts in Augmented Cognition continue to develop for military
and civilian applications, new opportunities are appearing on
the horizon to develop similar technologies for the training and
education environment. In particular, efforts at the Office of
Naval Research in the Human Performance Training and Education
thrust area are beginning to merge cognitive mediators with
augmented reality to begin to tackle the complex arena of
situational awareness in the mobile computing environment. This
talk will discuss new efforts underway and challenge the
audience to think about novel ways to maximize human performance
on the battlefield of tomorrow.
LCDR Dylan Schmorrow, PhD - DARPA and the Office of
Naval Research
LCDR Schmorrow is a U.S. Naval Officer and is an Aerospace
Experimental Psychologist in the Navy's Medical Service Corps.
He is currently a Program Manager serving at the Defense
Advanced Research Projects Agency (DARPA) where he is
responsible for executing cutting-edge basic science and
technology development. In this role, he supports DARPA's
mission to create and foster imaginative, innovative, and often
high-risk research ideas yielding revolutionary technological
advances in science and technology in support of the U.S.
military. He is currently directing both basic and applied,
research and development programs addressing human-technology
integration. His programs focus on improving warfighter
information intake under stress and extending agent based
computing into existing military computer technologies and
applications. LCDR Schmorrow also supports the Office of Naval
Research (ONR), where he coordinates, executes, and promotes
science and technology programs of the U.S. Navy and Marine
Corps. At ONR he is tasked with advancing basic and applied
science, and the transforming of promising technologies into
Navy and Marine Corps operational capabilities. His research
interests range from the integration of neuroscience,
human-factors and decision support technologies that foster
novel brain-machine-symbiosis for maximizing human cognition and
performance, to the alignment of DoD policy and emergent
technology capabilities.
Dr. Amy Kruse - Strategic Analysis
Dr. Amy Kruse currently serves as the Technical Director for the
Improving Warfighter Information Intake Under Stress-Augmented
Cognition program at DARPA. In this role she has championed the
use of neuroscientific techniques in the operational/applied
environment. In addition to her work at DARPA, Dr. Kruse serves
at the Technical Advisor to the Human Performance Training and
Education thrust area at the Office of Naval Research. Prior to
her work at DARPA, Dr. Kruse conducted NIH funded post-doctoral
research at the Beckman Institute on the University of Illinois
Urbana-Champaign campus. Her areas of expertise include learning
and memory, neurophysiology and fNIR technologies. Dr. Kruse
holds a BS in Cell and Structural Biology and a Ph.D. in
Neuroscience from the University of Illinois. During her doctoral
training she was an NSF Graduate Fellow in Neuroscience. She is a
member of IEEE, SPIE, the Cognitive Science Society and the
Society for Neuroscience.
Wearable Computers for Persons with Disabilities
Maribeth Gandy - Georgia Institute of Technology
Abstract:
The goal of this tutorial is to provide attendees with a survey of
issues related to wearable computing for persons with disabilities.
Topics to be covered will include a discussion of why accessibility
and universal design are important, along with information on current
accessibility techniques and relevant legislation. The research
activities of the recently established Wireless Rehabilitation
Engineering Research Center (RERC), which focuses on wireless and
mobile devices for people with disabilities, will be summarized.
There will also be a detailed look at previous wearable research in
the disability domain. The process of designing a wearable system for
persons with disabilities will be discussed along with a relevant case
study that illustrates some of the unique challenges in this type of
research. Lastly, the tutorial will focus on the future directions
that wearable computing for disabilities can take with the goal of
motivating attendees to consider these issues while carrying out their
research. The intended audience for this tutorial includes
researchers and developers in wearable or mobile computing who are
interested in learning more about how their work can aid those with
disabilities. However, a high level of technical knowledge will not
be necessary to benefit from this tutorial.
Bio:
Maribeth Gandy is a Research Scientist with the Interactive Media
Technology Center (IMTC) at Georgia Tech. She is involved with many
biomedical and wearable related research projects. In the past she
was a co-author on papers regarding such topics as audio-only wearable
computing and a wearable device for universal control. She is also
involved in an NIH funded grant to explore the use of computer vision,
digital signal processing, expert systems, and computer graphics to
automate a brain surgery procedure that alleviates the symptoms of
Parkinson's disease. Previously she oversaw a project which is
focused on developing a computerized system for assessing a patient's
dexterity as well as cognitive abilities with the goal of early
detection of disorders such as Alzheimer's disease and depression.
Maribeth is a project director and co-investigator in the
Wireless RERC, a 5-year
$5 mil grant from the National Institute for Disability and
Rehabilitation Research (NIDRR) in the area of wireless and
mobile devices for people with disabilities. Maribeth is a
co-investigator on RERC research such as user needs assessment
and on development projects for universal control and
cognitive prosthetics. Another of her duties with the RERC is
as the project director for the University Instruction portion
of the Center, which is an initiative to bring accessibility
and univers! al design concepts to researchers, students, and
industry. She has developed a clearinghouse of material
related to these topics and is a regular guest lecturer on
this topic in courses at Georgia Tech and other universities.
She has also presented tutorials on this topic at conferences
such as ISWC and HCII.
Making Sense
Chris Kasabach - BodyMedia
Abstract:
Participants will be provided a new wearable technology that
can continuously sense and present information about an
individual's health. After considering the characteristics of
this technology (the benefits it could provide wearers as well
as its physical characteristics) participants will develop an
"experience" around the technology in the form of a wearable
body monitor.
Inherent questions include:
- What is the clear benefit and value this product will
provide and to whom?
- How should the technology be configured and embodied
(where will it be worn?) to ensure physical comfort as well
as the correct perceptual message?
- What information will this product provide and in what form?
Bio:
Chris Kasabach directs BodyMedia's product design initiatives
and works closely with its partners to define new business
opportunities. Chris is a pioneer in the field of wearable
and mobile product development. In the early1990's he led the
development of the first user-centered wearable computers at
Carnegie Mellon and from 1993-7 co-directed the Mobile and
Wearable Computing group at the university's National Science
Foundation Design Research Center.
Throughout his career Chris has developed new products and
systems for groups including Boeing, Daimler Benz, Kodak,
Motorola and the United Nations Development Program. His work
has received numerous US patents, the Medical Design
Excellence Award, Gold Industrial Design Excellence Award, and
Fellisimo/New York Foundation for the Arts first prize. He
has presented widely in the United States, Europe and Africa
and his work has been featured in leading international
technology, business, design and art publications. Chris is a
1991 Thomas J. Watson Fellow.
Jonny Farringdon - BodyMedia
Jonny Farringdon directs Bodymedia's scientific initiatives,
inventing new sensors and technologies. He models millions of
physiological data points deriving context, health, and
biometric information. He led the wearable electronics
activity at Philips Research (1997-2001), and was product
manager (2000/1) for the Philips/Levi's ICD+ range of wearable
electronics jackets. Before this he was at University College
London as senior research fellow & tutor in computer science
(1995-97), and modelled human cognition (1990-95), memory,
learning, and the complexity of tasks. He studied in London,
Maths (BSc) and Artificial Intelligence (MSc distinction).
Jonny has been active in the commercial-research side of
wearables from an early point. A strong supporter of ISWC he
has presented many aspects of his work there spanning
augmented memory (2000), multi-modal displays (1999), and
introduced both accelerometers for context detection (1999,
2003) and sensor clothing (1999).
Wearable and Mobile Human Computer Interaction
Abstract:
Today more wearable computers (in the form of mobile phones and
MP3 players) are sold annually than desktop or laptop computers.
Given current technology trends, body-worn devices are quickly
approaching the performance of supercomputers in the early 1990's.
However, while the technology is improving rapidly and becoming
ubiquitous, mobile interfaces remain relatively undeveloped and
prevent the delivery of more sophisticated services to the mobile
user.
This tutorial will examine mobile devices from an Human-Computer
Interaction (HCI) perspective. While Mobile HCI is still
struggling to establish itself as a field, we will explore the
tools, techniques, and lessons from HCI that can be adapted for
the challenges of mobility. We will also discuss the current
state-of-the-art in Mobile HCI research with an emphasis on mobile
text entry, gestural interfaces, and voice input.
While prior experience in HCI would be helpful, it will not be expected.
Thad Starner - Georgia Institute of Technology
Thad Starner is an Assistant Professor in Georgia Tech's College
of Computing. He is a wearable computing pioneer, having worn a
wearable as an everyday personal assistant since 1993. Starner
holds four degrees from MIT, including his PhD from the MIT Media
Laboratory in 1999 on "Wearable Computing and Contextual
Awareness." Thad has authored over 80 scientific publications on
wearable computing, computer vision, and augmented reality and
received the best paper award at IEEE VR2000 and the prestigious
NSF Career Grant.
Kent Lyons - Georgia Institute of Technology
Kent Lyons is a PhD candidate in Georgia Tech's College of
Computing. His dissertation research is focused on enhancing
the support of conversations by improving mobile computing
input. Kent has authored several HCI papers on mobile and
wearable computing and presented user interface work at CHI,
UIST, ISWC and Mobile HCI. Kent is also an everyday wearable
computing user and has worn a wearable daily since 1999.
Ad Hoc Networking and Wireless Communications
David A. Maltz - Carnegie Mellon University
Abstract:
As the number and capabilities of wearable computing devices
have increased, so has the need to move information to and
from the devices --- wherever they may be. Advances in radio
technology and wireless communication protocols have created a
large tool-box of mechanisms that can be used to connect
wearable and mobile devices, but each of these mechanisms
comes with its own set of trade-offs. This tutorial is
targeted at helping attendees understand the capabilities and
limitations of today's wireless communication technologies,
and the direction those technologies are headed. It should be
of special interest to device and application designers who
want to know more about what they can and cannot expect from
wireless communications so that they can plan appropriately.
The tutorial will cover:
- Basics of radio wave propagation:
how and why walls, trees, and the environment affect
whether two nodes can communicate or not.
- Physical and link layer technologies: What is 802.1!
1a/b/g? How well do they really work? What is the
difference between WiFi and WiMax (802.16)? Bluetooth?
ZigBee?
- Building simple networks:
establishing communications when all nodes are in range of
each other
- Building multi-hop networks: How to build a
network using only the infrastructure your devices bring with
them. Overview of routing protocols in use in mobile ad hoc
networks (MANETs) including DSR, AODV, OLSR, and TBRPF.
Bio:
David A. Maltz developed the Dynamic Source Routing Protocol
for multi-hop wireless ad hoc networks --- DSR is on track to
become an RFC from the IETF Mobile Ad Hoc Network (MANET)
Working Group, and it has been used and extended by many groups.
He is an author of the mobile and wireless extensions to the
ns-2 simulator used by many research groups, and he lead the
development and evaluation of a full-scale wireless ad hoc
network testbed with car-mounted mobile nodes. He was the
network architect at FreeSpace Communications, a Silicon Valley
startup using a hybrid cellular/infrastructureless network to
provide residential Internet access at low cost and high
ease-of-use. He was the Founder and Director of Engineering at
AON Networks, which developed traffic management software for
carrier-class networks. David received his Ph.D. in May 2001
from Carnegie Mellon University while working in industry, and
returned there in 2003 as a Post-Doctoral Fellow.