OAE's Proteus Project: A New Learning Paradigm
From: Speaking of Computers - Stanford University - 09/22/2003 - page 5

It was born three years ago as an intriguing concept to better accommodate
students with disabilities. lt has since grown into a project that could
enhance the way all students learn, not only at Stanford but around the
world. The Proteus Learning Environment is the brainchild of the Office of
Accessible Education (OAE) and is the foundation of the newly-created Center
for Universal Design in Education.

...

Because it is important to integrate the teaching styles of instructors and
the learning needs of students, the Office of Accessible Education will be
including faculty and students as partners on the project development team.
If you are a faculty member with an "early adopter" frame of mind and are
interested in working with us, please contact Dr. Joan Bisagno, OAE Director
(joan.bisagno@stanford.edu, 650/723-1039) or Shelley Haven, Academic
Technology Specialist and Technology Coordinator for the OAE
(shelley.haven@stanford.edu, 650/725-6173). 


Scenario        Example of How Student Might Use Proteus Learning Environment

Student with    - Listens to lecture.
learning        - Synchronizes notes (taken by peer notetaker) with lecture
disabilities,     transcript.
ADHD            - Bookmarks stream for later referral (e.g.,"important
                  point", "Don't understand", 'Assignment").  


Student with    - Accesses lecture (transcribed via remote captioner) later
hearing           via speech recognition. 
impairment      - Uses closed captioned video.
                - Reviews Lecture transcript and notes available after class. 


Student who is  - Listens to lecture, bookmarks points.
legally blind   - Takes notes via Braille notetaker.
                - Uses refreshable (i.e., dynamic) Braille display to read
                  text on Web and handouts previously prepared by instructor. 


Any student,    - Reviews bookmarked sections of stream for follow-up.
studying after  - Reads AND listens to lecture transcript (text and audio
class             synchronized). 
                - Searches for material via text search,then replays
                  synchronized video, audio, whiteboard capture, etc. 
                - Annotates transcript,class notes using speech recognition,
                  typing, or handwriting recognition.


Read the entire article at:
http://academiccomputing.stanford.edu/pubs/SOC/Back_Issues/SOC63/highlights.html#high9
