Download full issue (PDF 2.1MB) - Council of European ...
Download full issue (PDF 2.1MB) - Council of European ...
Download full issue (PDF 2.1MB) - Council of European ...
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Information Technologies for Visually Impaired People<br />
3 Sketch Understanding<br />
In this section we describe the tool for understanding<br />
sketched drawings, its applications and the pattern recognition<br />
techniques used in it.<br />
3.1Purpose and Application Framework<br />
Usually, visually impaired children use a plastic sheet<br />
to make drawings so that the drawings can be seen but also<br />
perceived by touching the relief. Also, computers can "see"<br />
the drawings input by touching devices like a digital tablet<br />
or a tablet PC. Our idea was to join both drawing modes by<br />
adding a plastic sheet on the screen <strong>of</strong> a digital tablet. Thus,<br />
while the strokes made by the user are digitally acquired by<br />
the electronic device as a sequence <strong>of</strong> 2D coordinates, they<br />
are also marked in relief in the sheet so can be touched by<br />
the user.<br />
Our system allows teachers to define new forms as exercises<br />
to practice with the sketching input possibilities.<br />
These forms can be designed to interactively learn geometric<br />
shapes, graphical symbols and diagrams, letters, or any<br />
other symbol consisting <strong>of</strong> strokes. In addition, our system<br />
can be used to include diagrams in documents after cleaning<br />
up. Forms have different zones with associated functions.<br />
For example a zone can have a drawing and the associated<br />
function <strong>of</strong> aligning the input strokes with the lines<br />
<strong>of</strong> it. The aim would be to measure the ability <strong>of</strong> the user to<br />
follow (redraw) the lines <strong>of</strong> a given drawing. In addition<br />
this activity zone has two audio messages depending whether<br />
the action is right or not.<br />
Once the form has been designed, it can be printed with<br />
a special printer that makes the reliefs <strong>of</strong> the Braille dots<br />
and associates textures to colours. The form is printed on a<br />
plastic sheet that is placed on the digitizing tablet during<br />
the activity. Thus the child can "read" the instructions by<br />
touching the relief and interact with the system by touching<br />
the cells and getting audio feedback, or drawing with the<br />
digital pen on the tablet by following the drawing, or copying<br />
it into another cell, etc.<br />
The double system, relief plastic sheet and digital tablet,<br />
allows the visually impaired user first to touch the figure,<br />
and then to draw it with the digital pen. Digital strokes<br />
are captured from the computer and interpreted and the computer<br />
gives audio feedback accordingly. Thus, for example<br />
the system compares the input figure with a database <strong>of</strong><br />
models like square, circles, animal silhouettes, capital letters,<br />
etc.<br />
The teacher, when designing the forms, can record audio<br />
responses or visual ones like emoticons associated to<br />
activity actions. Visual outputs allow the system to be used<br />
by non-visually impaired or low vision users. An outline <strong>of</strong><br />
the system is shown in Figure 7. An interesting survey on<br />
similar interfaces can be found in [10].<br />
3. 2 A Framework Based on a Sketching Interface<br />
The technological basis <strong>of</strong> our system is the recognition<br />
<strong>of</strong> graphical symbols. It can be defined as the identification<br />
and interpretation <strong>of</strong> visual shapes depending on the do-<br />
main. It means that the same symbol can have different<br />
meanings depending on the context in which it appears.<br />
According to the input mode, we can classify the recognition<br />
<strong>of</strong> graphical symbols in two categories, namely <strong>of</strong>fline<br />
or static recognition, and on-line or dynamic recognition.<br />
The former uses a scanned document and only works<br />
with the spatial information <strong>of</strong> the image strokes. The latter<br />
interactively obtains the document, using an input device<br />
as a digitizing tablet or Tablet PC, so it additionally has<br />
dynamic information <strong>of</strong> the temporal sequence <strong>of</strong> strokes,<br />
speed, pressure, etc. Our work is based on on-line recognition<br />
<strong>of</strong> hand drawn graphical symbols by a digitizing tablet.<br />
Such type <strong>of</strong> man-machine interfaces are called sketching<br />
(or calligraphic) interfaces.<br />
Related works can be read in [11] [12]. On-line recognition<br />
consists <strong>of</strong> three steps: acquisition, primitive recognition<br />
and interpretation. The acquisition step consists <strong>of</strong><br />
extracting the strokes from the graphic device and encoding<br />
them as a sequence <strong>of</strong> coordinates or lines approximating<br />
the input as well as dynamic information such as speed,<br />
pressure, etc. Primitive recognition consists <strong>of</strong> associating<br />
the strokes with basic graphic structures as polylines, circumference<br />
arcs, basic geometric shapes, etc. Finally the<br />
interpretation step aims to associate a semantic meaning to<br />
the graphical structure consisting <strong>of</strong> basic primitives. This<br />
step requires the use <strong>of</strong> semantic knowledge <strong>of</strong> the working<br />
domain. Let us further describe these steps in our system.<br />
a) Acquisition <strong>of</strong> drawing strokes: when the user<br />
draws on the digitizing tablet, the electronics <strong>of</strong> it convert<br />
the input to a sequence <strong>of</strong> coordinates that are sent to the<br />
computer. Thus, each stroke <strong>of</strong> the drawing is encoded as a<br />
sequence <strong>of</strong> coordinates, see Figure 8. Taking as curves such<br />
sequences <strong>of</strong> coordinates, singular points as maximum curvature<br />
points, crossings, end points, etc. are detected. It allows<br />
conversion <strong>of</strong> the strokes in sequences <strong>of</strong> segments as<br />
Figure 8 illustrates.<br />
b) Primitive recognition: The extracted segments are<br />
grouped into primitives. These primitives depend on the<br />
application domain. For example, in flowchart diagrams<br />
primitives can be squares, circles, arrows, etc. In our case,<br />
the alphabet <strong>of</strong> primitives consists <strong>of</strong> the valid shapes as<br />
squares, diamonds, circumferences, etc. In the example<br />
shown in Figure 8 the detected primitive is an open polyline<br />
identified as a square<br />
c) Interpretation: The last step is the validation <strong>of</strong> the<br />
drawing after comparing it with possible models. For example,<br />
in the case <strong>of</strong> forms requesting a particular drawing,<br />
the input is compared with the ideal pattern with a distortion<br />
tolerance allowed for due to the hand drawn input, and<br />
an acceptance decision is made. In Figure 8 we show a shape<br />
that was drawn and recognized as square, and the corresponding<br />
model <strong>of</strong> Figure 7a). On the contrary in Figure<br />
7b), we can see another square but with a high degree <strong>of</strong><br />
distortion regarding the model and that has been rejected.<br />
In freehand drawings, each input shape is compared with<br />
all the models in the database and is recognized as the model<br />
with minimum distance, under a given threshold. The mod-<br />
60 UPGRADE Vol. VIII, No. 2, April 2007 © Novática