Kaynağa Gözat

Wrote 'Experiments' and started writing 'Design' chapters in report.

Taddeus Kroes 13 yıl önce
ebeveyn
işleme
7c17c78302
5 değiştirilmiş dosya ile 320 ekleme ve 113 silme
  1. 12 24
      docs/data/diagram.tex
  2. BIN
      docs/data/experimental_draw.png
  3. 33 0
      docs/report.bib
  4. 273 83
      docs/report.tex
  5. 2 6
      docs/rules.mk

+ 12 - 24
docs/data/server_scheme.tex → docs/data/diagram.tex

@@ -1,11 +1,3 @@
-\documentclass{minimal}
-
-\usepackage{tikz}
-\usetikzlibrary{shapes,arrows}
-
-\begin{document}
-\pagestyle{empty}
-
 % Block+line styles
 \tikzstyle{block} = [rectangle, draw, fill=blue!20, text width=5em,
     text centered, rounded corners, minimum height=4em]
@@ -17,18 +9,18 @@
 
 \begin{tikzpicture}[node distance = 7em, auto]
     % Servers
-    \node [block] (inputserver) {Input server};
-    \node [right of=inputserver] (inputserverdots) {$\cdots$};
+    \node [block] (eventserver) {Event server};
+    \node [right of=eventserver] (eventserverdots) {$\cdots$};
 
-    \node [impl, right of=inputserverdots] (tuioserver) {TUIO server};
+    \node [impl, right of=eventserverdots] (tuioserver) {TUIO server};
 
-    \node [block, below of=inputserver] (gestureserver) {Gesture server};
-    \path [line] (inputserver) -- node {triggers [down, move, up]}(gestureserver);
+    \node [block, below of=eventserver] (gestureserver) {Gesture erver};
+    \path [line] (eventserver) -- node {trigger events of all touch points}(gestureserver);
 
     % Window
     \node [block, below of=gestureserver] (window) {Window};
     \node [right of=window] (windowdots) {$\cdots$};
-    \path [line] (gestureserver) -- node {triggers [down, move, up]}(window);
+    \path [line] (gestureserver) -- node {trigger events of containing touch points}(window);
 
     \node [impl, right of=windowdots] (rectangularwindow) {Rectangular window};
     \node [impl, right of=rectangularwindow] (circularwindow) {Circular window};
@@ -36,7 +28,7 @@
     % Tracker
     \node [block, below of=window] (tracker) {Gesture tracker};
     \node [right of=tracker] (trackerdots) {$\cdots$};
-    \path [line] (window) -- node {triggers [down, move, up]}(tracker);
+    \path [line] (window) -- node {delegate events}(tracker);
 
     \node [impl, right of=trackerdots] (taptracker) {Tap tracker};
     \node [impl, right of=taptracker, text width=7em] (transformtracker) {Transformation tracker};
@@ -44,16 +36,12 @@
     % Gesture
     \node [block, below of=tracker] (gesture) {Gesture};
     \node [right of=gesture] (gesturedots) {$\cdots$};
-    \path [line] (tracker) -- node {triggers gesture}(gesture);
+    \path [line] (tracker) -- node {trigger by calling gesture handler}(gesture);
 
     % Client application
-    \node [app, left of=window, xshift=-14em] (app) {Client application};
-    \path [line, dashed] (app) -- node [near start] {binds gesture handler}(tracker);
-    \path [line, dashed] (tracker) -- node [near start] {triggers gesture handler}(app);
-
-    \path [line, dashed] (app) -- node [near start] {starts}(gestureserver);
-    \path [line, dashed] (app) -- node [near start] {adds to server}(window);
+    \node [app, left of=window, xshift=-11em] (app) {Client application};
+    \path [line, dashed] (app) -- node [left=20, near end] {add to window, bind gesture handler}(tracker);
 
+    \path [line, dashed] (app) -- node [near start] {start}(gestureserver);
+    \path [line, dashed] (app) -- node {add to gesture server}(window);
 \end{tikzpicture}
-
-\end{document}

BIN
docs/data/experimental_draw.png


+ 33 - 0
docs/report.bib

@@ -51,3 +51,36 @@
    note = {10.1007/978-3-540-73110-8\_78},
    year = {2007}
 }
+
+@article{OSC,
+    author = {Wright, Matthew},
+    title = {Open Sound Control: an enabling technology for musical networking},
+    journal = {Org. Sound},
+    issue_date = {December 2005},
+    volume = {10},
+    number = {3},
+    month = {dec},
+    year = {2005},
+    issn = {1355-7718},
+    pages = {193--200},
+    numpages = {8},
+    url = {http://dx.doi.org/10.1017/S1355771805000932},
+    doi = {10.1017/S1355771805000932},
+    acmid = {1109064},
+    publisher = {Cambridge University Press},
+    address = {New York, NY, USA},
+}
+
+@misc{pyOSC,
+    author = {v2lab},
+    title = {pyOSC - A Simple OpenSoundControl implementation, in Pure Python.},
+    year = {2008},
+    howpublished = {\url{https://trac.v2.nl/wiki/pyOSC}}
+}
+
+@misc{pyTUIO,
+    author = {Jannis Leidel},
+    title = {pyTUIO - A Python library that understands the TUIO protocol.},
+    year = {2008},
+    howpublished = {\url{http://code.google.com/p/pytuio/}}
+}

+ 273 - 83
docs/report.tex

@@ -2,7 +2,8 @@
 
 \usepackage[english]{babel}
 \usepackage[utf8]{inputenc}
-\usepackage{hyperref,graphicx,float}
+\usepackage{hyperref,graphicx,float,tikz}
+\usetikzlibrary{shapes,arrows}
 
 % Link colors
 \hypersetup{colorlinks=true,linkcolor=black,urlcolor=blue,citecolor=DarkGreen}
@@ -41,8 +42,8 @@ these frameworks have no access to their multi-touch events.
 
 % Aanleiding
 This problem was observed during an attempt to create a multi-touch
-``interactor'' class for the Visualization Toolkit (VTK \cite{VTK}). Because
-VTK provides the application framework here, it is undesirable to use an entire
+``interactor'' class for the Visualization Toolkit \cite[VTK]{VTK}. Because VTK
+provides the application framework here, it is undesirable to use an entire
 framework like Qt simultaneously only for its multi-touch support.
 
 % Ruw doel
@@ -65,7 +66,6 @@ events.
     % Deelvragen
     To design such a mechanism properly, the following questions are relevant:
     \begin{itemize}
-        \item What are the requirements of the mechanism to be universal?
         \item What is the input of the mechanism? Different touch drivers have
             different API's. To be able to support different drivers (which is
             highly desirable), there should probably be a translation from the
@@ -73,25 +73,33 @@ events.
         \item How can extendability be accomplished? The set of supported events
             should not be limited to a single implementation, but an application
             should be able to define its own custom events.
+        \item How can the mechanism be used by different programming languages?
+            A universal mechanism should not be limited to be used in only one
+            language.
         \item Can events be shared with multiple processes at the same time? For
             example, a network implementation could run as a service instead of
             within a single application, triggering events in any application that
             needs them.
-        \item Is performance an issue? For example, an event loop with rotation
-            detection could swallow up more processing resources than desired.
+        % FIXME: gaan we nog wat doen met onderstaand?
+        %\item Is performance an issue? For example, an event loop with rotation
+        %    detection could swallow up more processing resources than desired.
+        \item How can the mechanism be integrated in a VTK interactor?
     \end{itemize}
 
     % Afbakening
-    The scope of this thesis includes the design of an multi-touch triggering
-    mechanism, a reference implementation of this design, and its integration
-    into a VTK interactor. To be successful, the design should allow for
-    extensions to be added to any implementation. The reference implementation
-    is a Proof of Concept that translates TUIO events to some simple touch
-    gestures that are used by a VTK interactor.
+    The scope of this thesis includes the design of a universal multi-touch
+    triggering mechanism, a reference implementation of this design, and its
+    integration into a VTK interactor. To be successful, the design should
+    allow for extensions to be added to any implementation.
+
+    The reference implementation is a Proof of Concept that translates TUIO
+    events to some simple touch gestures that are used by a VTK interactor.
+    Being a Proof of Concept, the reference implementation itself does not
+    necessarily need to meet all the requirements of the design.
 
     \section{Structure of this document}
 
-    % TODO
+    % TODO: pas als het klaar is
 
 \chapter{Related work}
 
@@ -134,7 +142,25 @@ events.
     detected by different gesture trackers, thus separating gesture detection
     code into maintainable parts.
 
+    \section{Processing implementation of simple gestures in Android}
+
+    An implementation of a detection mechanism for some simple multi-touch
+    gestures (tap, double tap, rotation, pinch and drag) using
+    Processing\footnote{Processing is a Java-based development environment with
+    an export possibility for Android. See also \url{http://processing.org/.}}
+    can be found found in a forum on the Processing website
+    \cite{processingMT}. The implementation is fairly simple, but it yields
+    some very appealing results. The detection logic of all gestures is
+    combined in a single class. This does not allow for extendability, because
+    the complexity of this class would increase to an undesirable level (as
+    predicted by the GART article \cite{GART}). However, the detection logic
+    itself is partially re-used in the reference implementation of the
+    universal gesture detection mechanism.
+
+\chapter{Preliminary}
+
     \section{The TUIO protocol}
+    \label{sec:tuio}
 
     The TUIO protocol \cite{TUIO} defines a way to geometrically describe
     tangible objects, such as fingers or fiducials on a multi-touch table.  The
@@ -142,15 +168,13 @@ events.
     information is sent to the TUIO UDP port (3333 by default).
 
     For efficiency reasons, the TUIO protocol is encoded using the Open Sound
-    Control (OSC)\footnote{\url{http://opensoundcontrol.org/specification}}
-    format.  An OSC server/client implementation is available for Python:
-    pyOSC\footnote{\url{https://trac.v2.nl/wiki/pyOSC}}.
+    Control \cite[OSC]{OSC} format. An OSC server/client implementation is
+    available for Python: pyOSC \cite{pyOSC}.
 
-    A Python implementation of the TUIO protocol also exists:
-    pyTUIO\footnote{\url{http://code.google.com/p/pytuio/}}. However, the
-    execution of an example script yields an error regarding Python's built-in
-    \texttt{socket} library. Therefore, the reference implementation uses the
-    pyOSC package to receive TUIO messages.
+    A Python implementation of the TUIO protocol also exists: pyTUIO
+    \cite{pyTUIO}. However, the execution of an example script yields an error
+    regarding Python's built-in \texttt{socket} library. Therefore, the
+    reference implementation uses the pyOSC package to receive TUIO messages.
 
     The two most important message types of the protocol are ALIVE and SET
     messages. An ALIVE message contains the list of session id's that are
@@ -188,26 +212,10 @@ events.
         scale these values back to the actual screen dimension.
     \end{quote}
 
-    In other words, the design of the gesture detection mechanism should
-    incorporate a translation from driver-specific coordinates to pixel
-    coordinates.
-
-    \section{Processing implementation of simple gestures in Android}
+    \section{The Visualization Toolkit}
+    \label{sec:vtk}
 
-    An implementation of a detection mechanism for some simple multi-touch
-    gestures (tap, double tap, rotation, pinch and drag) using
-    Processing\footnote{Processing is a Java-based development environment with
-    an export possibility for Android. See also \url{http://processing.org/.}}
-    can be found found in a forum on the Processing website
-    \cite{processingMT}. The implementation is fairly simple, but it yields
-    some very appealing results. The detection logic of all gestures is
-    combined in a single class. This does not allow for extendability, because
-    the complexity of this class would increase to an undesirable level (as
-    predicted by the GART article \cite{GART}). However, the detection logic
-    itself is partially re-used in the reference implementation of the
-    universal gesture detection mechanism.
-
-% TODO
+    % TODO
 
 \chapter{Experiments}
 
@@ -227,6 +235,71 @@ events.
 
 % Proof of Concept: VTK interactor
 
+    \section{Experimenting with TUIO and event bindings}
+    \label{sec:experimental-draw}
+
+    When designing a software library, its API should be understandable and
+    easy to use for programmers. To find out the basic requirements of the API
+    to be usable, an experimental program has been written based on the
+    Processing code from \cite{processingMT}. The program receives TUIO events
+    and translates them to point \emph{down}, \emph{move} and \emph{up} events.
+    These events are then interpreted to be (double or single) \emph{tap},
+    \emph{rotation} or \emph{pinch} gestures. A simple drawing program then
+    draws the current state to the screen using the PyGame library. The output
+    of the program can be seen in figure \ref{fig:draw}.
+
+    \begin{figure}[H]
+        \center
+        \label{fig:draw}
+        \includegraphics[scale=0.4]{data/experimental_draw.png}
+        \caption{Output of the experimental drawing program. It draws the touch
+                 points and their centroid on the screen (the centroid is used
+                 as center point for rotation and pinch detection). It also
+                 draws a green rectangle which responds to rotation and pinch
+                 events.}
+    \end{figure}
+
+    One of the first observations is the fact that TUIO's \texttt{SET} messages
+    use the TUIO coordinate system, as described in section \ref{sec:tuio}.
+    The test program multiplies these with its own dimensions, thus showing the
+    entire screen in its window. Also, the implementation only works using the
+    TUIO protocol. Other drivers are not supported.
+
+    Though using relatively simple math, the rotation and pinch events work
+    surprisingly well. Both rotation and pinch use the centroid of all touch
+    points. A \emph{rotation} gesture uses the difference in angle relative to
+    the centroid of all touch points, and \emph{pinch} uses the difference in
+    distance. Both values are normalized using division by the number of touch
+    points. A pinch event contains a scale factor, and therefore uses a
+    division of the current by the previous average distance to the centroid.
+
+    There is a flaw in this implementation. Since the centroid is calculated
+    using all current touch points, there cannot be two or more rotation or
+    pinch gestures simultaneously. On a large multi-touch table, it is
+    desirable to support interaction with multiple hands, or multiple persons,
+    at the same time.
+
+    Also, the different detection algorithms are all implemented in the same
+    file, making it complex to read or debug, and difficult to extend.
+
+    \section{VTK interactor}
+
+    % TODO
+    % VTK heeft eigen pipeline, mechanisme moet daarnaast draaien
+
+    \section{Summary of observations}
+    \label{sec:observations}
+
+    \begin{itemize}
+        \item The TUIO protocol uses a distinctive coordinate system and set of
+            messages.
+        \item Touch events occur outside of the application window.
+        \item Gestures that use multiple touch points are using all touch
+            points (not a subset of them).
+        \item Code complexity increases when detection algorithms are added.
+        \item % TODO: VTK interactor observations
+    \end{itemize}
+
 % -------
 % Results
 % -------
@@ -234,49 +307,176 @@ events.
 \chapter{Design}
 
     \section{Requirements}
+    \label{sec:requirements}
 
-    % TODO
-    % ondersteunen van meerdere drivers
-    % gesture detectie koppelen aan bepaald gedeelte van het scherm
-    % scheiden van detectiecode voor verschillende gesture types
-    % eventueel te gebruiken in meerdere talen
-
-    \section{Input server}
+    From the observations in section \ref{sec:observations}, a number of
+    requirements can be specified for the design of the event mechanism:
 
-    % TODO
-    % vertaling driver naar point down, move, up
-    % TUIO in reference implementation
-
-    \section{Gesture server}
-
-    % TODO
-    % vertaling naar pixelcoordinaten
-    % toewijzing aan windows
+    \begin{itemize}
+        % vertalen driver-specifieke events naar algemeen formaat
+        \item To be able to support multiple input drivers, there must be a
+            translation from driver-specific messages to some common format
+            that can be used in gesture detection algorithms.
+        % events toewijzen aan GUI window (windows)
+        \item An application GUI window should be able to receive only events
+            occuring within that window, and not outside of it.
+        % scheiden groepen touchpoints voor verschillende gestures (windows)
+        \item To support multiple objects that are performing different
+            gestures at the same time, the mechanism must be able to perform
+            gesture detection on a subset of the active touch points.
+        % scheiden van detectiecode voor verschillende gesture types
+        \item To avoid an increase in code complexity when adding new detection
+            algorithms, detection code of different gesture types must be
+            separated.
+    \end{itemize}
 
-    \section{Windows}
+    \section{Components}
+
+        Based on the requirements from section \ref{sec:requirements}, a design
+        for the mechanism has been created. The design consists of a number of
+        components, each having a specific set of tasks.
+
+        \subsection{Event server}
+
+        % vertaling driver naar point down, move, up
+        % vertaling naar schermpixelcoordinaten
+        % TUIO in reference implementation
+
+        The \emph{event server} is an abstraction for driver-specific server
+        implementations, such as a TUIO server. It receives driver-specific
+        messages and tanslates these to a common set of events and a common
+        coordinate system.
+
+        A minimal example of a common set of events is $\{point\_down,
+        point\_move, point\_up\}$. This is the set used by the reference
+        implementation. Respectively, these events represent an object being
+        placed on the screen, moving along the surface of the screen, and being
+        released from the screen.
+
+        A more extended set could also contain the same three events for a
+        surface touching the screen. However, a surface can have a rotational
+        property, like the ``fiducials'' type in the TUIO protocol. This
+        results in as $\{point\_down, point\_move, point\_up, surface\_down,
+        surface\_move, surface\_up,\\surface\_rotate\}$.
+
+        An important note here, is that similar events triggered by different
+        event servers must have the same event type and parameters. In other
+        words, the output of the event servers should be determined by the
+        gesture servers (not the contrary).
+
+        The output of an event server implementation should also use a common
+        coordinate system, that is the coordinate system used by the gesture
+        server. For example, the reference implementation uses screen
+        coordinates in pixels, where (0, 0) is the upper left corner of the
+        screen.
+
+        The abstract class definition of the event server should provide some
+        functionality to detect which driver-specific event server
+        implementation should be used.
+
+        \subsection{Gesture trackers}
+
+        A \emph{gesture tracker} detects a single gesture type, given a set of
+        touch points. If one group of points on the screen is assigned to one
+        tracker and another group to another tracker, multiple gestures, an be
+        detected at the same time. For this assignment, the mechanism uses
+        windows.  These will be described in the next section.
+
+        % event binding/triggering
+        A gesture tracker triggers a gesture event by executing a callback.
+        Callbacks are ``bound'' to a tracker by the application. Because
+        multiple gesture types can have very similar detection algorithm, a
+        tracker can detect multiple different types of gestures. For instance,
+        the rotation and pinch gestures from the experimental program in
+        section \ref{sec:experimental-draw} both use the centroid of all touch
+        points.
+
+        If no callback is bound for a particular gesture type, no detection of
+        that type is needed. A tracker implementation can use this knowledge
+        for code optimization.
+
+        % scheiding algoritmiek
+        A tracker implementation defines the gesture types it can trigger, and
+        the detection algorithms to trigger them. Consequently, detection
+        algorithms can be separated in different trackers. Different
+        trackers can be saved in different files, reducing the complexity of
+        the code in a single file. \\
+        % extendability
+        Because tacker defines its own set of gesture types, the application
+        developer can define application-specific trackers (by extending a base
+        \texttt{GestureTracker} class, for example). In fact, any built-in
+        gesture trackers of an implementation are also created this way. This
+        allows for a plugin-like way of programming, which is very desirable if
+        someone would want to build a library of gesture trackers. Such a
+        library can easy be extended by others.
+
+        \subsection{Windows}
+
+        A \emph{window} represents a subset of the entire screen surface. The
+        goal of a window is to restrict the detection of certain gestures to
+        certain areas. A window contains a list of touch points, and a list of
+        trackers. A window server (defined in the next section) assigns touch
+        points to a window, but the window itself defines functionality to
+        check whether a touch point is inside the window. This way, new windows
+        can be defined to fit over any 2D object used by the application.
+
+        The first and most obvious use of a window is to restrict touch events
+        to a single application window. However, the use of windows can be used
+        in a lot more powerful way.
+
+        For example, an application contains an image with a transparent
+        background that can be dragged around. The user can only drag the image
+        by touching its foreground. To accomplish this, the application
+        programmer can define a window type that uses a bitmap to determine
+        whether a touch point is on the visible image surface. The tracker
+        which detects drag gestures is then bound to this window, limiting the
+        occurence of drag events to the image surface.
+
+        % toewijzen even aan deel v/h scherm:
+        % TUIO coördinaten zijn over het hele scherm en van 0.0 tot 1.0, dus
+        % moeten worden vertaald naar pixelcoördinaten binnen een ``window''
+        % TODO
+
+        \subsection{Gesture server}
+
+        % luistert naar point down, move, up
+        The \emph{gesture server} delegates events from the event server to the
+        set of windows that contain the touch points related to the events.
+
+        % toewijzing point (down) aan window(s)
+        The gesture server contains a list of windows. When the event server
+        triggers an event, the gesture server ``asks'' each window whether it
+        contains the related touch point. If so, the window updates its gesture
+        trackers, which can then trigger gestures.
+
+    \section{Diagram of component relations}
+
+    \begin{figure}[H]
+        \input{data/diagram}
+        % TODO: caption
+    \end{figure}
+
+    \section{Example usage}
 
     % TODO
-    % toewijzen even aan deel v/h scherm:
-    % TUIO coördinaten zijn over het hele scherm en van 0.0 tot 1.0, dus moeten
-    % worden vertaald naar pixelcoördinaten binnen een ``window''
+    % vertellen hoe je tracker aanmaakt, binnen een window
 
-    \section{Trackers}
+    %\section{Network protocol}
 
     % TODO
-    % event binding/triggering
-    % extendability
-
-% TODO: link naar appendix met schema
+    % ZeroMQ gebruiken voor communicatie tussen meerdere processen (in
+    % verschillende talen)
 
 \chapter{Reference implementation}
 
 % TODO
+% alleen window.contains op point down, niet move/up
 
 \chapter{Integration in VTK}
 
 % VTK interactor
 
-\chapter{Conclusions}
+%\chapter{Conclusions}
 
 % TODO
 % Windows zijn een manier om globale events toe te wijzen aan vensters
@@ -286,27 +486,17 @@ events.
 \chapter{Suggestions for future work}
 
 % TODO
-% Network protocol (ZeroMQ)
-% State machine
 
-\bibliographystyle{plain}
-\bibliography{report}{}
+% Network protocol (ZeroMQ) voor meerdere talen en simultane processen
+% Hierij ook: extra laag die gesture windows aanmaakt die corresponderen met window manager
 
-\appendix
-
-\chapter{Diagram of mechanism structure}
-\label{app:schema}
+% State machine
 
-\begin{figure}[H]
-    \hspace{-14em}
-    \includegraphics{data/server_scheme.pdf}
-    \caption{}
-    %TODO: caption
-\end{figure}
+% Window in boomstructuur voor efficientie
 
-\chapter{Supported events in reference implementation}
-\label{app:supported-events}
+\bibliographystyle{plain}
+\bibliography{report}{}
 
-% TODO
+%\appendix
 
 \end{document}

+ 2 - 6
docs/rules.mk

@@ -1,5 +1,5 @@
 REPORT := report
-SCHEME := data/server_scheme
+DIAGRAM := data/diagram
 PDFLATEX_FLAGS = -halt-on-error -interaction=nonstopmode \
 	-output-directory $(@D) -shell-escape
 LATEX = TEXINPUTS=$(d):$(b): pdflatex $(PDFLATEX_FLAGS)
@@ -19,11 +19,7 @@ $(b)%.pdf: $(d)%.tex
 		$(LATEX) $^; \
 	done
 
-$(b)$(REPORT).pdf: $(b)$(SCHEME).pdf
-
-$(b)$(SCHEME).pdf: $(d)$(SCHEME).tex
-	mkdir -p $(@D)
-	pdflatex $(PDFLATEX_FLAGS) $^
+$(b)$(REPORT).pdf: $(d)$(DIAGRAM).tex
 
 $(b)$(REPORT).bbl: $(d)$(REPORT).bib
 	BIBINPUTS=$(d) bibtex8 ${@:.bbl=.aux}