Update README.md
This commit is contained in:
parent
428df5cbe8
commit
1d030e7ccf
1 changed files with 5 additions and 7 deletions
12
README.md
12
README.md
|
@ -4,8 +4,6 @@ Appearance-based gaze estimation methods that only require an off-the-shelf came
|
||||||
|
|
||||||
To democratize their use in HCI, we present OpenGaze, the first software toolkit that is specifically developed for gaze interface designers. OpenGaze is open source and aims to implement state-of-the-art methods for camera-based gaze estimation and interaction.
|
To democratize their use in HCI, we present OpenGaze, the first software toolkit that is specifically developed for gaze interface designers. OpenGaze is open source and aims to implement state-of-the-art methods for camera-based gaze estimation and interaction.
|
||||||
|
|
||||||
<a href="https://perceptual.mpi-inf.mpg.de/" target="_blank"><img src="https://github.molgen.mpg.de/perceptual/opengaze/blob/master/imgs/logo_pui.png" height="80"></a><img src="https://github.molgen.mpg.de/perceptual/opengaze/blob/master/imgs/logo_osaka-u.png" height="80">
|
|
||||||
|
|
||||||
## Functionality
|
## Functionality
|
||||||
|
|
||||||
The toolkit is capable of performing the following gaze-related tasks:
|
The toolkit is capable of performing the following gaze-related tasks:
|
||||||
|
@ -29,10 +27,10 @@ Perform personal calibration and remap the gaze target on a screen.
|
||||||
<p> </p>
|
<p> </p>
|
||||||
|
|
||||||
## Installation
|
## Installation
|
||||||
[Unix Installation](https://github.molgen.mpg.de/perceptual/opengaze/wiki/Unix-Installation)
|
[Unix Installation](https://gitlab.hcics.simtech.uni-stuttgart.de/public-projects/opengaze/wikis/Unix-installation)
|
||||||
|
|
||||||
## Use
|
## Use
|
||||||
[Command line arguments](https://github.molgen.mpg.de/perceptual/opengaze/wiki/Command-line-arguments)
|
[Command line arguments](https://gitlab.hcics.simtech.uni-stuttgart.de/public-projects/opengaze/wikis/Command-line-arguments)
|
||||||
|
|
||||||
## Citation
|
## Citation
|
||||||
**If you use any of the resources provided on this page in any of your publications, please cite the following paper:**
|
**If you use any of the resources provided on this page in any of your publications, please cite the following paper:**
|
||||||
|
@ -42,7 +40,7 @@ Evaluation of Appearance-Based Methods and Implications for Gaze-Based Applicati
|
||||||
Xucong Zhang, Yusuke Sugano, Andreas Bulling
|
Xucong Zhang, Yusuke Sugano, Andreas Bulling
|
||||||
Proc. ACM SIGCHI Conference on Human Factors in Computing Systems (CHI), 2019
|
Proc. ACM SIGCHI Conference on Human Factors in Computing Systems (CHI), 2019
|
||||||
```
|
```
|
||||||
[arXiv](https://arxiv.org/abs/1901.10906)
|
[Project page](https://www.perceptualui.org/publications/zhang19_chi/)
|
||||||
|
|
||||||
@inproceedings{zhang19_chi,<br/>
|
@inproceedings{zhang19_chi,<br/>
|
||||||
title = {Evaluation of Appearance-Based Methods for Gaze-Based Applications},<br/>
|
title = {Evaluation of Appearance-Based Methods for Gaze-Based Applications},<br/>
|
||||||
|
@ -55,11 +53,11 @@ abstract = {Appearance-based gaze estimation methods that only require an off-th
|
||||||
|
|
||||||
## License
|
## License
|
||||||
|
|
||||||
The license agreement can be found in Copyright.txt
|
The license agreement can be found in [LICENSE](https://gitlab.hcics.simtech.uni-stuttgart.de/public-projects/opengaze/LICENSE).
|
||||||
|
|
||||||
You have to respect boost, OpenFace and OpenCV licenses.
|
You have to respect boost, OpenFace and OpenCV licenses.
|
||||||
|
|
||||||
Furthermore, you have to respect the licenses of the datasets used for [model training](:https://github.molgen.mpg.de/perceptual/opengaze/wiki/Model-training).
|
Furthermore, you have to respect the licenses of the datasets used for [model training](https://gitlab.hcics.simtech.uni-stuttgart.de/public-projects/opengaze/wikis/Model-training).
|
||||||
|
|
||||||
## Code layout
|
## Code layout
|
||||||
* caffe-layers: Our customized layers for the Caffe library.
|
* caffe-layers: Our customized layers for the Caffe library.
|
||||||
|
|
Loading…
Reference in a new issue