< Dongguk Open Databases >
---------------------------------------------
1. Dongguk
Drone Camera Database (DDroneC-DB1)
2. Dongguk
Visible Light & FIR Pedestrian Detection Database (DVLFPD-DB1) & CNN
Model
3. Dongguk
Open and Closed Eyes Database (DOCE-DB1) & CNN Model
4. Dongguk
Multi-national Currencies Database (DMC-DB1) & CNN Model
5. Dongguk
Finger-Vein Database (DFingerVein-DB1) & CNN Model
6. ISPR
database (real and presentation attack finger-vein images)
7. Dongguk
Night-time Human Detection Database (DNHD-DB1) & CNN Model
8. Dongguk
Body-based Person Recognition Database (DBPerson-Recog-DB1)
9. Dongguk
Body-based Gender Database (DBGender-DB2)
10.
Dongguk Activities & Actions Database
(DA&A-DB1)
11.
Dongguk Body-based Gender Database
(DBGender-DB1)
12.
Dongguk Face Database (DFace-DB1)
13.
Dongguk Banknote Database (DBanknote-DB1)
---------------------------------------------
1. Dongguk Drone Camera Database (DDroneC-DB1)
(1) Introduction
In our experiments, we used a DJI Phantom 4 quadcopter to capture
the video while the drone was landing or hovering. It includes a color camera
with a 1/2.3-inch-thick complementary metal–oxide–semiconductor (CMOS) sensor,
with a 94O-field-of-view (FOV) and an f/2.8 lens. The captured videos are in
mpeg-4 (MP4) format with 30 fps, and have a size of 1280 x 720 pixels. The
drone’s gimbal is adjusted 90° downward so that during landing, the camera can
be facing the ground. Our database (shown in Table 1) is divided in two sub
databases: drone landing on the marker and drone hovering over the same
position while the marker is moving on the ground. For each sub database, we
captured four videos at 10 AM, 2 PM, 6 PM, and 10 PM. We acquired videos in
varying types of environments (humidity level, wind velocity, temperature, and
weather). The marker was visible in the sequences for the morning, afternoon,
and evening, but it was barely seen in the night video.
Table 1.
Description of Description
of DDroneC-DB1
|
Kinds of sub-database |
Time |
Condition |
Description |
|
Sub-database 1 (drone landing) |
Morning |
Humidity: 41.5 %, wind speed: 1.4 m/s, temperature: 8.6 oC, spring, sunny |
- A sunny day with clear sky, which has affected the illumination on the marker - Landing speed: 4 m/s |
|
Afternoon |
Humidity: 73.8 %, wind speed: 2 m/s, temperature: -2.5 oC, winter,
cloudy |
- Low level of illumination observed in the winter time, which affected the intensity of background area. - Landing speed: 6 m/s |
|
|
Evening |
Humidity: 38.4 %, wind speed: 3.5 m/s, temperature: 3.5 oC, winter, windy |
- There is the change in the marker’s
position due to strong wind - Landing speed: 4 m/s |
|
|
Night |
Humidity: 37.5 %, wind speed: 3.2 m/s, temperature: 6.9 oC, spring, foggy |
- Marker cannot be seen owning low level of light at dark night - Landing speed: 6 m/s |
|
|
Sub-database 2 (drone hovering) |
Morning |
Humidity: 41.6 %, wind speed: 2.5 m/s, temperature: 11 oC, spring, foggy |
Drone hovers above the marker, and the marker is
manually moved and rotated while capturing videos |
|
Afternoon |
Humidity: 43.5 %, wind speed: 2.8 m/s, temperature: 13 oC, spring, sunny |
||
|
Evening |
Humidity: 42.9 %, wind speed: 2.9 m/s, temperature: 10 oC, spring, sunny |
||
|
Night |
Humidity: 41.5 %, wind speed: 3.1 m/s, temperature: 6oC, spring, dark
night |
(2) DDroneC-DB1 Request
To gain access to the database, download the following DDroneC-DB1 request form. Please scan
the request form and email to Mr. Phong Ha Nguyen (stormwindvn@dongguk.edu).
Any work that uses or incorporates the database must acknowledge the
authors by including the following reference.
Phong Ha Nguyen, Ki Wan Kim, Yong
Won Lee and Kang Ryoung Park, " Remote Marker-based Tracking for UAV Landing
Using Visible-Light Camera Sensor," in submission to Remote Sensing
===========================================================================================================================================================================================================
<
DDroneC-DB1 Request Form >
Please complete the following form to request access to the DDroneC-DB1 (All contents must be
completed). This database should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date
:
Name (signature)
2. Dongguk Visible Light & FIR Pedestrian Detection Database
(DVLFPD-DB1) && CNN model
(1) Introduction
There are 4 sub-databases, and the total number of frames of visible light images and FIR images is 4080 each. To obtain the images, this study used a duel camera system consisting of a FLIR Tau640 FIR camera (19 mm), and a Logitech C600 visible light web-camera. In order to record the filming conditions, a WH-1091 (wireless weather station) was used.
Table 1.
Description of Database
|
|
Sub-database 1 |
Sub-database 2 |
Sub-database 3 |
Sub-database 4 |
|
|
Number of image |
598 |
651 |
2364 |
467 |
|
|
Number of pedestrian candidate |
1123 |
566 |
2432 |
169 |
|
|
Number of non-pedestrian candidate |
763 |
734 |
784 |
347 |
|
|
(range of width) ´ (range of height) (pixels) |
Pedestrian |
(27 ~ 91) ´ (87 ~ 231) |
(47 ~ 85) ´ (85 ~ 163) |
(31 ~ 105) ´ (79 ~ 245) |
(30 ~ 40) ´ (90 ~ 120) |
|
Non-pedestrian |
(51 ~ 161) ´ (63 ~ 142) |
(29 ~ 93) ´ (49 ~ 143) |
(53 ~ 83) ´ (55 ~ 147) |
(60 ~ 170) ´ (50 ~ 110) |
|
|
Weather Conditions |
Surface Temperature: 30.4°C, Air Temperature: 22.5°C Wind speed: 10 km/h Sensory Temperature: 21.3°C |
Surface Temperature: 25.5°C Air Temperature: 24℃ Wind speed: 5
km/h Sensory Temperature: 23.5°C |
Surface Temperature: 20°C Air Temperature: 21°C Wind Speed: 6.1 km/h Sensory Temperature: 21°C |
Surface Temperature: 16°C Air Temperature: 20.5°C Wind Speed: 2.5 km/h Sensory Temperature: 20.8°C |
|
In addition, we obtained the
artificially degraded database by Gaussian noise and Gaussian blurring with the
original dataset of Table 1, and make this degraded dataset public, also.
Also, we made the classification
algorithm of pedestrian and non-pedestrian by convolutional neural network (CNN) trained with DVLFPD-DB1. We also
made this trained CNN model open to other researchers.
(2) DVLFPD-DB1 & CNN model Request
To gain access to the database and CNN models, download the following DVLFPD-DB1 & CNN model request form. Please scan the request form and email to Mr. Jin Kyu Kang (kangjinkyu@dgu.edu).
Any work that uses or incorporates the database & CNN Model must acknowledge the authors by including the following reference.
Jin Kyu Kang, Hyung Gil Hong, and Kang Ryoung Park, "Pedestrian Detection Based on Adaptive Selection of Visible Light or FIR Light Camera Image by Fuzzy Inference System and CNN-Based Verification," in submission to Sensors
===========================================================================================================================================================================================================
<
DVLFPD-DB1 & CNN model Request Form >
Please complete the following form to request access to the DVLFPD-DB1 & CNN model (All contents must be completed). This database and CNN model should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
3. Dongguk Open and Closed Eyes Database
(DOCE-DB1) & CNN model
(1) Introduction
DOCE-DB1 was created in an environment with a person watching TV indoors, with images taken at distances between 2 and 2.5 m. The camera used for capturing the images was a Logitech C600 web camera with a zoom lens attached. We used eye images obtained by extracting facial and eye features from images captured at a resolution of 1600 × 1200 pixels. Table 1 shows the descriptions of original and augmented DOCE-DB1.
Table 1.
Descriptions of original and augmented databases
|
|
Original database |
Augmented
database |
|
Open eye images |
2,062 |
103,100 |
|
Closed eye images |
4,763 |
238,150 |
|
Total |
6,825 |
341,250 |
In addition, we made the
classification algorithm of open and closed eye by convolutional neural network (CNN) trained with DOCE-DB1. We also
made this trained CNN model with other various (trained) CNN models (based on
AlexNet, VGG face, GoogLeNet) open to other researchers.
(2) DOCE-DB1 & CNN model Request
To gain access to the database and CNN models, download the following DOCE-DB1 & CNN model request form. Please scan the request form and email to Mr. Ki Wan Kim (yawara18@dgu.edu).
Any work that uses or incorporates the database & CNN Model must acknowledge the authors by including the following reference.
Ki Wan Kim, Hyung Gil Hong, Gi Pyo
Nam, and Kang Ryoung Park, "Deep Residual Network-based Classification of
Open and Closed Eyes using Visible Light Camera Sensor," in submission to
Sensors
===========================================================================================================================================================================================================
<
DOCE-DB1 & CNN model Request Form >
Please complete the following form to request access to the DOCE-DB1 & CNN model (All contents must be completed). This database and CNN model should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
4. Dongguk Multi-national Currencies
Database (DMC-DB1) & CNN model
(1) Introduction
Our multi-national banknote image database (DMC-DB1) containing
images from six national currencies, which are CNY, EUR, JPY, KRW, RUB and USD.
There were 64,668 images captured in both sides of 32,334 banknotes belonging
to 62 denominations of currencies from six countries. The number of classes is
four times of the number of denominations because of the inclusion of four
directions; therefore, there are totally 248 classes of banknote to be
classified in our study. In Table 1, we give the details numbers of images and
classes (denominations and directions) of each country’s
banknote in the dataset.
Table 1. Number of images and classes in our multi-national banknote database
|
Currency |
Number of Images |
Number of Classes |
|
CNY |
626 |
40 |
|
EUR |
4,324 |
44 |
|
JPY |
1,462 |
28 |
|
KRW |
536 |
28 |
|
RUB |
12,146 |
40 |
|
USD |
45,574 |
68 |
In addition, we made the
classification algorithm of multi-national currencies by convolutional neural network (CNN) trained with DMC-DB1. We also
made this trained CNN model open to other researchers.
(2) DMC-DB1 & CNN model Request
To gain access to the database and CNN model, download the following DMC-DB1 & CNN model request form. Please scan the request form and email to Prof. Tuyen Danh Pham (phamdanhtuyen@gmail.com).
Any work that uses or incorporates the database & CNN Model must acknowledge the authors by including the following reference.
Tuyen Danh Pham, Dong Eun Lee, and
Kang Ryoung Park, "Multi-National Banknote Classification based on
Visible-light Line Sensor and Convolutional Neural Network," in submission
to Sensors
===========================================================================================================================================================================================================
<
DMC-DB1 & CNN model Request Form >
Please complete the following form to request access to the DMC-DB1 & CNN model (All contents must be completed). This database and CNN model should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
5. Dongguk Finger-Vein Database (DFingerVein-DB1) & CNN Model
(1) Introduction
DFingerVein-DB1 includes two databases. The first database (good-quality database) was acquired with two guide bars (fingertip guide bar and finger side guide bar) attached, when the images were captured using the finger-vein capturing device made by our lab. Ten images each were captured from the index, middle, and ring fingers of the left and right hands of 20 persons. Thus, this database has a total of 1,200 images consisting of 20 people × 2 hands × 3 fingers × 10 images. This is called good-quality database, because the images have few finger-vein misalignments by finger translation owing to the two guide bars, and there is little shading in the finger-vein images.
For the second database (mid-quality database), the two guide bars were removed when the images were captured with the same device used for the collection of good-quality database. Ten images each were captured from the index, middle, and ring fingers of the left and right hands of 33 persons. Thus, this database has a total of 1,980 images consisting of 33 people × 2 hands × 3 fingers × 10 images. This is called mid-quality database because the images were captured with no guide bar, and the images have finger-vein misalignments by finger translation; however, there is little shading in the finger-vein images similar to the good-quality database.
Table 1.
Descriptions of the collected databases
|
|
Good-quality database |
Mid-quality database |
|
|
Original images |
# of images |
1,200 |
1,980 |
|
# of people |
20 |
33 |
|
|
# of hands |
2 |
2 |
|
|
# of fingers* |
3 |
3 |
|
|
# of classes (# of images per class) |
120 (10) |
198 (10) |
|
In addition, we made the
finger-vein recognition algorithm by convolutional
neural network (CNN) trained with DFingerVein-DB1. We also made this trained
CNN model open to other researchers.
(2) DFingerVein-DB1 & CNN Model Request
To gain access to the database & CNN Model, download the following DFingerVein-DB1 & CNN Model request form. Please scan the request form and email to Mr. Hyung Gil Hong (hell@dongguk.edu).
Any work that uses or incorporates the database & CNN Model must acknowledge the authors by including the following reference.
Hyung Gil Hong, Min Beom Lee, and Kang Ryoung Park, "Convolutional Neural Network-based Finger-vein Recognition Using NIR Image Sensor," in submission to Sensors
===========================================================================================================================================================================================================
<
DFingerVein-DB1 & CNN Model Request Form
>
Please complete the following form to request access to the DFingerVein-DB1 & CNN Model (All contents must be completed). This database & CNN Model should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
===========================================================================================================================================================================================================
6. ISPR database (real and presentation
attack finger-vein images)
(1) Introduction
The ISPR database consists of 3300 and 7560 images for real and presentation attack finger-vein images, respectively. The real finger-vein database was collected by capturing finger-vein images from 33 people, including both male and female. For each person, all 10 fingers were used and 10 trials were captured for each finger. Consequently, the real finger-vein database contains 3300 (33 people × 10 fingers × 10 trials) real finger-vein images for our experiments. Among 3300 real finger-vein images, we selected 56 fingers which have clear vein pattern for making the presentation attack finger-vein images.
There are two reasons for this selection scheme. First, the users should normally use a finger with clear vein pattern for finger-vein recognition system in order to guarantee the security level of biometric feature. Therefore, if an attacker steals a finger-vein pattern image from user, it will normally be a clear finger-vein image. Second, with some fingers which normally have poor vein pattern such as thumbs or little fingers it is very hard for attackers to produce a clear presentation attack finger-vein image. Instead, the presentation attack finger-vein images can contain poor vein pattern and much of noise due to the reproducing process. As a result, the attacking task can be failed by rejection rate of finger-vein recognition system.
The presentation attack finger-vein image database was collected by re-capturing the printed versions of 56 selected real finger-vein images in three different printing materials including A4 paper, MAT paper and OHP film. In addition, we used three different values of the printing resolution of low resolution (300dpi), middle resolution (1200dpi) and high resolution (2400dpi). By using this scheme, we can collect presentation attack finger-vein images which contain various characteristics about printing materials and printing resolution. Finally, in order to simulate the attacking process, we captured presentation attack finger-vein images at three z-distances (the distance between camera and finger-vein sample) by little changing the z-distance during image acquisition and 5 trials for each z-distance.
As a result, a presentation attack finger-vein image database of 7560 images (56 real image × 3 printing materials × 3 printing resolutions × 3 z-distances × 5 trials) was collected.
Table 1. Description of ISPR presentation attack finger-vein image database
|
Image Making Protocol |
Real Access |
Printed Access |
|||||
|
Train Set |
Test Set |
Total |
Train Set |
Test Set |
Total |
||
|
Material |
Printed on A4 Paper (ISPR-DB1) |
1700 |
1600 |
3300 |
1440 |
1080 |
2520 |
|
Printed on MAT Paper (ISPR-DB2) |
1700 |
1600 |
3300 |
1440 |
1080 |
2520 |
|
|
Printed on OHP Film (ISPR-DB3) |
1700 |
1600 |
3300 |
1440 |
1080 |
2520 |
|
|
Printer Resolution |
Printed using 300 DPI
Resolution Printer (ISPR-DB4) |
1700 |
1600 |
3300 |
1440 |
1080 |
2520 |
|
Printed using 1200 DPI
Resolution Printer. (ISPR-DB5) |
1700 |
1600 |
3300 |
1440 |
1080 |
2520 |
|
|
Printed using 2400 DPI
Resolution Printer (ISPR-DB6) |
1700 |
1600 |
3300 |
1440 |
1080 |
2520 |
|
|
Entire Database (ISPR-DB) |
1700 |
1600 |
3300 |
4320 |
3120 |
7560 |
|
(2) ISPR database Request
To gain access to the database, download the following ISPR database request form. Please scan the request form and email to Prof. Dat Tien Nguyen (nguyentiendat@dongguk.edu).
Any work that uses or incorporates the database must acknowledge the authors by including the following reference.
Dat Tien Nguyen, Hyo Sik Yoon and Kang Ryoung Park, "Presentation Attack Detection for Finger-vein Biometric System Based on Convolutional Neural Network and Support Vector Machine," in submission to Sensors
===========================================================================================================================================================================================================
<
ISPR database Request Form >
Please complete the following form to request access to the ISPR database (All contents must be completed). This database should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
===========================================================================================================================================================================================================
7. Dongguk Night-time Human Detection
Database (DNHD-DB1) & CNN model
(1) Introduction
This database was obtained
through a nighttime visible light camera (Logitech C600 web camera
(https://support.logitech.com/en_us/product/5869)). DNHD-DB1 was constructed
using images obtained from cameras fixed at various locations.
Table 1. Description of
DNHD-DB1
|
|
DNHD-DB1 |
|
|
Number of images |
Human |
19760 |
|
Background |
19760 |
|
|
#of channel |
Color (3 channels) |
|
|
Width of human (background) image (min. ~ max.) (pixels) |
15 ~ 219 |
|
|
Height of human (background) image (min. ~ max.) (pixels) |
45 ~ 313 |
|
|
Environment of database collection |
-
Image acquisition using a static camera in surveillance environment -
The height of the camera from the ground was about 6–10 m -
Database was collected at various places (at 8–10 pm) |
|
In addition, we made human
detection algorithm in nighttime image by convolutional
neural network (CNN) trained with DNHD-DB1. We also made this trained CNN
model open to other researchers.
(2) DNHD-DB1 & CNN model Request
To gain access to the database and CNN model, download the following DNHD-DB1 & CNN model request form. Please scan the request form and email to Mr. Jong Hyun Kim (zzingae@naver.com).
Any work that uses or incorporates the database must acknowledge the authors by including the following reference.
Jong Hyun Kim, Hyung Gil Hong, and Kang Ryoung Park, "Convolutional Neural Network-based Human Detection in Nighttime Image Using Visible Light Camera Sensor," accepted for publication in Sensors
===========================================================================================================================================================================================================
<
DNHD-DB1 & CNN model Request Form >
Please complete the following form to request access to the DNHD-DB1 & CNN model (All contents must be completed). This database and CNN model should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
===========================================================================================================================================================================================================
8. Dongguk Body-based Person Recognition
Database (DBPerson-Recog-DB1)
(1) Introduction
This database was collected from 412 persons with different body-view of subjects such as front, back, and side views. The cameras (visible light and thermal cameras) were placed near to each other to create a dual-camera set up, and placed at a height of roughly 6 meters. As the visible light camera, a webcam camera (C600) (https://support.logitech.com/en_us/product/5869) was used, and a Tau2 camera (http://www.flir.com/cores/display/?id=54717) was used for capturing thermal images. The visible camera captures images with a size of 800 × 600 pixels. The thermal camera captures images with a size of 640 × 480 pixels.
For each person, 10 images
were respectively captured by visible light and thermal cameras to simulate the
variation of body gaits. In total, we collected a database including 8,240
images (4,120 visible light images and 4,120 corresponding thermal images).
Among the 412 subjects, there were 254 males and 158 females.
In addition, the following
augmented images are included in this database.
Table 1. Description
of the training and testing databases in our experiments
|
Database |
Males |
Females |
Total |
|
|
Training
Database |
Number of Persons |
204 (persons) |
127 (persons) |
331 (persons) |
|
Number of Original Images |
4080 images (204×20) |
2540 images (127×20) |
6620 (images) |
|
|
Number of Artificial Images |
20400 images (204×20×5) |
10160 images (127×20×5) |
33100 images |
|
|
Testing Database |
Number of Persons |
50 (persons) |
31 (persons) |
81 (persons) |
|
Number of Original Images |
1000 images (50×20) |
620 images (31×20) |
1620 images |
|
|
Number of Artificial Images |
5000 images (50×20×5) |
3100 images (31×20×5) |
8100 images |
|
(2) DBPerson-Recog-DB1 Request
To gain access to the database, download the following DBPerson-Recog-DB1 request form. Please scan the request form and email to Dr. Dat Tien Nguyen (nguyentiendat@dongguk.edu).
Any work that uses or incorporates the database must acknowledge the authors by including the following reference.
Dat Tien Nguyen, Hyung Gil Hong, Ki Wan Kim, and Kang Ryoung Park, "Person Recognition System Based on a Combination of Body Images from Visible Light and Thermal Cameras," Sensors, Vol. 17, Issue 3(605), pp. 1-29, March 2017
===========================================================================================================================================================================================================
<
DBPerson-Recog-DB1 Request Form >
Please complete the following form to request access to the DBPerson-Recog-DB1 (All contents must be completed). This database should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
===========================================================================================================================================================================================================
9. Dongguk Body-based Gender Database
(DBGender-DB2)
(1) Introduction
This database was collected from 412 persons with different body-view of subjects such as front, back, and side views. The cameras (visible light and thermal cameras) were placed near to each other to create a dual-camera set up, and placed at a height of roughly 6 meters. As the visible light camera, a webcam camera (C600) (https://support.logitech.com/en_us/product/5869) was used, and a Tau2 camera (http://www.flir.com/cores/display/?id=54717) was used for capturing thermal images. The visible camera captures images with a size of 800 × 600 pixels. The thermal camera captures images with a size of 640 × 480 pixels.
For each person, 10 images
were respectively captured by visible light and thermal cameras to simulate the
variation of body gaits. In total, we collected a database including 8,240
images (4,120 visible light images and 4,120 corresponding thermal images).
Among the 412 subjects, there were 254 males and 158 females.
In addition, the following
augmented images are included in this database.
Table 1. Description of the training and testing sub-databases and the
corresponding augmented databases for our experiments
|
Database |
Males |
Females |
Total |
||
|
Augmented
database |
Learning
database |
Number of
persons |
204
(persons) |
127 (persons) |
331
(persons) |
|
Number of
images |
73,440 |
76,200 |
149,640
(images) |
||
|
Testing
database |
Number of
persons |
50 (persons) |
31 (persons) |
81 (persons) |
|
|
Number of
images |
18,000 |
18,600 |
36,600
(images) |
||
(2) DBGender-DB2 Request
To gain access to the database, download the following DBGender-DB2 request form. Please scan the request form and email to Dr. Dat Tien Nguyen (nguyentiendat@dongguk.edu).
Any work that uses or incorporates the database must acknowledge the authors by including the following reference.
Dat Tien Nguyen, Ki Wan Kim, Hyung Gil Hong, Ja Hyung Koo, Min Cheol Kim and Kang Ryoung Park, "Gender Recognition from Human-Body Images Using Visible-Light and Thermal Camera Videos Based on a Convolutional Neural Network for Image Feature Extraction," Sensors, Vol. 17, Issue 3(637), pp. 1-22, March 2017
===========================================================================================================================================================================================================
<
DBGender-DB2 Request Form >
Please complete the following form to request access to the DBGender-DB2 (All contents must be completed). This database should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
===========================================================================================================================================================================================================
10. Dongguk Activities & Actions
Database (DA&A-DB1)
(1) Introduction
Our database (shown in Table
1) consists of 12 datasets that have been acquired in different environments
and 11 different types of behavior have been included in it. Here, the 11
behaviors include “lying down”, “sitting”, “standing”, “walking”, “running”, “approaching”, “leaving”, “hand waving with two hands”, “hand waving with one hand”, “punching”, and “kicking”. As shown in Table 1, the
database has been collected in six different places (“Behind the building
including many windows,”
… “Car road”) during the night and day,
in varying environments (humidity level, wind velocity, temperature, and
weather) in four seasons, using different positions of the camera setup of
Table 2.
The database includes both
visible and thermal images. The database consists of 194,094 images. In our
database, the human width varies between 28 and 117 pixels and the height
varies in between 50 and 265 pixels.
The cameras (visible light and thermal cameras) were placed near to each other to create a dual-camera set up. As the visible light camera, a webcam camera (C600) (https://support.logitech.com/en_us/product/5869) was used, and a Tau2 camera (http://www.flir.com/cores/display/?id=54717) was used for capturing thermal images. The visible camera captures images with a size of 800 × 600 pixels. The thermal camera captures images with a size of 640 × 480 pixels.
Table 1. Description of the all 12 datasets.
|
Dataset |
Place |
Condition |
Detail
Description |
|
I |
Behind the building
including many windows |
humidity 73.0%,
wind 1.6 m/s, 1.2 °C, morning, winter |
- In winter time, the window of the building has affected the intensity of background area - The gray level of the window is similar to that of human in
the thermal image |
|
II |
humidity
73.0%, wind 1.5 m/s, −1.0 °C, evening, winter |
- In winter time, the window of the building has affected the intensity of background area - Object cannot be seen in the visible light image |
|
|
III |
Forest of low mountain |
humidity
50.6%, wind 1.7 m/s, 1.0 °C, afternoon, cloudy, spring |
- Leaves and trees have affected the intensity of background area |
|
IV |
humidity
50.6%, wind 1.8 m/s, −2.0 °C, dark night, spring |
- Leaves and trees have affected the intensity of background area - Object cannot be seen in the visible light image |
|
|
V |
Narrow walk road between
two buildings |
humidity
43.4%, wind 3.1 m/s, 14.0 °C, afternoon, sunny, autumn |
- The temperature outside the building is increased by the air heating system of the
building in the image |
|
VI |
humidity
43.4%, wind 3.1 m/s, 5.0 °C, dark night, autumn |
- Object cannot be seen in the visible light image |
|
|
VII |
Small square-1 in front
of building |
humidity
39.6%, wind 1.9 m/s, −6.0 °C, afternoon, cloudy, winter |
- Object is small in the visible light image due to the far position
of the camera, which makes it difficult for behavior recognition |
|
VIII |
humidity
39.6%, wind 1.7 m/s, −10.0 °C, dark night, winter |
- Halo effect generated near the foot of the human area makes it
difficult to detect the correct human region from the background
area - Object cannot be seen in the visible light image |
|
|
IX |
Small square-2 in front
of building |
humidity
62.6%, wind 1.3 m/s, 21.9 °C, afternoon, cloudy, autumn |
- Object is small in the visible light image due to the far position
of the camera, which makes it difficult for behavior recognition - The high
background temperature makes it difficult to segment the human area from the background when the human is
positioned in the left-upper part of the background |
|
X |
humidity
48.3%, wind 2.0 m/s, −10.9 °C, dark night, autumn |
- Object cannot be seen in the visible light image - Halo effect occurs near the foot of the
human area, which makes it difficult to detect the correct human region from the background
area |
|
|
XI |
Car road |
humidity
60.0%, wind 1.0 m/s, 27.0 °C, afternoon, sunny, summer |
- During summer, the temperature of road is higher
than the human which causes the human to
appear darker than the road |
|
XII |
humidity
58.6%, wind 1.2 m/s, 20.2 °C, dark night, summer |
- During summer, the temperature of road is higher
than the human which causes the human to appear darker than the road - Object cannot be seen in the visible light image |
Table 2. Details of our camera setup used to collect the 12 datasets (unit:
meters).
|
Datasets |
Height |
Horizontal
Distance |
Z Distance |
|
Datasets I and II |
8 |
10 |
12.8 |
|
Datasets III and IV |
7.7 |
11 |
13.4 |
|
Datasets V and VI |
5 |
15 |
15.8 |
|
Datasets VII and VIII |
10 |
15 |
18 |
|
Datasets IX and X |
10 |
15 |
18 |
|
Datasets XI and XII |
6 |
11 |
12.5 |
(2) DA&A-DB1 Request
To gain access to the database, download the following DA&A-DB1 request form. Please scan the request form and email to Ganbayar Batchuluun (ganabata87@gmail.com).
Any work that uses or incorporates the database must acknowledge the authors by including the following reference.
Ganbayar Batchuluun, Jong Hyun Kim, Hyung Gil Hong, Jin Kyu Kang, and Kang Ryoung Park, “Fuzzy System based Human Behavior Recognition by Combining Behavior Prediction and Recognition,” Expert Systems with Applications, Vol. 81, pp. 108-133, September 2017.
===========================================================================================================================================================================================================
< DA&A-DB1 Request Form >
Please complete the following form to request access to the DA&A-DB1 (All contents must be completed). This database should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
===========================================================================================================================================================================================================
11. Dongguk Body-based Gender Database
(DBGender-DB1)
(1) Introduction
This database was collected from 412 persons with different body-view of subjects such as front, back, and side views. The cameras (visible light and thermal cameras) were placed near to each other to create a dual-camera set up, and placed at a height of roughly 6 meters. As the visible light camera, a webcam camera (C600) (https://support.logitech.com/en_us/product/5869) was used, and a Tau2 camera (http://www.flir.com/cores/display/?id=54717) was used for capturing thermal images. The visible camera captures images with a size of 800 × 600 pixels. The thermal camera captures images with a size of 640 × 480 pixels.
For each person, 10 images were respectively captured by visible light and thermal cameras to simulate the variation of body gaits. In total, we collected a database including 8,240 images (4,120 visible light images and 4,120 corresponding thermal images). Among the 412 subjects, there were 254 males and 158 females.
(2) DBGender-DB1 Request
To gain access to the database, download the following DBGender-DB1 request form. Please scan the request form and email to Dr. Dat Tien Nguyen (nguyentiendat@dongguk.edu).
Any work that uses or incorporates the database must acknowledge the authors by including the following reference.
Dat Tien Nguyen and Kang Ryoung Park, “Enhanced Gender Recognition System Using an Improved Histogram of Oriented Gradient (HOG) Feature from Quality Assessment of Visible Light and Thermal Images of the Human Body,” Sensors, Vol. 16, Issue 7(1134), pp. 1-25, July 2016
===========================================================================================================================================================================================================
< DBGender-DB1
Request Form >
Please complete the following form to request access to the DBGender-DB1 (All contents must be completed). This database should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
===========================================================================================================================================================================================================
12. Dongguk Face Database (DFace-DB1)
(1) Introduction
This database was created by collecting face images from 15 people
using conventional web-camera. In order to capture images of users looking at
the TV screen naturally, each participant was instructed to watch TV without
any restrictions. As a result, we captured a total of 1,350 frames (database I)
(15 persons × 2 quantities of
participants (1 person or 3 persons) × 3 seating
positions (left, middle, and right) × 3 Z distances
(1.5, 2, and 2.5 m) × 5 trials (looking naturally)).
In
addition, a total
of 300 images (database II) (5
persons × 3 Z distances (1.5 m, 2 m, and 2.5 m) × 2
lying directions (left and right) × 10 images) were
collected for experiment when each people is lying on his or her side.
For face registration for recognition, additional 75 frames (15 persons × 5 TV gaze points) were obtained at the Z distance of 2 m.
(2) DFace-DB1 Request
To gain access to the database, download the following DFace-DB1 request form. Please scan the request form and email to Hyung Gil Hong (hell@dongguk.edu).
Any work that uses or incorporates the database must acknowledge the authors by including the following reference.
H. G. Hong, W. O. Lee, Y. G. Kim, K. W. Kim, and K. R. Park, “Fuzzy System-Based Face Detection Robust to In-Plane Rotation Based on Symmetrical Characteristics of a Face,” Symmetry-Basel, Vol. 8, Issue 8(75), pp. 1-28, August 2016
===========================================================================================================================================================================================================
< DFace-DB1 Request Form >
Please complete the following form to request access to the DFace-DB1 (All contents must be completed). This database should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
===========================================================================================================================================================================================================
13. Dongguk Banknote Database
(DBanknote-DB1)
(1) Introduction
This database was collected for measuring the performance of classifying fitness and unfitness caused by soiling level on banknote. The database includes three datasets such as US dollar (USD), Korean won (KRW), and Indian rupee (INR) datasets. Detail descriptions of each dataset are shown in Tables 1 ~ 4. In all the Tables, direction A shows an image of the banknote in the front side and forward direction, direction B shows an image in the front side and backward direction, C represents an image in the back side and forward direction. Direction D shows the image of the banknote in the back side and backward direction. Based on actual densitometer measurement values about soiling levels, human experts classified the banknotes in the database as fit banknotes (capable of being used) or unfit banknotes (incapable of use).
Table 1.
Total number of images in USD dataset
|
Direction Denomination |
A |
B |
C |
D |
|
$1 |
148 |
|||
|
$2 |
20 |
|||
|
$5 |
88 |
|||
|
$10 |
136 |
|||
|
$20 |
160 |
|||
|
$50 |
212 |
|||
|
$100 |
200 |
|||
Table 2.
Total number of images in KRW dataset
|
Direction Denomination |
A |
B |
C |
D |
|
₩1000 |
300 |
|||
|
₩5000 |
293 |
|||
|
₩10000 |
198 |
|||
|
₩50000 |
198 |
|||
Table 3.
Total number of images in INR dataset
|
Direction Denomination |
A |
B |
C |
D |
|
₨10 |
112 |
|||
|
₨20 |
72 |
|||
|
₨50 |
67 |
|||
|
₨100 |
171 |
|||
|
₨500 |
101 |
|||
|
₨1000 |
52 |
|||
Table 4. Number of
images and classes for USD, KRW and INR datasets
|
Currency |
Number of images |
Number of classes |
|
USD |
3856 |
28 |
|
KRW |
3956 |
16 |
|
INR |
2300 |
24 |
(2) DBanknote-DB1 Request
To gain access to the database, download the following DBanknote-DB1 request form. Please scan the request form and email to Seung Yong Kwon (sbaru07@dgu.edu).
Any work that uses or incorporates the database must acknowledge the authors by including the following reference.
Seung Yong Kwon, Tuyen Danh Pham, Kang Ryoung Park, Dae Sik Jeong, and Sungsoo Yoon, "Recognition of Banknote Fitness Based on a Fuzzy System Using Visible Light Reflection and Near-infrared Light Transmission Images," Sensors, Vol. 16, Issue 6(863), pp. 1-18, June 2016
===========================================================================================================================================================================================================
< DBanknote-DB1
Request Form >
Please complete the following form to request access to the DBanknote-DB1 (All contents must be completed). This database should not be used for commercial use.
Name :
Contact : (Email)
(Telephone)
Organization Name :
Organization Address :
Purpose :
Date :
Name (signature)
===========================================================================================================================================================================================================