dc.contributor.authors |
Aydemir, Emrah; Yalcinkaya, Mehmet Ali; Barua, Prabal Datta; Baygin, Mehmet; Faust, Oliver; Dogan, Sengul; Chakraborty, Subrata; Tuncer, Turker; Acharya, U. Rajendra |
|
dc.date.accessioned |
2023-01-24T12:08:55Z |
|
dc.date.available |
2023-01-24T12:08:55Z |
|
dc.date.issued |
2022 |
|
dc.identifier.uri |
http://dx.doi.org/10.3390/ijerph19041939 |
|
dc.identifier.uri |
https://hdl.handle.net/20.500.12619/99704 |
|
dc.description |
Bu yayın 06.11.1981 tarihli ve 17506 sayılı Resmî Gazete’de yayımlanan 2547 sayılı Yükseköğretim Kanunu’nun 4/c, 12/c, 42/c ve 42/d maddelerine dayalı 12/12/2019 tarih, 543 sayılı ve 05 numaralı Üniversite Senato Kararı ile hazırlanan Sakarya Üniversitesi Açık Bilim ve Açık Akademik Arşiv Yönergesi gereğince telif haklarına uygun olan nüsha açık akademik arşiv sistemine açık erişim olarak yüklenmiştir. |
|
dc.description.abstract |
Mask usage is one of the most important precautions to limit the spread of COVID-19. Therefore, hygiene rules enforce the correct use of face coverings. Automated mask usage classification might be used to improve compliance monitoring. This study deals with the problem of inappropriate mask use. To address that problem, 2075 face mask usage images were collected. The individual images were labeled as either mask, no masked, or improper mask. Based on these labels, the following three cases were created: Case 1: mask versus no mask versus improper mask, Case 2: mask versus no mask + improper mask, and Case 3: mask versus no mask. This data was used to train and test a hybrid deep feature-based masked face classification model. The presented method comprises of three primary stages: (i) pre-trained ResNet101 and DenseNet201 were used as feature generators; each of these generators extracted 1000 features from an image; (ii) the most discriminative features were selected using an improved RelieF selector; and (iii) the chosen features were used to train and test a support vector machine classifier. That resulting model attained 95.95%, 97.49%, and 100.0% classification accuracy rates on Case 1, Case 2, and Case 3, respectively. Having achieved these high accuracy values indicates that the proposed model is fit for a practical trial to detect appropriate face mask use in real time. |
|
dc.language |
English |
|
dc.language.iso |
eng |
|
dc.publisher |
MDPI |
|
dc.relation.isversionof |
10.3390/ijerph19041939 |
|
dc.subject |
Environmental Sciences & Ecology |
|
dc.subject |
Public, Environmental & Occupational Health |
|
dc.subject |
face mask detection |
|
dc.subject |
ResNet101 |
|
dc.subject |
DenseNet201 |
|
dc.subject |
transfer learning |
|
dc.subject |
hybrid feature selector |
|
dc.subject |
support vector machine |
|
dc.title |
Hybrid Deep Feature Generation for Appropriate Face Mask Use Detection |
|
dc.type |
Article |
|
dc.contributor.authorID |
Aydemir, Emrah/0000-0002-8380-7891 |
|
dc.contributor.authorID |
DOGAN, Sengul/0000-0001-9677-5684 |
|
dc.contributor.authorID |
chakraborty, subrata/0000-0002-0102-5424 |
|
dc.contributor.authorID |
Faust, Oliver/0000-0002-3979-4077 |
|
dc.contributor.authorID |
TUNCER, Turker/0000-0002-5126-6445 |
|
dc.contributor.authorID |
Barua, Prabal Datta/0000-0001-5117-8333 |
|
dc.contributor.authorID |
Acharya, U Rajendra/0000-0003-2689-8552 |
|
dc.identifier.volume |
19 |
|
dc.relation.journal |
INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH |
|
dc.identifier.issue |
4 |
|
dc.identifier.doi |
10.3390/ijerph19041939 |
|
dc.identifier.eissn |
1660-4601 |
|
dc.contributor.author |
Aydemir, Emrah |
|
dc.contributor.author |
Yalcinkaya, Mehmet Ali |
|
dc.contributor.author |
Barua, Prabal Datta |
|
dc.contributor.author |
Baygin, Mehmet |
|
dc.contributor.author |
Faust, Oliver |
|
dc.contributor.author |
Dogan, Sengul |
|
dc.contributor.author |
Chakraborty, Subrata |
|
dc.contributor.author |
Tuncer, Turker |
|
dc.contributor.author |
Acharya, U. Rajendra |
|
dc.relation.publicationcategory |
Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı |
|
dc.rights.openaccessdesignations |
Green Published, gold, Green Accepted |
|