Distilling knowledge of neural networks for image analysis, model compression, data protection and minimization

dc.contributor.advisor Töreyin, Behçet Uğur
dc.contributor.author Keser, Reyhan Kevser
dc.contributor.authorID 708182005
dc.contributor.department Information and Communication Engineering
dc.date.accessioned 2025-01-03T06:39:12Z
dc.date.available 2025-01-03T06:39:12Z
dc.date.issued 2024-07-04
dc.description Thesis (Ph.D.) -- Istanbul Technical University, Graduate School, 2024
dc.description.abstract Knowledge distillation is an effective tool for model training, which refers to the process of knowledge transfer between models. In the context of knowledge distillation, the model to be trained with the injected knowledge is named student, where the teacher refers to the model whose knowledge is acquired. It can be exploited for various aims including improving model performance, accelerating the model, and reducing model parameters. Further, with the advent of diverse distillation schemes, it can be efficiently applied in various scenarios and problems. Thus, it has a wide range of application fields including computer vision and natural language processing. This thesis comprises the studies conducted on numerous problems of knowledge distillation, as well as the literature review. The first problem we focus on is hint position selection as an essential element in hint distillation, which is transferring features extracted in intermediate layers, namely hints. First, we demonstrate the importance of the determination of the hint positions. Then, we propose an efficient hint point selection methodology based on layer clustering. For this purpose, we exploit the k-means algorithm with specially designed metrics for layer comparison. We validate our approach by conducting comprehensive experiments utilizing various architectures for teacher-student pairs, hint types, and hint distillation methods, on two well-known image classification datasets. The results indicate that the proposed method achieves superior performance compared to the conventional approach. Another problem focused on in this thesis is model stealing, which refers to acquiring knowledge of a model that is desired to be protected due to the privacy concerns or commercial purposes. Since knowledge distillation can be exploited for model stealing, the concept of the undistillable teacher has been introduced recently, which aims to protect the model from stealing its knowledge via distillation. To contribute to this field, we propose an approach called averager student, whose goal is distilling the undistillable teacher, in this thesis. We evaluate the proposed approach for given teachers which are undistillable or normal. The results suggest that the proposed method outperforms the compared methods whose aim is the same as ours. The last problem we addressed is cross distillation, which means the distillation process between teacher and student models that operate on different modalities. In this work, we introduce a cross distillation scheme that transfers the compressed domain knowledge to the pixel domain. Further, we employ hint distillation which utilizes our previously proposed hint selection method. We evaluate our approach on two computer vision tasks, that are object detection and recognition. The results demonstrate that compressed domain knowledge can be efficiently exploited in a task in the pixel domain via the proposed approach. The proposed approaches in the context of the thesis, contribute to studies on image analysis, model compression, data protection, and minimization. First, our study on the selection of efficient hint positions aims to improve model compression performance, although the proposed approach can also be employed for other distillation schemes. The gains of our method in terms of model compression are presented as well as the performance results of the proposed algorithm. Then, our work on model stealing targets to contribute to the literature on model intellectual property (IP) protection and data protection, where we introduce an algorithm to distill a protected model's knowledge. Moreover, our study on cross distillation provides a contribution to data protection and minimization studies, where we propose a distillation methodology that utilizes compressed domain knowledge on pixel domain problems. Our approach demonstrates a technique that expands limited knowledge by employing different modality data instead of more samples. Since we utilize compressed domain images and eliminate the need for more samples to boost performance, we prevent the use of more data that may be personal or sensitive.
dc.description.degree Ph.D.
dc.identifier.uri http://hdl.handle.net/11527/26091
dc.language.iso en_US
dc.publisher Graduate School
dc.sdg.type Goal 7: Affordable and Clean Energy
dc.sdg.type Goal 9: Industry, Innovation and Infrastructure
dc.subject image analysis
dc.subject görüntü analizi
dc.subject neural networks
dc.subject sinir ağları
dc.subject data protection
dc.subject veri koruma
dc.subject artificial neural networks
dc.subject yapay sinir ağları
dc.title Distilling knowledge of neural networks for image analysis, model compression, data protection and minimization
dc.title.alternative Görüntü analizi, model sıkıştırma, veri koruma ve minimizasyonu için yapay sinir ağlarının bilgisinin damıtılması
dc.type Doctoral Thesis
Dosyalar
Orijinal seri
Şimdi gösteriliyor 1 - 1 / 1
thumbnail.default.alt
Ad:
708182005.pdf
Boyut:
2.66 MB
Format:
Adobe Portable Document Format
Açıklama
Lisanslı seri
Şimdi gösteriliyor 1 - 1 / 1
thumbnail.default.placeholder
Ad:
license.txt
Boyut:
1.58 KB
Format:
Item-specific license agreed upon to submission
Açıklama