论文标题

通过Meta-Heuristics微调改善预训练的权重

Improving Pre-Trained Weights Through Meta-Heuristics Fine-Tuning

论文作者

de Rosa, Gustavo H., Roder, Mateus, Papa, João Paulo, Santos, Claudio F. G. dos

论文摘要

在过去的十年中,已经对机器学习算法进行了广泛的研究,从而在广泛的应用程序中取得了前所未有的进步,例如图像分类和重建,对象识别和文本分类。尽管如此,大多数机器学习算法都是通过基于衍生的优化器(例如随机梯度下降)训练的,导致可能的局部最佳捕集,并抑制它们实现适当的性能。以生物风格的替代方案,称为元热疗法的传统优化技术,由于其简单性和避免局部最佳监禁的能力而受到了极大的关注。在这项工作中,我们建议使用元热疗法来微调预训练的权重,探索搜索空间的其他区域并提高其有效性。实验评估包括两个分类任务(图像和文本),并在四个文献数据集下进行评估。实验结果表明,自然风格的算法在探索预训练重量的邻居方面的能力比其对应的预训练的架构获得了优越的结果。此外,对不同体系结构的彻底分析,例如多层感知器和经常性神经网络,试图可视化并提供更精确的见解,以在学习过程中对最关键的权重进行微调。

Machine Learning algorithms have been extensively researched throughout the last decade, leading to unprecedented advances in a broad range of applications, such as image classification and reconstruction, object recognition, and text categorization. Nonetheless, most Machine Learning algorithms are trained via derivative-based optimizers, such as the Stochastic Gradient Descent, leading to possible local optimum entrapments and inhibiting them from achieving proper performances. A bio-inspired alternative to traditional optimization techniques, denoted as meta-heuristic, has received significant attention due to its simplicity and ability to avoid local optimums imprisonment. In this work, we propose to use meta-heuristic techniques to fine-tune pre-trained weights, exploring additional regions of the search space, and improving their effectiveness. The experimental evaluation comprises two classification tasks (image and text) and is assessed under four literature datasets. Experimental results show nature-inspired algorithms' capacity in exploring the neighborhood of pre-trained weights, achieving superior results than their counterpart pre-trained architectures. Additionally, a thorough analysis of distinct architectures, such as Multi-Layer Perceptron and Recurrent Neural Networks, attempts to visualize and provide more precise insights into the most critical weights to be fine-tuned in the learning process.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源