论文标题

学习软件:小型型号做得很大

Learnware: Small Models Do Big

论文作者

Zhou, Zhi-Hua, Tan, Zhi-Hao

论文摘要

有人抱怨当前的机器学习技术,例如需要大量的培训数据和熟练的培训技巧,持续学习的困难,灾难性遗忘的风险,数据隐私/专有性的泄漏等。大多数研究工作都集中在分别关注的问题上,对大多数问题的关注较少,因为大多数问题在实践中遇到了大多数问题。普遍的大型模型范式在自然语言处理和计算机视觉应用中取得了令人印象深刻的结果,但尚未解决这些问题,而成为严重的碳排放源。本文概述了学习软件范式,该范式试图使用户无需从头开始构建机器学习模型,希望重复使用小型模型,甚至超出其原始目的,而关键成分是规范,可以使受过训练的模型充分地根据未来的用户进行充分识别,以至于未来对模型的需求不了解该模型。

There are complaints about current machine learning techniques such as the requirement of a huge amount of training data and proficient training skills, the difficulty of continual learning, the risk of catastrophic forgetting, the leaking of data privacy/proprietary, etc. Most research efforts have been focusing on one of those concerned issues separately, paying less attention to the fact that most issues are entangled in practice. The prevailing big model paradigm, which has achieved impressive results in natural language processing and computer vision applications, has not yet addressed those issues, whereas becoming a serious source of carbon emissions. This article offers an overview of the learnware paradigm, which attempts to enable users not need to build machine learning models from scratch, with the hope of reusing small models to do things even beyond their original purposes, where the key ingredient is the specification which enables a trained model to be adequately identified to reuse according to the requirement of future users who know nothing about the model in advance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源