Improving Multi-Domain Learning by Balancing Batches With Domain Information

Alexandre Thurow Bender, Emillyn Mellyne Gobetti Souza, Ihan Belmonte Bender,Ulisses Brisolara Corrêa,Ricardo Matsumura Araujo

WebMedia '23: Proceedings of the 29th Brazilian Symposium on Multimedia and the Web(2023)

引用 0|浏览4
暂无评分
摘要
Collections of data obtained or generated under similar conditions are referred to as domains or data sources. The distinct conditions of data acquisition or generation are often neglected, but understanding them is vital to address any phenomena emerging from these differences that might hinder model generalization. Multi-domain learning seeks to find the best way to train a model so that it performs adequately in all domains used during training. This paper explores multi-domain learning techniques that use explicit information about the domain of an example, in addition to its class. This study evaluates a general approach (Stew) by mixing all available data and also proposes two novel batch domain-regularization methods: Balanced Domains and Loss Sum. We train machine learning models with the listed approaches using datasets with multiple sources of data for image and audio classification tasks. The results suggest training a model using the Loss Sum method improves the results of models otherwise trained in a mix of all available data.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要