首页> 外文期刊>Neural computing & applications >One deep music representation to rule them all? A comparative analysis of different representation learning strategies
【24h】

One deep music representation to rule them all? A comparative analysis of different representation learning strategies

机译:一个深入的音乐表示来统治它们吗? 不同代表学习策略的比较分析

获取原文
获取原文并翻译 | 示例
           

摘要

Inspired by the success of deploying deep learning in the fields of Computer Vision and Natural Language Processing, this learning paradigm has also found its way into the field of Music Information Retrieval. In order to benefit from deep learning in an effective, but also efficient manner, deep transfer learning has become a common approach. In this approach, it is possible to reuse the output of a pre-trained neural network as the basis for a new learning task. The underlying hypothesis is that if the initial and new learning tasks show commonalities and are applied to the same type of input data (e.g., music audio), the generated deep representation of the data is also informative for the new task. Since, however, most of the networks used to generate deep representations are trained using a single initial learning source, their representation is unlikely to be informative for all possible future tasks. In this paper, we present the results of our investigation of what are the most important factors to generate deep representations for the data and learning tasks in the music domain. We conducted this investigation via an extensive empirical study that involves multiple learning sources, as well as multiple deep learning architectures with varying levels of information sharing between sources, in order to learn music representations. We then validate these representations considering multiple target datasets for evaluation. The results of our experiments yield several insights into how to approach the design of methods for learning widely deployable deep data representations in the music domain.
机译:灵感来自在计算机视觉和自然语言处理领域部署深度学习的成功,这一学习范式也发现了进入音乐信息检索领域的方式。为了使深深的学习中受益于有效,也有效的方式,深度转移学习已成为一种共同的方法。在这种方法中,可以将预先训练的神经网络的输出重用作为新学习任务的基础。底层假设是,如果初始和新的学习任务显示共性,并且应用于相同类型的输入数据(例如,音频音频),则数据的生成深度表示也是新任务的信息。然而,由于使用的大多数用于生成深度表示的网络使用单个初始学习来源训练,因此它们的表示不太可能是所有可能的未来任务的信息。在本文中,我们提出了我们调查的结果,对音乐领域的数据和学习任务产生深刻表现的最重要因素的结果。我们通过广泛的实证研究进行了这一调查,涉及多个学习来源,以及多个深入学习架构,具有不同级别的信息之间的信息共享,以便学习音乐表示。然后,考虑多个目标数据集进行评估,我们验证这些表示。我们的实验结果会产生多种洞察,如何探讨如何在音乐域中学习广泛部署的深度数据表示的方法设计。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号