首页    期刊浏览 2024年07月08日 星期一
登录注册

文章基本信息

  • 标题:Decision Aggregation in Distributed Classification by a Transductive Extension of Maximum Entropy/Improved Iterative Scaling
  • 本地全文:下载
  • 作者:David J. Miller ; Yanxin Zhang ; George Kesidis
  • 期刊名称:EURASIP Journal on Advances in Signal Processing
  • 印刷版ISSN:1687-6172
  • 电子版ISSN:1687-6180
  • 出版年度:2008
  • 卷号:2008
  • DOI:10.1155/2008/674974
  • 出版社:Hindawi Publishing Corporation
  • 摘要:

    In many ensemble classification paradigms, the function which combines local/base classifier decisions is learned in a supervised fashion. Such methods require common labeled training examples across the classifier ensemble. However, in some scenarios, where an ensemble solution is necessitated, common labeled data may not exist: (i) legacy/proprietary classifiers, and (ii) spatially distributed and/or multiple modality sensors. In such cases, it is standard to apply fixed (untrained) decision aggregation such as voting, averaging, or naive Bayes rules. In recent work, an alternative transductive learning strategy was proposed. There, decisions on test samples were chosen aiming to satisfy constraints measured by each local classifier. This approach was shown to reliably correct for class prior mismatch and to robustly account for classifier dependencies. Significant gains in accuracy over fixed aggregation rules were demonstrated. There are two main limitations of that work. First, feasibility of the constraints was not guaranteed. Second, heuristic learning was applied. Here, we overcome these problems via a transductive extension of maximum entropy/improved iterative scaling for aggregation in distributed classification. This method is shown to achieve improved decision accuracy over the earlier transductive approach and fixed rules on a number of UC Irvine datasets.

国家哲学社会科学文献中心版权所有