Abstract

Convolution kernels, such as sequence and tree kernels, are advantageous for both the concept and accuracy of many natural language processing (NLP) tasks. Experiments have, however, shown that the over-fitting problem often arises when these kernels are used in NLP tasks. This paper discusses this issue of convolution kernels, and then proposes a new approach based on statistical feature selection that avoids this issue. To enable the proposed method to be executed efficiently, it is embedded into an original kernel calculation process by using sub-structure mining algorithms. Experiments are undertaken on real NLP tasks to confirm the problem with a conventional method and to compare its performance with that of the proposed method.

Keywords

Computer scienceKernel (algebra)Convolution (computer science)Artificial intelligenceSelection (genetic algorithm)Feature selectionProcess (computing)Feature (linguistics)Pattern recognition (psychology)Natural language processingKernel methodMachine learningAlgorithmSupport vector machineMathematicsProgramming languageArtificial neural network

Affiliated Institutions

Related Publications

Publication Info

Year
2004
Type
article
Pages
119-es
Citations
37
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

37
OpenAlex

Cite This

Jun Suzuki, Hideki Isozaki, Eisaku Maeda (2004). Convolution kernels with feature selection for natural language processing tasks. , 119-es. https://doi.org/10.3115/1218955.1218971

Identifiers

DOI
10.3115/1218955.1218971