Author Search Result

[Author] Changki LEE(3hit)

1-3hit
  • LSTM-CRF Models for Named Entity Recognition

    Changki LEE  

     
    PAPER-Natural Language Processing

      Pubricized:
    2017/01/20
      Vol:
    E100-D No:4
      Page(s):
    882-887

    Recurrent neural networks (RNNs) are a powerful model for sequential data. RNNs that use long short-term memory (LSTM) cells have proven effective in handwriting recognition, language modeling, speech recognition, and language comprehension tasks. In this study, we propose LSTM conditional random fields (LSTM-CRF); it is an LSTM-based RNN model that uses output-label dependencies with transition features and a CRF-like sequence-level objective function. We also propose variations to the LSTM-CRF model using a gate recurrent unit (GRU) and structurally constrained recurrent network (SCRN). Empirical results reveal that our proposed models attain state-of-the-art performance for named entity recognition.

  • Extracting Events from Web Documents for Social Media Monitoring Using Structured SVM

    Yoonjae CHOI  Pum-Mo RYU  Hyunki KIM  Changki LEE  

     
    LETTER-Natural Language Processing

      Vol:
    E96-D No:6
      Page(s):
    1410-1414

    Event extraction is vital to social media monitoring and social event prediction. In this paper, we propose a method for social event extraction from web documents by identifying binary relations between named entities. There have been many studies on relation extraction, but their aims were mostly academic. For practical application, we try to identify 130 relation types that comprise 31 predefined event types, which address business and public issues. We use structured Support Vector Machine, the state of the art classifier to capture relations. We apply our method on news, blogs and tweets collected from the Internet and discuss the results.

  • Pegasos Algorithm for One-Class Support Vector Machine

    Changki LEE  

     
    LETTER-Artificial Intelligence, Data Mining

      Vol:
    E96-D No:5
      Page(s):
    1223-1226

    Training one-class support vector machines (one-class SVMs) involves solving a quadratic programming (QP) problem. By increasing the number of training samples, solving this QP problem becomes intractable. In this paper, we describe a modified Pegasos algorithm for fast training of one-class SVMs. We show that this algorithm is much faster than the standard one-class SVM without loss of performance in the case of linear kernel.

FlyerIEICE has prepared a flyer regarding multilingual services. Please use the one in your native language.