NLP领域的任务的输入输出有以下几种:
输入 | 输出 | 示例 | 备注 |
---|---|---|---|
Seq | label | ||
Seq1 | Seq2 | 所有Seq2Seq问题,如翻译、Chatbox、序列生成等 | |
Seq1 + Seq2 | label | Pairwise类问题,比如判断2个Seq的关系或相似度、Chatbox等 |
除7.2外,其他模型结构都是或类似于双胞胎网络(Siamese Network),2个网络的结构是完全一致的,但其参数,有时共享,有时不同?
研究的模型有:InferSent, SSE, DecAtt, ESIM, PWIM
Article: 基于神经网络模型的释义识别、语义文本相似性、自然语言推理和问题回答
https://github.com/brightmart/nlu_sim (Tensorflow)
all kinds of baseline models for modeling tasks with pair of sentences: semantic text similarity(STS), natural language inference(NLI), paraphrase identification(PI), question answering(QA)
模型有:DualTextCNN, DualBiLSTM, DualBiLSTMCNN, ESIM, SSE, BiLSTM with Attention
Structure: Input(Seq EOS Seq) -> Embeddding -> BiLSTM -> Average -> Softmax
Same with TextRNN, but input is special designed.
e.g. input: "How much is the computer ? EOS Price of laptop", where 'EOS' is a special token splitted input1 and input2
Structure: (Input1 -> Embedding -> TextCNN) * 2 -> Concatenate -> Softmax
产品词关系项目中使用的模型与此类似,在此基础上增加了第3个Input(结构化输入)。
又叫 Dual Encoder LSTM Network ?
Structure: Seq1(Input1 -> Embedding -> BiLSTM) + Seq2(Input2 -> Embedding -> BiLSTM) -> Dot Product -> Softmax
Dot Product作用:To measure the similarity of the predicted response r' and the actual response r by taking the dot product of these two vectors. A large dot product means the vectors are similar and that the response should receive a high score. We then apply a sigmoid function to convert this score into a probability. Similarity --> Probability
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。