lecture 6

作者: CCCCCCC7 | 来源:发表于2018-02-22 04:00 被阅读0次

Continue from lecture 5: tagging as classification, combining searching and learning

Transition probability: count(t_{i}, t_{i-1}) / count(t_{i-1})
Emission probability: count(w_i, t_i) / count(t_i)
-code: Viterbi-tagger.ipynb
-those probabilities often are very small numbers.
Reason of dropping words: comparing long sentences with short ones, longer sentence has too small numbers related to short ones

Conclusion

-Cons: word with different meanings under the same tag, local information doesn't help:(plant)
All the model been talked so far only use local informations. But there are long term ambiguity.
Pipe line to handle different kinds of ambiguities.

Continue with the problem talked last time: when maximizing the likelihood...

Observe some words and find tags that maximizes the likelihood.
Since we're finding the sequence that maximizing the likelihood, the sequence at any time must have the maximum likelihood till this specific time(only one path from the start point till now), a DP problem.

A small code example on wikipedia of Viterbi algorithm.

Complexity Analysis

-num of words:
-num of tags:
Sometimes changing tag set helps.

Trigram Tagging

Similar with trigram with language models on the transition probability side, emission probability stays the same.
This is a tutorial about tagging with Markov models.

Cons: Couldn't generalize well, sparsity makes it easy to blow up with unseen transitions.

Review -- Search

Guidance: give guidance, algorithm will improve. e.g: in the example(path planing) mentioned in class: guide = cost + heuristic guise
A* search wiki.

Analogy to Sequence Tagging

Transform probability into neg-log-prob.
Heuristic is a lower bound .... not clear....
Compare A* , Viterbi and beam search.

相关文章

  • Lecture 6

    ![459K$J631L9@43322)RQ]51.png](https://upload-images.jian...

  • lecture 6

    Continue from lecture 5: tagging as classification, combi...

  • Lecture 6

    第三章 内部存储器 3.1 存储器概述 3.1.1 存储器的分类 3.1.2 存储器的分级 3.1.3 存储时间 ...

  • Lecture 6 Growth

    有了好的产品、团队后,唯一要考虑的事情就是增长了。 那增长最关键的因素是什么?留存率,这是唯一的衡量标准。如果一个...

  • ISL视频课程学习笔记

    视频列表 Lecture 1.1 Lecture 1.2 Lecture 2.1 Lecture 2.2 Lect...

  • Princeton Alogorithm COS226 Week

    Overview this lecture is the most interesting lecture in ...

  • Lecture 5-6

    [toc] Lecture 5-6 P2.What is complexity analysis? 什么是算法复杂...

  • 雅思阅读笔记(第三节)

    curriculum 课程大纲: Lecture 1 总纲、单词题 Lecture 2 句子题 Lecture 3...

  • 雅思阅读笔记(第一节)

    curriculum 课程大纲: Lecture 1 总纲、单词题 Lecture 2 句子题 Lecture 3...

  • 雅思阅读笔记(第二节)

    curriculum 课程大纲: Lecture 1 总纲、单词题 Lecture 2 句子题 Lecture 3...

网友评论

      本文标题:lecture 6

      本文链接:https://www.haomeiwen.com/subject/adratftx.html