Natural language processing: Difference between revisions
No edit summary |
|||
(6 intermediate revisions by the same user not shown) | |||
Line 4: | Line 4: | ||
==Classical NLP== | ==Classical NLP== | ||
The Classical NLP consists of creating a pipeline using processors to create | The Classical NLP consists of creating a pipeline using processors to create annotations from text files.<br> | ||
Below is an example of a few processors.<br> | Below is an example of a few processors.<br> | ||
* Tokenization | * Tokenization | ||
Line 10: | Line 10: | ||
* Part-of-speech annotation | * Part-of-speech annotation | ||
* Named Entity Recognition | * Named Entity Recognition | ||
==Machine Learning== | ==Machine Learning== | ||
===Datasets and Challenges=== | |||
====SQuAD==== | ====SQuAD==== | ||
[https://rajpurkar.github.io/SQuAD-explorer/ Link]<br> | [https://rajpurkar.github.io/SQuAD-explorer/ Link]<br> | ||
The Stanford Question Answering Dataset. There are two | The Stanford Question Answering Dataset. There are two versions of this dataset, 1.1 and 2.0. | ||
===Transformer=== | ===Transformer=== | ||
[https://arxiv.org/abs/1706.03762 Attention is all you need paper] | {{ main | Transformer (machine learning model)}} | ||
A neural network architecture by Google. | [https://arxiv.org/abs/1706.03762 Attention is all you need paper] | ||
It is currently the best at NLP tasks and has mostly replaced RNNs for these tasks. | |||
A neural network architecture by Google which uses encoder-decoder attention and self-attention. | |||
It is currently the best at NLP tasks and has mostly replaced RNNs for these tasks. | |||
However, it's computational complexity is quadratic in the number of input and output tokens due to attention. | |||
;Guides and explanations | |||
* [https://nlp.seas.harvard.edu/2018/04/03/attention.html The Annotated Transformer] | |||
* [https://www.youtube.com/watch?v=iDulhoQ2pro Youtube Video] | |||
===Google Bert=== | ===Google Bert=== | ||
{{main| BERT (language model)}} | |||
[https://github.com/google-research/bert Github Link] | [https://github.com/google-research/bert Github Link] | ||
[https://arxiv.org/abs/1810.04805 Paper] | [https://arxiv.org/abs/1810.04805 Paper] | ||
Line 28: | Line 37: | ||
A pretrained NLP neural network. | A pretrained NLP neural network. | ||
Note the code is written in TensorFlow 1. | Note the code is written in TensorFlow 1. | ||
====Albert==== | |||
[https://github.com/google-research/google-research/tree/master/albert Github]<br> | |||
;A Lite BERT for Self-supervised Learning of Language Representations | |||
This is a parameter reduction on Bert. | |||
==Libraries== | ==Libraries== | ||
===Apache OpenNLP=== | ===Apache OpenNLP=== | ||
[https://opennlp.apache.org/ Link] | [https://opennlp.apache.org/ Link] |