Context Prediction Using Neural Networks

Authors

  • Rohith Manda SRM Institute of Science and Technology Author

DOI:

https://doi.org/10.61841/gskqsp10

Keywords:

Prediction, Ne.ural Networks

Abstract

Word-prediction can be thought of as predicting a word, given its context. It is similar to predictive text used by multiple devices. The concept of word-embeddings is ideal for this type of a problem. Embeddings are an alternate form of word representation, such that it preserves the syntactic and semantic information. It extends the functionality to extract or study the relationships between various words, given in a corpus. The objective can be viewed as, (i) generation of quality embeddings,(ii) using neural networks to realise a model, capable of predicting words accurately, given the context as input. The generation of embeddings are a tedious task and requires a good amount of computation. To use these in an environment that is supported by machine learning algorithms requires transformation of these embeddings and corpus into vector space, i.e, a numeric form of representation. Word prediction can be of many types, depending on the model. For example, if the objective is to predict the (n+1)th word, given the previous n words; The automaton can consider only the previous n or (n,n-1) or (n,n-1,n-2…) words for predicting the next word, the earlier two are referred as bigram and trigram model respectively.

 

Downloads

Download data is not yet available.

References

1. Y. Bengio, R. Ducharme, P. Vincent. A neural probabilistic language model. Journal of Machine Learning Research, 3:1137-1155, 2003.

2. Y. Bengio, Y. LeCun. Scaling learning algorithms towards AI. In: Large-Scale Kernel Machines, MIT Press, 2007.

3. Sebastian Sudholt, Gernot A. Fink “Evaluating Word String Embeddings and Loss Functions for NN- Based Word Spotting”. IAPR International Conference on Document Analysis and Recognition (ICDAR) 2017.

4. Wen-tau Yih, Geoffrey Zweig “Linguistic Regularities in Continuous Space Word Representations”.

North American Chapter of the Association for Computational Linguistics(NAACL) 2014.

5. Zhila, W.T. Yih, C. Meek, G. Zweig, T. Mikolov. Combining Heterogeneous Models for Measuring Relational Similarity. NAACL HLT 2013.

6. Tomas Mikolov, Kai Chen, Greg Corrado, Jeffrey Dean” Efficient Estimation of Word Representations in Vector Space”. The International Conference on Learning Representations (ICLR) 2013.

7. D.A. Jurgens, S.M. Mohammad, P.D. Turney, K.J. Holyoak. Semeval-2012 task 2: Measuring degrees of relational similarity. In: Proceedings of the 6th International Workshop on Semantic Evaluation (SemEval 2012), 2012.

8. A.L. Maas, R.E. Daly, P.T. Pham, D. Huang, A.Y. Ng, and C. Potts. Learning word vectors for sentiment analysis. In Proceedings of ACL, 2011.

9. T. Mikolov, A. Deoras, D. Povey, L. Burget, J. Cernock ˇ y. Strategies for Training Large Scale ´ Neural Network Language Models, In: Proc. Automatic Speech Recognition and Understanding, 2011.

10. T. Mikolov, W.T. Yih, G. Zweig. Linguistic Regularities in Continuous Space Word Representations.

NAACL HLT 2013. H. Schwenk. Continuous space language models. Computer Speech and Language, vol. 21, 2007.

Downloads

Published

30.06.2020

How to Cite

Manda, R. (2020). Context Prediction Using Neural Networks. International Journal of Psychosocial Rehabilitation, 24(6), 4151-4159. https://doi.org/10.61841/gskqsp10