Random indexing

Random indexing is a dimension reduction method and computational framework for Distributional semantics, based on the insight that very-high-dimensional Vector Space Model implementations are impractical, that models need not grow in dimensionality when new items (e.g. new terminology) is encountered, and that a high-dimensional model can be projected into a space of lower dimensionality without compromising L2 distance metrics if the resulting dimensions are chosen appropriately, which is the original point of the random projection approach to dimension reduction first formulated as the Johnson–Lindenstrauss lemma. Locality-sensitive hashing has some of the same starting points. Random indexing, as used in representation of language, originates from the work of Pentti Kanerva[1] on Sparse distributed memory, and can be described as an incremental formulation of a random projection.[2]

It can be also verified that random indexing is a random projection technique for the construction of Euclidean spaces---i.e. L2 normed vector spaces.[3] In Euclidean spaces, random projections are elucidated using the Johnson–Lindenstrauss lemma.[4]

TopSig [5] extends the Random Indexing model to produce bit vectors for comparison with the Hamming distance similarity function. It is used for improving the performance of information retrieval and document clustering. In a similar line of research, Random Manhattan Integer Indexing[6] is proposed for improving the performance of the methods that employ the Manhattan distance between text units. Many random indexing methods primarily generate similarity from co-occurrence of items in a corpus. Reflexive Random Indexing [7] generates similarity from co-occurrence and from shared occurrence with other items.

References

  1. Joshi, Aditya, Johan Halseth, and Pentti Kanerva. "Language Recognition using Random Indexing." arXiv preprint arXiv:1412.7026 (2014).
  2. Recchia, Gabriel, et al. "Encoding sequential information in vector space models of semantics: Comparing holographic reduced representation and random permutation." (2010): 865-870.
  3. QasemiZadeh, B. & Handschuh, S. (2014) Random Manhattan Indexing, In Proceedings of the 25th International Workshop on Database and Expert Systems Applications.
  4. Johnson, W. and Lindenstrauss, J. (1984) Extensions of Lipschitz mappings into a Hilbert space, in Contemporary Mathematics. American Mathematical Society, vol. 26, pp. 189–206.
  5. Geva, S. & De Vries, C.M. (2011) TopSig: Topology Preserving Document Signatures, In Proceedings of Conference on Information and Knowledge Management 2011, 24-28 October 2011, Glasgow, Scotland.
  6. QasemiZadeh, B. & Handschuh, S. (2014) , In Proceedings of EMNLP'14.
  7. Cohen T., Schvaneveldt R. & Widdows D. (2009) Reflective Random Indexing and indirect inference: a scalable method for discovery of implicit connections, Journal of Biomedical Informatics, 43(2):240-56.


This article is issued from Wikipedia - version of the 11/9/2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.