SMART Information Retrieval System

The SMART (System for the Mechanical Analysis and Retrieval of Text) Information Retrieval System is an information retrieval system developed at Cornell University in the 1960s.[1] Many important concepts in information retrieval were developed as part of research on the SMART system, including the vector space model, relevance feedback, and Rocchio classification.

Gerard Salton led the group that developed SMART. Other contributors included Mike Lesk.

The SMART system also provides a set of corpora, queries and reference rankings, taken from different subjects, notably

  • ADI: publications from information science reviews
  • Computer science
  • Cranfield collection: publications from aeronautic reviews
  • Forensic science: library science
  • MEDLARS collection: publications from medical reviews
  • Time magazine collection: archives of the generalist review Time in 1963

To the legacy of the SMART system belongs the so-called SMART triple notation, a mnemonic scheme for denoting tf-idf weighting variants in the vector space model. The mnemonic for representing a combination of weights takes the form ddd.qqq, where the first three letters represents the term weighting of the collection document vector and the second three letters represents the term weighting for the query document vector. For example, ltc.lnn represents the ltc weighting applied to a collection document and the lnn weighting applied to a query document.

The following tables establish the SMART notation:[2]

Symbols and notation
D i = { w i 1 , w i 2 , , w i t } {\textstyle D_{i}=\{w_{i_{1}},w_{i_{2}},\ldots ,w_{i_{t}}\}} represents a document vector, where w i k {\textstyle w_{i_{k}}} is the weight of the term T k {\textstyle T_{k}} in D i {\textstyle D_{i}} and t {\displaystyle t} is the number of unique terms in D i {\textstyle D_{i}} . Positive features characterize terms that are present in a document, and the weight of zero is used for terms that are absent from a document.
f i k {\textstyle f_{i_{k}}} Occurrence frequency of term T k {\textstyle T_{k}} in document D i {\textstyle D_{i}} u i {\textstyle u_{i}} Number of unique terms in document D i {\textstyle D_{i}}
N {\displaystyle N} Number of collection documents avg ( u ) {\displaystyle \operatorname {avg} (u)} Average number of unique terms in a document
n k {\textstyle n_{k}} Number of documents with term T k {\textstyle T_{k}} present b t {\displaystyle b_{t}} Number of characters in document D i {\displaystyle D_{i}}
max ( f i k ) {\displaystyle \max(f_{i_{k}})} Occurrence frequency of the most common term in document D i {\displaystyle D_{i}} avg ( b ) {\textstyle \operatorname {avg} (b)} Average number of characters in a document
avg ( f i k ) {\displaystyle \operatorname {avg} (f_{i_{k}})} Average occurrence frequency of a term in document D i {\displaystyle D_{i}} G {\textstyle G} Global collection statistics
s {\displaystyle s} The slope in the context of pivoted document length normalization[3]
Smart term-weighting triple notation
Term frequency tf ( f i k ) {\textstyle {\text{tf}}(f_{i_{k}})} Document frequency df ( N , n k ) {\textstyle {\text{df}}(N,n_{k})} Document length normalization g ( G , D i ) {\textstyle g(G,D_{i})}
b 1 {\textstyle 1} Binary weight x n 1 {\textstyle 1} Disregards the collection frequency x n 1 {\textstyle 1} No document length normalization
t n f i k {\textstyle f_{i_{k}}} Raw term frequency f log 2 ( N n k ) {\displaystyle \log _{2}\left({\frac {N}{n_{k}}}\right)} Inverse collection frequency c k = 1 t w i k 2 {\displaystyle {\sqrt {\sum _{k=1}^{t}w_{i_{k}}^{2}}}} Cosine normalization
a 0.5 + 0.5 f i k max ( f i k ) {\textstyle 0.5+0.5{\frac {f_{i_{k}}}{\max(f_{i_{k}})}}} Augmented normalized term frequency t log 2 ( N + 1 n k ) {\displaystyle \log _{2}\left({\frac {N+1}{n_{k}}}\right)} Inverse collection frequency u 1 s + s u i avg ( u ) {\displaystyle 1-s+s{\frac {u_{i}}{\operatorname {avg} (u)}}} Pivoted unique normalization[3]
l 1 + log 2 f i k {\displaystyle 1+\log _{2}f_{i_{k}}} Logarithm p log 2 ( N n k n k ) {\displaystyle \log _{2}\left({\frac {N-n_{k}}{n_{k}}}\right)} Probabilistic inverse collection frequency b 1 s + s b i avg ( b ) {\displaystyle 1-s+s{\frac {b_{i}}{\operatorname {avg} (b)}}} Pivoted characted length normalization[3]
L 1 + log 2 ( f i k ) 1 + log 2 ( avg ( f i k ) ) {\displaystyle {\frac {1+\log _{2}(f_{i_{k}})}{1+\log _{2}(\operatorname {avg} (f_{i_{k}}))}}} Average-term-frequency-based normalization[3]
d 1 + log 2 ( 1 + log 2 ( f i k ) ) {\displaystyle 1+\log _{2}(1+\log _{2}(f_{i_{k}}))} Double logarithm

The gray letters in the first, fifth, and ninth columns are the scheme used by Salton and Buckley in their 1988 paper.[4] The bold letters in the second, sixth, and tenth columns are the scheme used in experiments reported thereafter.

References

  1. ^ Salton, G, Lesk, M.E. (June 1965). "The SMART automatic document retrieval systems—an illustration". Communications of the ACM. 8 (6): 391–398. doi:10.1145/364955.364990.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  2. ^ Palchowdhury, Sauparna (2016). "On The Provenance of tf-idf". sauparna.sdf.org. Retrieved 2019-07-29.
  3. ^ a b c d Singhal, A., Buckley, C., & Mitra, M. (1996). Pivoted Document Length Normalization. SIGIR Forum, 51, 176-184.
  4. ^ Salton, G., & Buckley, C. (1988). Term-Weighting Approaches in Automatic Text Retrieval. Inf. Process. Manage., 24, 513-523.

External links

  • Software and test collections[dead link] (FTP at Cornell University)
  • Interactive SMART tutorial


  • v
  • t
  • e