Zach Solan, Eytan Ruppin, David Horn, Shimon Edelman
The distributional principle according to which morphemes that occur in identical contexts belong, in some sense, to the same category  has been advanced as a means for extracting syntactic structures from corpus data. We extend this principle by applying it recursively, and by us- ing mutual information for estimating category coherence. The resulting model learns, in an unsupervised fashion, highly structured, distributed representations of syntactic knowledge from corpora. It also exhibits promising behavior in tasks usually thought to require representations anchored in a grammar, such as systematicity.