syntactic category learning as iterative prototype-driven

12
Proceedings of the Society for Computation in Linguistics Volume 1 Article 6 2018 Syntactic Category Learning as Iterative Prototype- Driven Clustering Jordan Kodner University of Pennsylvania, [email protected] Follow this and additional works at: hps://scholarworks.umass.edu/scil Part of the Computational Linguistics Commons is Paper is brought to you for free and open access by ScholarWorks@UMass Amherst. It has been accepted for inclusion in Proceedings of the Society for Computation in Linguistics by an authorized editor of ScholarWorks@UMass Amherst. For more information, please contact [email protected]. Recommended Citation Kodner, Jordan (2018) "Syntactic Category Learning as Iterative Prototype-Driven Clustering," Proceedings of the Society for Computation in Linguistics: Vol. 1 , Article 6. DOI: hps://doi.org/10.7275/R5TQ5ZQ4 Available at: hps://scholarworks.umass.edu/scil/vol1/iss1/6

Upload: others

Post on 06-Nov-2021

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Syntactic Category Learning as Iterative Prototype-Driven

Proceedings of the Society for Computation in Linguistics

Volume 1 Article 6

2018

Syntactic Category Learning as Iterative Prototype-Driven ClusteringJordan KodnerUniversity of Pennsylvania, [email protected]

Follow this and additional works at: https://scholarworks.umass.edu/scil

Part of the Computational Linguistics Commons

This Paper is brought to you for free and open access by ScholarWorks@UMass Amherst. It has been accepted for inclusion in Proceedings of theSociety for Computation in Linguistics by an authorized editor of ScholarWorks@UMass Amherst. For more information, please [email protected].

Recommended CitationKodner, Jordan (2018) "Syntactic Category Learning as Iterative Prototype-Driven Clustering," Proceedings of the Society forComputation in Linguistics: Vol. 1 , Article 6.DOI: https://doi.org/10.7275/R5TQ5ZQ4Available at: https://scholarworks.umass.edu/scil/vol1/iss1/6

Page 2: Syntactic Category Learning as Iterative Prototype-Driven

Syntactic Category Learning as Iterative Prototype-Driven Clustering

Jordan KodnerUniversity of PennsylvaniaDepartment of Linguistics

Department of Computer and Information [email protected]

Abstract

We lay out a model for minimally supervisedsyntactic category acquisition which com-bines psychologically plausible concepts fromstandard NLP part-of-speech tagging applica-tions with simple cognitively motivated dis-tributional statistics. The model assumes asmall set of seed words (Haghighi and Klein,2006), an approach with motivation in (Pinker,1984)’s semantic bootstrapping hypothesis,and repeatedly constructs hierarchical ag-glomerative clusterings over a growing lexi-con. Clustering is performed on the basis ofword-adjacent syntactic frames alone (Mintz,2003) with no reference to word-internal fea-tures, which has been shown to yield qualita-tively coherent POS clusters (Redington et al.,1998). A prototype-driven labeling processbased on tree-distance yields results compara-ble to unsupervised algorithms based on com-plex statistical optimization while maintainingits cognitive underpinnings.

1 Introduction

Supervised part-of-speech (POS) tagging is one ofstatistical NLP’s classic problems (Meteer et al.,1991), but its reliance on large POS-annotated cor-pora makes it impractical to extend to new domainsand unsuitable for low-resource languages with-out pre-existing annotation. Unsupervised taggingpresents an interesting alternative because it does notrequire labelled training data, but it is a more dif-ficult problem, and typical performance is substan-tially lower. The fundamental problem of trainingwithout examples aside, unsupervised tagging algo-rithms induce some number of clusters which must

be mapped onto a desired tag set. This mapping neednot be one-to-one, so error may be introduced goingfrom algorithm output to final results.

Within NLP, unsupervised POS tagging is typi-cally approached as a statistical optimization prob-lem, though implementations vary widely. Brownet al. (1992) and Clark (2003) induce clusters viaclass-based n-gram models, the latter including mor-phological information. While these perform rea-sonably well, more recent approaches have insteadbeen centered around HMMs. Goldwater and Grif-fiths (2007) and Johnson (2007) both place Dirichletpriors over the multinomial parameters of roughlytypical HMM POS taggers. Berg-Kirkpatrick et al.(2010) define a feature-based HMM instead whichallows the inclusion of discriminative orthographicinformation.1 Haghighi and Klein (2006) implementa model based on Markov random fields (MRFs), anundirected generalization of HMMs.

Haghighi & Klein (H&K) also differs from theabove models in that it is minimally supervised withprototypes. In their implementation, three wordsper tag are defined as prototypes or seeds and la-belled with their correct tags, and the MRF sorts outwhich prototypes unknown words are most similarto. They choose the most frequent words per tagimposing the requirement that each seed must onlysupport a single tag, so given the 45-tag Penn Tree-bank tag set, this yields 135 seeds, a tiny fractionof the tens of thousands of types in the Wall StreetJournal corpus. Their MRF is trained on both distri-butional features and orthographic features follow-

1See Christodoulopoulos et al. (2010) for a fuller summaryof such models.

44Proceedings of the Society for Computation in Linguistics (SCiL) 2018, pages 44-54.

Salt Lake City, Utah, January 4-7, 2018

Page 3: Syntactic Category Learning as Iterative Prototype-Driven

ing Smith and Eisner (2005). This achieves impres-sive results on English, but their reliance on ortho-graphic features curtails the model’s performance onChinese.

H&K’s and the other computational models aboveare engineered primarily with performance in mindrather than cognitive plausibility. The complex opti-mization models in particular are slow, taking tens ofhours to complete, while the simpler n-gram basedmodels run in tens of minutes (Christodoulopouloset al., 2010). But fundamentally, they are all toolswhich take advantage of statistical, especially distri-butional information.

For well over half a century now, it has beenunderstood that children make use of distributionalcues in the process of syntactic category (roughlyPOS) assignment as well. For example, Brown(1957)’s classic study found that children recognizea nonce word “sib” as a noun when if it is intro-duced to them in a sentence like “This is a sib,” butprefer to label it as a verb if it is introduced in “Itis sibbing.” Along similar lines, Shi and Melancon(2010) present nonce words to year-old children andwatch whether they then tend to focus on an im-age depicting an object or an action. They find thatchildren presented with “the mige” prefer to look atthe object image over the action image, consistentwith them understanding that the word should be anoun. These experiments demonstrate that learnersmake use of distributional information (e.g., “a/the

”) even with limited exposure. Studies of child-directed corpora suggest the presence of distribu-tional cues as well in a more naturalistic setting(Maratsos, 1979; Redington et al., 1998).

Children’s sensitivity to such distributional infor-mation has been operationalized through the notionof frequent frames, single-word contexts on eitherside of an item (Mintz, 2003). Experimental evi-dence demonstrates that children who are exposed toitems within the same frame, (e.g, “the is”) treatthose items as members of the same class. How-ever, the large number classes induced from frequentframes do not provide a clean one-to-one mappingto syntactic categories (Chemla et al., 2009). Syn-tactic frames can be seen as a purely structural cue,but semantic information play a role as well. Asdescribed by the semantic bootstrapping hypothesis(Pinker, 1984), children have innate rules for map-

ping real-world semantic onto syntactic categories.For example, actions should be verbs, objects shouldbe nouns, and so on. These then serve as anchorsin the input to provide early distributional context.The validity of semantic bootstrapping’s claims ofinnateness and the exact nature and number of syn-tactic classes are not critical for the present work,rather it is sufficient that bootstrapping guides chil-dren to some kind of categorization. Experimentalwork has long confirmed that semantic bootstrap-ping basic prediction holds and that children reallydo associate actions and concrete objects with verbsand nouns (Gleitman et al., 2005; Pinker, 1984; Ron-dal et al., 1987).

We develop a computational implementation forsemantic bootstrapping in syntactic frames whichdraws inspiration from hierarchical clustering andfrom Haghighi and Klein (2006)’s prototype-drivenmodel for POS tagging. Since we know that chil-dren do not wait patiently to build up large vocabu-laries replete with distributional information beforeattempting to assign syntactic categories, our algo-rithm runs iteratively on the lexicon as it grows,revising category assignments as more evidencecomes in. Additionally, we discard all word-internalmorphological and phonological information to testthe distributional cues on their own. The rest of thispaper is organized as follows: section 2 introducesthe iterative prototype-driven clustering model fortagging along with the basic insights behind it. Sec-tion 3 discusses the problem of evaluation for un-supervised POS tagging, provides results for child-directed English under various conditions, compara-tive results across nine other languages, and a com-parison with H&K on English and Chinese. Section4 reviews the model’s cognitive plausibility and dis-cusses possible extensions to the algorithm.

2 The Model

The primary insight for this work comes from theobservation that words may be grouped into roughpart-of-speech clusters on the basis of their single-word contexts. This is an implicit assumption ofn-gram tagging models (Clark, 2003; Brown et al.,1992) for POS tagging, and the feasibility of distri-butional contexts for distinguishing between syntac-tic categories has been explicitly studied in cognitive

45

Page 4: Syntactic Category Learning as Iterative Prototype-Driven

Figure 1: Example of the cutting problem from Parkes et al.,

(1998) with additional markup for readability

literature as well (Redington et al., 1998).Parkes et al. (1998) provide a straightforward

implementation for simple agglomerative clusteringwithout orthographic input that serves to highlightthe benefits and drawbacks of the approach. Foreach of the 400 most frequent words in the WallStreet Journal corpus, Parkes et al. tabulate the fre-quencies of left and right-context types. Using sym-metricized KL-divergence (KL(p||q)+KL(q||p)) asa distance metric, they perform agglomerative clus-tering on those 400 types and note the qualitative pu-rity of the resulting clusters. The immediately obvi-ous problem is that there is no perfect mapping fromclusters to tags. Figure 1 demonstrates this. Here,there are two pure VBD (past tense verb) clusters, butthey cannot be joined without including a cluster ofVBZ (present tense verb). Any simple cutting algo-rithm fails by either incorrectly postulating two VBD

classes (VBD1, VBD2, VBZ), or postulating a mixedVBD/VBZ class. This is a general problem, indepen-dent of the choice of tag set, so while the WSJ VBD

and VBZ should probably be collapsed into a singleV class from a cognitive point of view, that alonedoes not solve it.

The simple counts over single-word contextswhich Parkes et al. (1998) employ are equivalent tothe syntactic frames described in the acquisition lit-erature. So while the Parkes et al. study is neitherpresented as nor performs as a tagging algorithm, itholds promise from a cognitive perspective and pro-

vides clusters similar to those in the cognitive liter-ature. The next two sections describe a prototype-driven labeling algorithm built on top of these clus-ters. The algorithm is described in two parts, firstthe inner loop basic algorithm which operates on avocabulary of fixed size, then the outer loop, an iter-ative extension that operates on a growing lexicon.

2.1 Prototype-Driven labeling

Prototype-driven labeling operates over the treescreated by hierarchical clustering and makes up theinner loop of the category labeling algorithm. Twovectors, one each of immediate left and right con-text type counts, are tabulated for the top k most fre-quent words in a corpus. Next, these k types aregrouped by agglomerative clustering. This is slow(O(k2 log(k)) in a naıve implementation, but thatcan be mitigated with incremental clustering or otherapproaches. KL-divergence over the concatenatedleft and right context vectors serves as the distancemetric. KL is not symmetric, so for each pair of vec-tors v and w, the sum of the KL-divergence betweenv and w and w and v is used as in Equation 1, wherea and b are two words, and v and w are their corre-sponding context vectors.2

d(a, b) = KL(v||w) + KL(w||v) (1)

=∑

i

vi logviwi

+ wi logwi

vi

In agglomerative clustering, it is necessary to de-fine a linkage criterion that allows the constructedsubtrees to be compared to each other along withindividual words. The distance between the sub-trees here is taken as the average distances betweentheir members. This only performs slightly betterthan taking the minimum distance between any twomembers of the subtrees, and it can be updated onthe fly as new members are added to the tree. Equa-tion 2 gives a formal description of the average dis-tance criterion. Clustering greedily joins whateverpair of words or subtrees has the smallest distancebetween them at each step. As a result, the final tree

2This is symmetric and is equivalent to Jenson-Shannon-divergence for the purpose of rank ordering, but it is marginallyeasier to compute.

46

Page 5: Syntactic Category Learning as Iterative Prototype-Driven

can be thought of as a rank ordering of word similar-ities. There is no need to track the actual vector val-ues of individual leaves once they have been joinedinto subtrees.

D(A,B) =1

|A||B|∑

a∈A

b∈Bd(a, b) (2)

Before beginning the category labeling process,seed words are labelled in the tree to simulate se-mantic bootstrapping. This can be done by pick-ing the three most frequent words per actual labelas in H&K, or seeds can be chosen by contextualsaliency. For a set of n tags, there is a maximumof 3n seeds in the tree. But if k is small, the ac-tual number will usually be smaller. It is easy to seewhy when looking at the WSJ corpus and tag set, butthe same is true of smaller cognitively-motivated tagsets as well. The three seeds with the FW “foreignword” tag are perestroika, de, and kanji.3 None ofthese is in the top thousand most frequent words, sofor k = 1000, there must be fewer than 3n seeds.

Once the seeds are assigned, the labeling can pro-ceed for the remaining words. This can be equiva-lently implemented during the join steps of agglom-erative clustering itself or on the completed agglom-erative tree. Initially, the leaves of the tree are treatedas labelled if they are seeds and unlabelled other-wise. Iterating over joins in the order that they oc-cur from the leaves upward, if an unlabelled subtreeis joined with a labelled one, each of its leaves isassigned the most common label from the labelledsubtree. This is easiest to explain by example. Inthe simplest case, a seed is joined with an unlabelledword, and that word is assigned the same label asthe seed (Figure 2 (left)). In the general case, whena labelled subtree (containing ≥ 1 seed) is joinedwith an unlabelled subtree, every word in the unla-belled subtree is assigned the most common labelfrom the labelled subtree. For example, if a subtreethat is 90% nouns and 10% verbs is joined with anunlabelled tree, every word in the unlabelled tree istagged as a noun (Figure 2 (right)). This approachprovides a solution to the cutting problem: it is notan issue if multiple clusters map to the same part-of-speech because what matters is the clusters’ proxim-

3The frequent FW words effectively date this corpus.

Figure 2: Visual depiction of assignment. (Left) leaves. (Right)

subtrees.

ity to the seeds, not to each other. Referring back tothe Parkes et al. example in Figure 1, the model willlabel the tree perfectly as long as each small clusteris centered around a seed.

2.2 Iterative Prototype-Driven labeling

The basic prototype-driven labeling algorithm oper-ates on a hierarchical clustering of fixed size k. Thelabeling step itself is fast, but for large k, comput-ing the pairwise distance function becomes becomesslow. It also runs afoul of experimental evidence oncategory learning. Children do not wait patientlyto build a large vocabulary complete with distribu-tional information before even attempting to assignsyntactic categories. Rather, they assign what theycan when they can, potentially revising their assign-ments as they go along (Pinker, 1984; Gleitman etal., 2005).

An iterative extension to the labeling algorithmsolves both the practical and empirical problems.Instead of running once on a fixed large k, thealgorithm is run multiple times on a sequence oflexicons with monotonically increasing sizes K =(k0, k1, . . . , km) where each lexicon contains the topki most frequent types in the corpus of study. Thisis meant to approximate which words a learner isstatistically most likely to have heard early in devel-opment. The algorithm begins by labeling a smallki, which is always quick to compute. Then it hasworking hypotheses for the first ki words while itre-evaluates those words and learns new words upto ki+1. While not purely online (Hewitt, 2017),this provides the sort of incremental advancementseen in the literature and is amenable to online im-plementation (Guedalia et al., 1999). Additionally,

47

Page 6: Syntactic Category Learning as Iterative Prototype-Driven

once a partial tree has been built, it becomes unnec-essary to compute pairwise distance between everynew vocabulary item and existing item. Once it isdiscovered which seeds are close to a new word, itis sufficient to only compute distances between thenew words and those words which are close to itsnearest seeds to achieve a kind of clustering. Thisdovetails nicely with some efficient agglomerativeclustering implementations, e.g., (McCallum et al.,2000)’s canopies.

The algorithm attempts classification on frequentwords multiple times, so it has the opportunity to useevidence from early classification to inform later at-tempts before picking its best guess in the end. Toaccomplish this, a confidence value is set for eachsubtree assignment operation. This is defined asthe purity of the assigning tree, so a subtree that is100% adjectives assigns members of an unlabelledsubtree with a confidence of 1.0, while a 40% ad-jective, 30% noun, 30% verb subtree only assignswith a confidence of 0.4. Intuitively, a pure sub-tree is likely to represent an actual category cluster,while an impure tree is either a higher-level cluster-ing of category clusters or garbage. Then at the endof each iteration, all words assigned with a confi-dence above some fixed threshold are added to theseed set and become prototypes for the next itera-tion. All other words are reassigned in subsequent it-erations until they become prototypes themselves orthe experiment finishes. At the end, all words whichwere never added to the prototype set are assignedthe highest confidence label that they encounteredduring any iteration.

This iterative extension greatly improves perfor-mance because it expands the seed set as the sizeof the trees increase. This guarantees that the ra-tio of seeds to non-seeds is always relatively high,which improves assignment performance as long asthe augmented seed set remains high accuracy. Theratio of seeds to non-seeds is always high for smallk, but without this process, the ratio is prohibitivelylow at large k. Cognitively, this corresponds to thefact that children build up a lexicon of known wordsas they are available to them as evidence when cate-gorizing words learned later.

3 Evaluation and Results

In this section, the iterative prototype-driven label-ing model is evaluated on corpora from a wide rangeof languages. First, we discuss results on the EnglishBrown (Harvard) corpus within the CHILDES cor-pus of child-directed speech (MacWhinney, 2000),since this serves as an in-domain application. Per-formance is compared on a range of tag sets, seedset sizes, and seed selection procedures. Next, wetest on a range of languages from the Universal De-pendency Treebank (McDonald et al., 2013) in orderto gain insight into how language specific factors in-fluence results while keeping extra-linguistic factorsconstant to the extent possible. Finally, we run onthe WSJ portion of the Penn Treebank (Marcus etal., 1993) and on the Penn Chinese Treebank (Xueet al., 2005) in order to compare against H&K’sprototype-driven model and ground this work in thewider field.

3.1 Evaluation MetricsEvaluation for supervised POS tagging is straight-forward. Sentences from a test corpus are labelled,then for each token, the proposed label is comparedto the human annotated label. The simplest of suchmetrics is one-to-one token accuracy. On the otherhand, the problem of evaluation is more complicatedfor unsupervised algorithms. Results are broadly notcomparable across experiments because a wide ofevaluation metrics are employed.4

Token-based metrics are an excellent option whenthe task is to automatically annotate running textwith part-of-speech tags, but they have undesirabletraits when applied to syntactic category learning.If syntactic category learning is analogous to label-ing types in a dictionary or lexicon, then labelingsequences of text just obfuscates results. Type fre-quencies are not uniform across a corpus, so token-based metrics weight assignments to frequent typeshigher than assignments to infrequent types. Forexample, an algorithm which labels the pronoun“you” incorrectly will be punished more severelythan one which labels the pronoun “whomever” in-correctly simply because “you” is more commonthan “whomever” in most corpora. This is partic-ularly problematic because word frequencies follow

4See (Christodoulopoulos et al., 2010).

48

Page 7: Syntactic Category Learning as Iterative Prototype-Driven

a distribution where just a few types are hundredsof times more frequent than most others. Mislabel-ing any pronoun is hundreds of times more damag-ing than labeling almost any noun from our dailylives (like “table” or “bug”) incorrectly. Mislabel-ing any common noun is much more damaging thanmislabeling a rare noun, and noun frequencies canbe highly corpus specific. This makes it difficult togauge the relative performance of different models.

A one-to-many type accuracy is a better choicefor scoring syntactic category learning because eachitem is weighted the same for scoring. In natural lan-guage, types can potentially support multiple labels,so a one-to-many metric is needed to account forthis. For example, “bat” could be a noun or a verb,so an algorithm which classifies it as either shouldbe correct. The many-to-one accuracy used in someunsupervised models does not make sense here be-cause the model does not output clusters which needto be mapped. The algorithm never demarcates clus-ters and instead assigns labels to individual items.

The difference between token and type-basedmetrics becomes clear when calculating the base-lines for our model. Scoring by type accuracy, thebaseline is arrived at by scoring the initial seeds ascorrect and marking everything else wrong. Thistypically yields a score of under 10%, often under1%, and corresponds directly to the proportion oftypes which are selected as seeds. However, thistends to correspond to a > 50% token accuracy be-cause seeds are drawn from the most frequent types.A final type score of, say, 70% is more meaningfulthan a final token score of 70% because the improve-ment over the baseline is greater.

3.2 Experiments on English CHILDES

The CHILDES corpus contains transcriptions of nat-uralistic child-directed speech. The English subsetstudied here consists of all caregiver text extractedfrom Adam, Eve, and Sarah of the Harvard (Brown)corpus, yielding 8,307 types and 588,888 tokens.The tag set used in the Brown corpus consists of55 idiosyncratic tags, and this is tested along with amapping that reduces these to an 8-tag (+SKIP5) set(DT, IN, JJ, NN, PRP, RB, VB, SKIP). Table 1 reports

5Tags which do not correspond to any of the 8 are mappedto SKIP and not scored.

k # Seeds Baseline % Type Acc.100 58 58.0 94.01000 100 10.0 81.28307 130 1.6 62.8

Table 1: CHILDES type accuracy by tree size. Baseline indi-

cates the contribution from the seeds alone. Accuracy presented

as percents

the impact of k on performance and indicates thebest achieved type accuracy results training on thefull Brown tag set then mapping to the reduced set(Brown-to-Reduced) for scoring for final k = 100,k = 1000, and k = 8307.6 As expected, percentcorrect decreases for higher k because infrequentwords which occur only once or twice in the cor-pus provide weaker distributional information thanfrequent words do. The number of seeds used neverreaches 55*3 because some tags do not appear inthe top k words. This is the same as the WSJ corpusFW-problem described earlier.

To determine what impact the choice of tag setand number of seeds have on the results, the exper-iment was run and evaluated directly on the Browntag set as well as the reduced tag set. Reduced tagset experiments were run with 3 seeds per tag and11 per tag. The Brown-to-reduced tests smooth outthe difference between related tags when comput-ing accuracy scores. The difference between B-to-Rand straight Brown in Table 2 implies that the modelstruggles to differentiate some of the Brown tags.This could be because some of the more eccentricBrown tags do not actually distinguish distribution-ally coherent classes. The difference between the 3-seed and 11-seed results indicates that performancelargely depends on the number of seeds. Note, how-ever, that the baseline is still quite low even with 11seeds per tag.

The point of syntactic frames is that they are avail-able as primary evidence early on, and nobody whoworks on them would argue that they are the onlysource of evidence, so it is unsurprising that perfor-mance declines using frames alone for large k. Ifsyntactic frames are indeed useful, then we wouldexpect their application on early vocabulary to carrybenefits downstream, and this this is what Table 3

6k sequence K = (100, 500, 900, 1000, 2000, 4000,6000, 8307) was used.

49

Page 8: Syntactic Category Learning as Iterative Prototype-Driven

k Tag Set # Seeds Baseline Type Acc.1000 B-to-R 100 10.0 81.21000 Brown 100 10.0 70.31000 Reduced 24 2.4 51.81000 Reduced 85 8.5 80.68307 B-to-R 130 1.6 62.88307 Brown 130 1.6 44.08307 Reduced 24 0.3 25.38307 Reduced 85 1.0 53.3

Table 2: CHILDES type accuracy by tag set and seed set size

Tag Set # Seeds Basic Acc. Iter. Acc.B-to-R 130 44.2 62.8Brown 130 27.7 44.0Reduced 24 9.3 25.3Reduced 85 44.0 53.3

Table 3: Comparison between CHILDES basic and iterative

prototype-driven labeling performance

shows. It compares the results of iterative applica-tion from k = 100 to k = 8307 used for the previ-ous experiments to a single non-iterative applicationk = 8307. The iterative results are 9 to 18 pointshigher, demonstrating that syntactic frames appliedto early vocabulary set up better performance later.The iterative application of the algorithm which ex-pands the seed set as the lexicon grows is critical tomodel performance.

Up to this point, seeds have been selected post hocby corpus frequency, which cannot possibly be howchildren use semantic bootstrapping. To correct forthis, a set of 82 lower frequency but high saliencyseed words was selected for comparison based onstudies of salience in child-directed speech (Carlsonet al., 2014). Tables 5 lays out the result achievedwith salient seeds on the reduced tag set, which canbe compared to the 11 frequent seed results from Ta-ble 2. The type accuracy is lower and roughly whatis to be expected given the size of the seed set eventhough the seeds themsleves are of lower frequencyon average.

Tag Set # Seeds Baseline Token Acc.B-to-R 130 50.2 82.7Brown 130 48.0 73.4Reduced 24 28.8 60.0Reduced 85 52.3 83.5

Table 4: CHILDES one-to-one token accuracy performance

k # Seeds Baseline Type Acc.1000 74 7.4 73.48307 82 1.0 49.5

Table 5: CHILDES type accuracy with salient seeds

Language # Seeds k=1,000 k=10,000French 28 77.92 62.07German 30 79.04 26.52Indonesian 30 75.84 65.21Italian 26 54.26 37.08Japanese 24 47.78 48.31Korean 26 33.47 39.19Portuguese 28 65.40 49.44Spanish 29 63.41 46.14Swedish 37 51.10 33.96

Table 6: UTB type accuracy by language

3.3 Experiments on Other Languages

The CHILDES results suggest that information fromsyntactic frames is useful for assigning syntactic cat-egories from English child-directed speech. In or-der to compare performance across other languagesas well, we apply the algorithm to nine languagesfrom the Universal Dependency Treebank: German,Spanish, French, Indonesian, Italian, Japanese, Ko-rean, Brazilian Portuguese, and Swedish. These cor-pora share a 10-tag tag set (excluding punctuationand non-word tags) and were compiled for the sametask, which means that different performances re-flect differences in the languages themselves to theextent possible. In order to more closely align theseexperiments to syntactic category learning, no seedswere provided for the punctuation ( , MAD, MID,PAD) or non-word (X) tags, and punctuation andnon-words were not scored.7

Performance varies substantially across lan-guages. At k = 1000, type accuracy ranges from thelow 30s (Korean), to the high 70s (German, French,Indonesian), and at k = 10000 from the mid 20s(German) to the 60s (French, Indonesian). Much ofthis variation can be explained linguistically with theimportant caveat that extra-linguistic factors in thecorpora must still be at play.

Languages with complicated inflection are at a

7Each language was tested with confidence = 0.3,0.5, 0.7, 0.9, and 1.0. Only the best confidence foreach language is reported. The same k sequence K =(100, 500, 900, 1000, 2000, 5000, 9000, 10000) was used eachtime.

50

Page 9: Syntactic Category Learning as Iterative Prototype-Driven

disadvantage because the distributional context in-formation of their roots is spread across inflectionalforms, and without reference to word-internal fea-tures, it is impossible to group these forms by rootto pool that information. For example, while allthe distributional information for English BLUE iscollected in the contexts for the word “blue,” Ger-man spreads its distributional information across thecontexts for its six inflected forms, “blau,” “blauer,”“blauen,” “blauem,” “blaue,” and “blaues,” and ouralgorithm has no way to combine them.

Another way to think about this is to compare to-ken/type ratios between languages, which are equiv-alent to how many syntactic frames on average con-tribute to the context vectors for each type. Alow ratio indicates that the typical context vectoris sparser because it accounts for information fromfewer frames. Languages with complex inflectionnaturally have more types and a lower token/type ra-tio.

The particularly poor performance for Korean andJapanese is at least partially due to the UTB cor-pus tokenization which attaches phrase final syntac-tic clitics to the preceding word as opposed to stan-dalone tokens. This is in line with the traditionalconception of “word” boundaries in these languages(called bunsetsu in Japanese and eojeol in Korean),but is not an obviously correct choice from a cogni-tive standpoint, and would be equivalent to not seg-menting the possessive ’s clitic in English. As shownin Table 7, bunsetsu tokenization creates multiplewords for APPLE and PEAR and two example clitics,each with disjoint right contexts, while standalonetokenization would yield one word each for APPLE

and PEAR with both clitics as right contexts. Thistechnical choice effectively renders the clitics as in-flections like in German rather than as useful syn-tactic context information. Also, it is unclear whyKorean and Japanese perform better at high k thanlow k at the reported confidences.8

Languages with freer word order are also at a dis-advantage because the entire premise of syntacticframes assumes that the syntax forces categories toappear adjacent to certain other categories. A freerword order means more violations of this assump-

8This only happens for Japanese and Korean, and only atc = 0.7, 0.9 for both.

Tokenization Text Strings Right FramesBunsetsu ringo-ga X ringo-ga: {X}

ringo-wo Y ringo-wo: {Y}nashi-ga Z nashi-ga: {Z}nashi-wo W nashi-wo: {W}

Standalone ringo ga X ringo: {ga, wo}ringo wo Y nashi: {ga, wo}nashi ga Z ga: {X, Z}nashi wo W wo: {Y, W}

Table 7: Right frames for apple-NOM X, apple-ACC Y, pear-

NOM Z, and pear-ACC W

tion relative to the number of frames attested, whichmanifests as more uniform and less discriminablecontext vectors.

3.4 Comparison with Haghighi & Klein

In order to ground the performance of this algorithmin the broader research context, we compare resultswith Haghighi and Klein (2006) since it is similarlysemi-supervised. The H&K PROTO model repre-sents the fairest comparison because they were cal-culated on atypically small datasets, fractions of theEnglish Wall Street Journal and Mandarin ChineseTreebank. They report one-to-one token accuracy,so we follow suit.

The WSJ dataset contains 16,839 types across193,000 tokens9, capitalization removed.10 Themodel takes 3 seeds for each of the 45 tags in thePenn Treebank tag set. Table 8 presents WSJ typeaccuracy at k = 1000 and k = 10000. The1000-type model takes about five minutes to train,while the 10,000-type model takes just over eighthours under a naıve clustering implementation. Theperformance is overall worse than for CHILDES,likely because the type diversity is much higher thanwhat would be expected in child-directed speech.CHILDES has a token/type ratio of 71.4, and theWSJ’s ratio is almost three times lower at 27.8.

The CTB dataset contains 8,842 types across only60,000 tokens and is annotated with a 33-tag set sim-ilar to the Penn Treebank tag set. Table 8 comparesthe Chinese type results to English. The WSJ modeloutperforms the CTB model, which, among factorslike corpus size, is probably related to the number ofseeds present and the token/type ratio.

9Starting from section 2. H&K report 18,423 types.10H&K retain capitalization as a model feature.

51

Page 10: Syntactic Category Learning as Iterative Prototype-Driven

Corpus k # Seeds Baseline Type Acc.WSJ 1000 95 9.5 57.9WSJ 10000 95 1.0 30.2CTB 1000 74 7.4 50.4CTB 8842 74 0.8 27.5

Table 8: Wall Street Journal and Chinese Treebank Type Accu-

racy

Model # Seeds Base Top k All All+k=1000 95 40.5 74.3 54.7 60.2k=10000 95 40.5 63.2 60.9 61.4H&K06 135 41.3 - 68.8 -

Table 9: Wall Street Journal token accuracy and comparison

The CTB model was trained on all types, but theWSJ model was only trained up to k = 10000(accounting for 96.46% of tokens) because of timeconsiderations, which makes token accuracy lessstraightforward to compute. Three numbers are pro-vided: Top k only scores words appearing in the topk by frequency, All scores the top k as before andmarks all other tokens incorrect, and All+ insteadassigns tokens outside the top k the type of theirnearest seed by symmetricized KL-distance. Table9 compares WSJ results with H&K. Overall perfor-mance is lower but still above 60.

Table 10 compares against H&K’s CTB results.Here, the iterative prototype-driven labeling modelis clearly superior. Even the k = 1000 All score is 8points higher than H&K’s MRF model, and the bestk = 10000 All+ score is over 15 points higher. Thatlarge discrepancy is due to H&K’s reliance on or-thographic features. Since suffix n-grams are mean-ingless in Chinese, they were forced to discard themfrom their model. It seems that the reason why H&Koutperformed our model by over 7 points on Englishwas that their model made reference to the word-internal features which ours discards. Chinese repre-sents a more level playing field which demonstrateshow our model makes good use of sparse distribu-tional information.

Model # Seeds Base Top k All All+k=1000 74 29.5 62.80 46.9 50.4k=8841 74 29.5 - 54.1 -H&K06 99 34.4 - 39.0 -

Table 10: Chinese Treebank token accuracy and comparison

4 Discussion and Future Work

The minimally supervised iterative prototype-drivenlabeling algorithm laid out in this paper leveragesonly simple distributional statistics over adjacentword-forms to perform syntactic category labeling.Nevertheless it achieves English and Chinese re-sults comparable to and surpassing the more com-plex H&K model respectively, and similar perfor-mance on French and Indonesian, though it struggleson other languages.

The most glaring deficiency in the model is that itlacks any notion of word-internal features. Throw-ing away this critical information prevents it fromidentifying and utilizing affixes as cues for syntac-tic categories and from pooling evidence from word-forms that share common roots. One promising av-enue of research therefore, is determining how tocleanly incorporate morphological information intothe clustering algorithm. This will be helpful forlanguages like Japanese under bunsetsu tokeniza-tion or German, and absolutely critical for languagefamilies with complex agglutinative or polysyntheticmorphology like Turkic, Eskimo-Aleut, or Bantu.

Perhaps less obviously, the model is missing outon an important generalization by only training onlexical syntactic frames. Children are not only sen-sitive to the specific lexical items surrounding aword, but also the syntactic category of those items(Reeder et al., 2013). For example, a word precededby a determiner and followed by a noun (DT N)is almost certainly an adjective, regardless of whichdeterminer and which noun it is, so even though theadjectives in “the shiny ball” and “a scary cube”have different lexical contexts, they share the samecategory context. Relevant to our algorithm, cate-gory contexts reveal distributional similarities thatare hidden by lexical contexts alone.

Acknowledgments

We would like to thank Charles Yang and MitchMarcus for their helpful insights along with the restof the Penn LORELEI team. This work was fundedby the DARPA LORELEI program under Agree-ment No. HR0011-15-2-0023 and by the U.S. ArmyResearch Office (ARO) through an NDSEG fellow-ship awarded to the author.

52

Page 11: Syntactic Category Learning as Iterative Prototype-Driven

ReferencesTaylor Berg-Kirkpatrick, Alexandre Bouchard-Cote,

John DeNero, and Dan Klein. 2010. Painless unsu-pervised learning with features. In Human LanguageTechnologies: The 2010 Annual Conference of theNorth American Chapter of the Association for Com-putational Linguistics, pages 582–590. Association forComputational Linguistics.

Peter F Brown, Peter V Desouza, Robert L Mercer, Vin-cent J Della Pietra, and Jenifer C Lai. 1992. Class-based n-gram models of natural language. Computa-tional linguistics, 18(4):467–479.

Roger W Brown. 1957. Linguistic determinism and thepart of speech. The Journal of Abnormal and SocialPsychology, 55(1):1.

Matthew T Carlson, Morgan Sonderegger, and MaxBane. 2014. How children explore the phonologicalnetwork in child-directed speech: A survival analysisof childrens first word productions. Journal of mem-ory and language, 75:159–180.

Emmanuel Chemla, Toben H Mintz, Savita Bernal, andAnne Christophe. 2009. Categorizing words usingfrequent frames: what cross-linguistic analyses revealabout distributional acquisition strategies. Develop-mental science, 12(3):396–406.

Christos Christodoulopoulos, Sharon Goldwater, andMark Steedman. 2010. Two decades of unsupervisedpos induction: How far have we come? In Proceed-ings of the 2010 Conference on Empirical Methods inNatural Language Processing, pages 575–584. Asso-ciation for Computational Linguistics.

Alexander Clark. 2003. Combining distributional andmorphological information for part of speech induc-tion. In Proceedings of the tenth conference on Eu-ropean chapter of the Association for ComputationalLinguistics-Volume 1, pages 59–66. Association forComputational Linguistics.

Lila R Gleitman, Kimberly Cassidy, Rebecca Nappa,Anna Papafragou, and John C Trueswell. 2005.Hard words. Language Learning and Development,1(1):23–64.

Sharon Goldwater and Tom Griffiths. 2007. A fullybayesian approach to unsupervised part-of-speech tag-ging. In Annual meeting-association for computa-tional linguistics, volume 45, page 744.

Isaac David Guedalia, Mickey London, and Michael Wer-man. 1999. An on-line agglomerative clusteringmethod for nonstationary data. Neural computation,11(2):521–540.

Aria Haghighi and Dan Klein. 2006. Prototype-drivenlearning for sequence models. In Proceedings ofthe main conference on Human Language Technol-ogy Conference of the North American Chapter of the

Association of Computational Linguistics, pages 320–327. Association for Computational Linguistics.

Yang Charles Hewitt, John. 2017. Bootstrapping for syn-tactic categories. In Cognitive Science Society, Lon-don, UK.

Mark Johnson. 2007. Why doesn’t em find good hmmpos-taggers? In EMNLP-CoNLL, pages 296–305.

Brian MacWhinney. 2000. The CHILDES project: Thedatabase, volume 2. Psychology Press.

Michael Maratsos. 1979. How to get from words to sen-tences. Perspectives in Psycholinguistics.

Mitchell P Marcus, Mary Ann Marcinkiewicz, and Beat-rice Santorini. 1993. Building a large annotated cor-pus of english: The penn treebank. Computational lin-guistics, 19(2):313–330.

Andrew McCallum, Kamal Nigam, and Lyle H Ungar.2000. Efficient clustering of high-dimensional datasets with application to reference matching. In Pro-ceedings of the sixth ACM SIGKDD international con-ference on Knowledge discovery and data mining,pages 169–178. ACM.

Ryan T McDonald, Joakim Nivre, Yvonne Quirmbach-Brundage, Yoav Goldberg, Dipanjan Das, KuzmanGanchev, Keith B Hall, Slav Petrov, Hao Zhang, OscarTackstrom, et al. 2013. Universal dependency annota-tion for multilingual parsing. In ACL (2), pages 92–97.

Marie Meteer, Richard Schwartz, and Ralph Weischedel.1991. Studies in part of speech labelling. In Pro-ceedings of the workshop on Speech and Natural Lan-guage, pages 331–336. Association for ComputationalLinguistics.

Toben H Mintz. 2003. Frequent frames as a cue forgrammatical categories in child directed speech. Cog-nition, 90(1):91–117.

Cornelia Parkes, Alexander Malek, Er M Malek, andMitchell Marcus. 1998. Towards unsupervised ex-traction of verb paradigms from large corpora.

Steven Pinker. 1984. Language learnability and lan-guage development.

Martin Redington, Nick Chater, and Steven Finch.1998. Distributional information: A powerful cuefor acquiring syntactic categories. Cognitive science,22(4):425–469.

Patricia A Reeder, Elissa L Newport, and Richard NAslin. 2013. From shared contexts to syntactic cate-gories: The role of distributional information in learn-ing linguistic form-classes. Cognitive psychology,66(1):30–54.

Jean A Rondal, Martine Ghiotto, Serge Bredart, and Jean-Francois Bachelet. 1987. Age-relation, reliability andgrammatical validity of measures of utterance length.Journal of child language, 14(3):433–446.

53

Page 12: Syntactic Category Learning as Iterative Prototype-Driven

Rushen Shi and Andreane Melancon. 2010. Syntac-tic categorization in french-learning infants. Infancy,15(5):517–533.

Noah A Smith and Jason Eisner. 2005. Contrastive esti-mation: Training log-linear models on unlabeled data.In Proceedings of the 43rd Annual Meeting on Associ-ation for Computational Linguistics, pages 354–362.Association for Computational Linguistics.

Naiwen Xue, Fei Xia, Fu-Dong Chiou, and Marta Palmer.2005. The penn chinese treebank: Phrase structure an-notation of a large corpus. Natural language engineer-ing, 11(2):207–238.

54