Knowledge

Analogical modeling

Source đź“ť

613:
A supracontext may exhibit several behaviors, but contain no exemplars that occur in any more specific supracontext (that is, in any of its subcontexts); in this case it is non-deterministically homogeneous and is included. Here there is no great evidence that a systematic behavior occurs, but also no counterargument. Finally, a supracontext may be heterogeneous, meaning that it exhibits behaviors that are found in a subcontext (closer to the given context), and also behaviors that are not. Where the ambiguous behavior of the nondeterministically homogeneous supracontext was accepted, this is rejected because the intervening subcontext demonstrates that there is a better theory to be found. The heterogeneous supracontext is therefore excluded. This guarantees that we see an increase in meaningfully consistent behavior in the analogical set as we approach the given context.
1779: 2018: 944: 549:
consist of the letters of a word. Each exemplar in the dataset is stored with an outcome, such as a phoneme or phone to be generated. When the model is presented with a novel situation (in the form of an outcome-less feature vector), the engine algorithmically sorts the dataset to find exemplars that helpfully resemble it, and selects one, whose outcome is the model's prediction. The particulars of the algorithm distinguish one exemplar-based modeling system from another.
459: 617:
exemplar to which it points provides the outcome. This gives each supracontext an importance proportional to the square of its size, and makes each exemplar likely to be selected in direct proportion to the sum of the sizes of all analogically consistent supracontexts in which it appears. Then, of course, the probability of predicting a particular outcome is proportional to the summed probabilities of all the exemplars that support it.
1931:", which in the dataset consists of "3 1 0 e" and "3 1 1 r". There are 2 disagreements in this subcontext; 1 pointing from each of the exemplars to the other (see the pointer network pictured above). Therefore, only supracontexts containing this subcontext will contain any disagreements. We use a simple rule to identify the homogeneous supracontexts: 2367:
Though analogical modeling aims to create a model free from rules seen as contrived by linguists, in its current form it still requires researchers to select which variables to take into consideration. This is necessary because of the so-called "exponential explosion" of processing power requirements
530:
approaches, in that it is data-based rather than abstraction-based; but it is distinguished by its ability to cope with imperfect datasets (such as caused by simulated short term memory limits) and to base predictions on all relevant segments of the dataset, whether near or far. In language modeling,
612:
The engine now chooses the analogical set from among the supracontexts. A supracontext may contain exemplars that only exhibit one behavior; it is deterministically homogeneous and is included. It is a view of the data that displays regularity, or a relevant theory that has never yet been disproven.
604:
supracontexts. Each supracontext is a set of exemplars in which one or more variables have the same values that they do in the given context, and the other variables are ignored. In effect, each is a view of the data, created by filtering for some criteria of similarity to the given context, and the
548:
engine and a problem-specific dataset. Within the dataset, each exemplar (a case to be reasoned from, or an informative past experience) appears as a feature vector: a row of values for the set of parameters that define the problem. For example, in a spelling-to-sound task, the feature vector might
1541:
Behavior can only be predicted for a given context; in this example, let us predict the outcome for the context "3 1 2". To do this, we first find all of the contexts containing the given context; these contexts are called supracontexts. We find the supracontexts by systematically eliminating the
568:
of supracontexts, and probabilistically selects an exemplar from the analogical set with a bias toward those in large supracontexts. This multilevel search exponentially magnifies the likelihood of a behavior's being predicted as it occurs reliably in settings that specifically resemble the given
616:
With the analogical set chosen, each appearance of an exemplar (for a given exemplar may appear in several of the analogical supracontexts) is given a pointer to every other appearance of an exemplar within its supracontexts. One of these pointers is then selected at random and followed, and the
608:
It is important to note that the supracontexts are not equal peers one with another; they are arranged by their distance from the given context, forming a hierarchy. If a supracontext specifies all of the variables that another one does and more, it is a subcontext of that other one, and it lies
1953:
Only one subcontext contains any data. The subcontext does not have to be deterministic for the supracontext to be homogeneous. For example, while the supracontexts "3 1 -" and "- 1 2" are deterministic and only contain one non-empty subcontext, "3 - -" contains only the subcontext "3 1
1973:
There is actually a 4th type of homogeneous supracontext: it contains more than one non-empty subcontext and it is non-deterministic, but the frequency of outcomes in each sub-context is exactly the same. Analogical modeling does not consider this situation, however, for 2 reasons:
1784:
The next step is to determine which exemplars belong to which contexts in order to determine which of the contexts are homogeneous. The table below shows each of the subcontexts, their behavior in terms of the given exemplars, and the number of disagreements within the behavior:
609:
closer to the given context. (The hierarchy is not strictly branching; each supracontext can itself be a subcontext of several others, and can have several subcontexts.) This hierarchy becomes significant in the next step of the algorithm.
1497: 2012:
Next we construct the analogical set, which consists of all of the pointers and outcomes from the homogeneous supracontexts. The figure below shows the pointer network with the homogeneous contexts highlighted.
901: 2005:
test. This is the only test of homogeneity that requires arithmetic, and ignoring it allows our tests of homogeneity to become statistically free, which makes AM better for modeling human reasoning.
1943:
The supracontext is empty. This is the case for "3 - 2", which contains no data points. There can be no increase in the number of disagreements, and the supracontext is trivially homogeneous.
1935:
If the number if disagreements in the supracontext is greater than the number of disagreements in the contained subcontext, we say that it is heterogeneous; otherwise, it is homogeneous.
911:
This terminology is best understood through an example. In the example used in the second chapter of Skousen (1989), each context consists of three variables with potential values 0-3
1350: 552:
In AM, we think of the feature values as characterizing a context, and the outcome as a behavior that occurs within that context. Accordingly, the novel situation is known as the
1292: 1231: 809: 605:
total set of supracontexts exhausts all such views. Alternatively, each supracontext is a theory of the task or a proposed rule whose predictive power needs to be evaluated.
770: 1126: 1170: 531:
AM has successfully predicted empirically valid forms for which no theoretical explanation was known (see the discussion of Finnish morphology in Skousen et al. 2002).
2003: 715: 1536: 1082: 1047: 1013: 1571: 677: 602: 1946:
The supracontext is deterministic, meaning that only one type of outcome occurs in it. This is the case for "- 1 2" and "- - 2", which contain only data with the
979: 646: 949:
The solid lines represent pointers between exemplars with matching outcomes; the dotted lines represent pointers between exemplars with non-matching outcomes.
1958:". This subcontext contains "3 1 0 e" and "3 1 1 r", making it non-deterministic. We say that this type of supracontext is unobstructed and non-deterministic. 2017: 2335:
Skousen's proposal appears to address that criticism by proposing an explicit mechanism for analogy, which can be tested for psychological validity.
577:
AM performs the same process for each case it is asked to evaluate. The given context, consisting of n variables, is used as a template to generate
2202:
it is 9/13 or 69.2%. We can create a more detailed account by listing the pointers for each of the occurrences in the homogeneous supracontexts:
1357: 1778: 2194:. In AM, a pointer is randomly selected and the outcome it points to is predicted. With a total of 13 pointers, the probability of the outcome 1962:
The only two heterogeneous supracontexts are "- 1 -" and "- - -". In both of them, it is the combination of the non-deterministic "3 1
560:), and extracts from the dataset the exemplars that belong to each. The engine then discards those supracontexts whose outcomes are 943: 816: 2008:
It is an extremely rare situation, and thus ignoring it will can be expected not to have a large effect on the predicted outcome.
527: 486: 376: 2435: 556:
Given the known features of the context, the AM engine systematically generates all contexts that include it (all of its
2544: 2539: 2489: 2470: 2424: 2484:(Human Cognitive Processing vol. 10). Amsterdam/Philadelphia: John Benjamins Publishing Company. p. x+417pp. 1927:
Analyzing the subcontexts in the table above, we see that there is only 1 subcontext with any disagreements: "3 1
336: 2331:
and others have more recently criticized analogy as too vague to really be useful (Bańko 1991), an appeal to a
396: 341: 114: 371: 62: 1299: 2396: 316: 182: 522:. It is applicable to language modeling and other categorization tasks. Analogical modeling is related to 436: 142: 1242: 1181: 777: 2381: 479: 426: 326: 152: 726: 2348: 545: 515: 331: 274: 89: 24: 1093: 2391: 1137: 431: 269: 246: 2372:
could provide the solution to such performance bottlenecks (Skousen et al. 2002, see pp 45–47).
1981: 688: 381: 348: 301: 217: 197: 177: 79: 57: 52: 2368:
of the computer software used to implement analogical modeling. Recent research suggests that
1507: 1054: 2416: 2410: 2324: 1019: 985: 157: 1549: 655: 580: 472: 401: 311: 192: 137: 34: 8: 958: 242: 172: 147: 119: 631: 462: 441: 411: 366: 321: 289: 279: 167: 162: 620:(Skousen 2002, in Skousen et al. 2002, pp. 11–25, and Skousen 2003, both passim) 2485: 2466: 2420: 2369: 2323:
Analogy has been considered useful in describing language at least since the time of
458: 306: 284: 227: 2549: 406: 237: 232: 207: 202: 187: 17: 2479: 507: 251: 2499: 2450: 2533: 2519: 2386: 1573:
supracontexts. The following table lists each of the sub- and supracontexts;
523: 511: 2328: 1492:{\displaystyle n_{r}(n-n_{r})+n_{e}(n-n_{e})=n^{2}-(n_{r}^{2}+n_{e}^{2})=8} 561: 294: 84: 564:(this measure of consistency will be discussed further below), leaving an 2352: 519: 446: 421: 42: 2186:
4 of the pointers in the analogical set are associated with the outcome
416: 99: 2344: 391: 386: 222: 212: 104: 94: 2480:
Royal Skousen; Deryle Lonsdale; Dilworth B. Parkinson, eds. (2002).
2343:
Analogical modeling has been employed in experiments ranging from
1939:
There are 3 situations that produce a homogeneous supracontext:
2514: 2356: 109: 1978:
Determining whether this 4 situation has occurred requires a
2504:. Presented at the Berkeley Linguistics Society conference. 2501:
Analogical Modeling: Exemplars, Rules, and Quantum Computing
896:{\displaystyle \sum {n_{i}(n-n_{i})}=n^{2}-\sum {n_{i}^{2}}} 2482:
Analogical Modeling: An exemplar-based approach to language
1984: 1552: 1510: 1360: 1302: 1245: 1184: 1140: 1096: 1057: 1022: 988: 961: 819: 780: 729: 691: 658: 634: 583: 2415:(hardcover). Dordrecht: Kluwer Academic Publishers. 2023:
The pointers are summarized in the following table:
1774:
These contexts are shown in the venn diagram below:
514:, professor of Linguistics and English language at 2198:being predicted is 4/13 or 30.8%, and for outcome 1997: 1565: 1530: 1491: 1344: 1286: 1225: 1164: 1120: 1076: 1041: 1007: 973: 895: 803: 764: 709: 671: 640: 596: 2531: 952:The statistics for this example are as follows: 572: 2433: 2460: 2408: 480: 2515:Analogical Modeling Research Group Homepage 2315:of each of the instances in the data set. 487: 473: 2436:"Review: Analogical Modeling of Language" 939:We define a network of pointers like so: 2465:. Dordrect: Kluwer Academic Publishers. 1970:outcome which causes the heterogeneity. 1966:" with other subcontexts containing the 936:3 1 0 e 0 3 2 r 2 1 0 r 2 1 2 r 3 1 1 r 510:based analogical reasoning, proposed by 2532: 2190:, and the other 9 are associated with 1345:{\displaystyle n_{r}^{2}+n_{e}^{2}=17} 544:An exemplar-based model consists of a 2318: 1542:variables in the given context; with 925:The two outcomes for the dataset are 1235:number of disagreements for outcome 1174:number of disagreements for outcome 719:number of disagreements for outcome 377:Conservative and innovative language 1546:variables, there will generally be 13: 14: 2561: 2508: 1130:number of agreements for outcome 1086:number of agreements for outcome 681:number of agreements for outcome 534: 2016: 1777: 1287:{\displaystyle n_{e}(n-n_{e})=4} 1226:{\displaystyle n_{r}(n-n_{r})=4} 942: 804:{\displaystyle \sum {n_{i}^{2}}} 457: 2412:Analogical Modeling of Language 2338: 1354:total number of disagreements: 813:total number of disagreements: 1480: 1444: 1425: 1406: 1390: 1371: 1275: 1256: 1214: 1195: 853: 834: 765:{\displaystyle n_{i}(n-n_{i})} 759: 740: 1: 2402: 1504:or fraction of disagreement: 573:Analogical modeling in detail 2449:(2): 246–248. Archived from 2434:Miroslaw BaĹ„ko (June 1991). 2397:k-nearest neighbor algorithm 2222: 1920: 1901: 1884: 1867: 1850: 1836: 1821: 1807: 1296:total number of agreements: 1121:{\displaystyle n_{r}^{2}=16} 774:total number of agreements: 317:Functional discourse grammar 183:Ethnography of communication 7: 2375: 2362: 1165:{\displaystyle n_{e}^{2}=1} 623: 539: 437:Second-language acquisition 10: 2566: 2520:LINGUIST List Announcement 2036: 2033: 2028: 1796: 1793: 1790: 1051:total number of pairings: 906: 652:total number of pairings: 115:Syntax–semantics interface 22: 15: 2545:Computational linguistics 2540:Classification algorithms 2498:Skousen, Royal. (2003). 2443:Computational Linguistics 2382:Computational Linguistics 1998:{\displaystyle \chi ^{2}} 933:, and the exemplars are: 710:{\displaystyle n_{i}^{2}} 427:Philosophy of linguistics 327:Interactional linguistics 2349:morphology (linguistics) 1531:{\displaystyle 8/25=.32} 1077:{\displaystyle n^{2}=25} 546:general-purpose modeling 516:Brigham Young University 506:) is a formal theory of 25:Analogy (disambiguation) 16:Not to be confused with 2526:, Skousen et al. (2002) 2392:Instance-based learning 1042:{\displaystyle n_{e}=1} 1008:{\displaystyle n_{r}=4} 2461:Royal Skousen (1992). 2409:Royal Skousen (1989). 1999: 1937: 1567: 1532: 1493: 1346: 1288: 1227: 1166: 1122: 1078: 1043: 1009: 975: 897: 805: 766: 711: 673: 642: 598: 264:Theoretical frameworks 218:Philosophy of language 198:History of linguistics 2463:Analogy and Structure 2145:"2 1 2 r", "0 3 2 r" 2121:"3 1 0 e", "3 1 1 r" 2073:"3 1 0 e", "3 1 1 r" 2000: 1933: 1568: 1566:{\displaystyle 2^{m}} 1533: 1494: 1347: 1289: 1228: 1167: 1123: 1079: 1044: 1010: 976: 898: 806: 767: 712: 674: 672:{\displaystyle n^{2}} 643: 628:Given a context with 599: 597:{\displaystyle 2^{n}} 158:Conversation analysis 2311:We can then see the 1982: 1550: 1508: 1358: 1300: 1243: 1182: 1138: 1094: 1055: 1020: 986: 959: 817: 778: 727: 689: 656: 632: 581: 402:Internet linguistics 312:Construction grammar 23:For other uses, see 2524:Analogical Modeling 1577:means "not x", and 1479: 1461: 1335: 1317: 1155: 1111: 974:{\displaystyle n=5} 921:Variable 3: 0,1,2,3 918:Variable 2: 0,1,2,3 915:Variable 1: 0,1,2,3 891: 799: 706: 500:Analogical modeling 337:Systemic functional 132:Applied linguistics 74:General linguistics 2319:Historical context 1995: 1581:means "anything". 1563: 1528: 1489: 1465: 1447: 1342: 1321: 1303: 1284: 1223: 1162: 1141: 1118: 1097: 1074: 1039: 1005: 971: 893: 877: 801: 785: 762: 707: 692: 669: 638: 594: 442:Theory of language 412:Origin of language 367:Autonomy of syntax 322:Grammaticalization 168:Discourse analysis 163:Corpus linguistics 2370:quantum computing 2313:analogical effect 2309: 2308: 2184: 2183: 2180: 2179: 2159: 2158: 2135: 2134: 2111: 2110: 2087: 2086: 2063: 2062: 1925: 1924: 1819:3 1 0 e, 3 1 1 r 1772: 1771: 641:{\displaystyle n} 497: 496: 285:Distributionalism 228:Psycholinguistics 2557: 2495: 2476: 2457: 2455: 2440: 2430: 2333:deus ex machina. 2296: 2280: 2264: 2248: 2232: 2205: 2204: 2170: 2169: 2149: 2148: 2142: 2125: 2124: 2118: 2101: 2100: 2094: 2077: 2076: 2070: 2049: 2048: 2026: 2025: 2020: 2004: 2002: 2001: 1996: 1994: 1993: 1965: 1957: 1930: 1915: 1912: 1909: 1895: 1892: 1879: 1875: 1862: 1859: 1844: 1830: 1816: 1788: 1787: 1781: 1768: 1765: 1762: 1758: 1755: 1751: 1747: 1743: 1740: 1736: 1732: 1728: 1722: 1714: 1711: 1707: 1703: 1697: 1690: 1686: 1682: 1678: 1672: 1665: 1662: 1658: 1654: 1648: 1640: 1634: 1626: 1620: 1613: 1607: 1597: 1584: 1583: 1576: 1572: 1570: 1569: 1564: 1562: 1561: 1537: 1535: 1534: 1529: 1518: 1498: 1496: 1495: 1490: 1478: 1473: 1460: 1455: 1440: 1439: 1424: 1423: 1405: 1404: 1389: 1388: 1370: 1369: 1351: 1349: 1348: 1343: 1334: 1329: 1316: 1311: 1293: 1291: 1290: 1285: 1274: 1273: 1255: 1254: 1232: 1230: 1229: 1224: 1213: 1212: 1194: 1193: 1171: 1169: 1168: 1163: 1154: 1149: 1127: 1125: 1124: 1119: 1110: 1105: 1083: 1081: 1080: 1075: 1067: 1066: 1048: 1046: 1045: 1040: 1032: 1031: 1014: 1012: 1011: 1006: 998: 997: 980: 978: 977: 972: 946: 902: 900: 899: 894: 892: 890: 885: 869: 868: 856: 852: 851: 833: 832: 810: 808: 807: 802: 800: 798: 793: 771: 769: 768: 763: 758: 757: 739: 738: 716: 714: 713: 708: 705: 700: 678: 676: 675: 670: 668: 667: 647: 645: 644: 639: 603: 601: 600: 595: 593: 592: 528:nearest neighbor 489: 482: 475: 461: 407:LGBT linguistics 397:Internationalism 372:Compositionality 233:Sociolinguistics 208:Neurolinguistics 203:Interlinguistics 188:Ethnomethodology 30: 29: 18:Analogical model 2565: 2564: 2560: 2559: 2558: 2556: 2555: 2554: 2530: 2529: 2511: 2492: 2473: 2453: 2438: 2427: 2405: 2378: 2365: 2341: 2321: 2294: 2278: 2262: 2246: 2230: 2224: 2219: 2214: 2212: 2140: 2116: 2092: 2068: 2038: 2030: 1989: 1985: 1983: 1980: 1979: 1963: 1955: 1928: 1913: 1910: 1907: 1893: 1890: 1877: 1873: 1860: 1857: 1842: 1828: 1814: 1766: 1763: 1760: 1756: 1753: 1749: 1745: 1741: 1738: 1734: 1730: 1726: 1720: 1712: 1709: 1705: 1701: 1695: 1688: 1684: 1680: 1676: 1670: 1663: 1660: 1656: 1652: 1646: 1638: 1632: 1624: 1618: 1611: 1605: 1595: 1574: 1557: 1553: 1551: 1548: 1547: 1514: 1509: 1506: 1505: 1474: 1469: 1456: 1451: 1435: 1431: 1419: 1415: 1400: 1396: 1384: 1380: 1365: 1361: 1359: 1356: 1355: 1330: 1325: 1312: 1307: 1301: 1298: 1297: 1269: 1265: 1250: 1246: 1244: 1241: 1240: 1208: 1204: 1189: 1185: 1183: 1180: 1179: 1150: 1145: 1139: 1136: 1135: 1106: 1101: 1095: 1092: 1091: 1062: 1058: 1056: 1053: 1052: 1027: 1023: 1021: 1018: 1017: 993: 989: 987: 984: 983: 960: 957: 956: 937: 909: 886: 881: 876: 864: 860: 847: 843: 828: 824: 823: 818: 815: 814: 794: 789: 784: 779: 776: 775: 753: 749: 734: 730: 728: 725: 724: 701: 696: 690: 687: 686: 663: 659: 657: 654: 653: 633: 630: 629: 626: 588: 584: 582: 579: 578: 575: 542: 537: 493: 452: 451: 362: 354: 353: 265: 257: 256: 252:Writing systems 143:Anthropological 133: 125: 124: 75: 67: 28: 21: 12: 11: 5: 2563: 2553: 2552: 2547: 2542: 2528: 2527: 2517: 2510: 2509:External links 2507: 2506: 2505: 2496: 2490: 2477: 2471: 2458: 2456:on 2003-08-02. 2431: 2425: 2404: 2401: 2400: 2399: 2394: 2389: 2384: 2377: 2374: 2364: 2361: 2340: 2337: 2320: 2317: 2307: 2306: 2303: 2300: 2297: 2291: 2290: 2287: 2284: 2281: 2275: 2274: 2271: 2268: 2265: 2259: 2258: 2255: 2252: 2249: 2243: 2242: 2239: 2236: 2233: 2227: 2226: 2221: 2216: 2209: 2182: 2181: 2178: 2177: 2174: 2167: 2164: 2161: 2160: 2157: 2156: 2153: 2146: 2143: 2137: 2136: 2133: 2132: 2129: 2122: 2119: 2113: 2112: 2109: 2108: 2105: 2098: 2095: 2089: 2088: 2085: 2084: 2081: 2074: 2071: 2065: 2064: 2061: 2060: 2055: 2046: 2044: 2041: 2040: 2035: 2032: 2010: 2009: 2006: 1992: 1988: 1960: 1959: 1951: 1944: 1923: 1922: 1919: 1916: 1904: 1903: 1900: 1897: 1887: 1886: 1883: 1880: 1870: 1869: 1866: 1863: 1853: 1852: 1849: 1846: 1839: 1838: 1835: 1832: 1824: 1823: 1820: 1817: 1810: 1809: 1806: 1803: 1799: 1798: 1797:Disagreements 1795: 1792: 1770: 1769: 1723: 1717: 1716: 1698: 1692: 1691: 1673: 1667: 1666: 1649: 1643: 1642: 1635: 1629: 1628: 1621: 1615: 1614: 1608: 1602: 1601: 1598: 1592: 1591: 1588: 1560: 1556: 1539: 1538: 1527: 1524: 1521: 1517: 1513: 1499: 1488: 1485: 1482: 1477: 1472: 1468: 1464: 1459: 1454: 1450: 1446: 1443: 1438: 1434: 1430: 1427: 1422: 1418: 1414: 1411: 1408: 1403: 1399: 1395: 1392: 1387: 1383: 1379: 1376: 1373: 1368: 1364: 1352: 1341: 1338: 1333: 1328: 1324: 1320: 1315: 1310: 1306: 1294: 1283: 1280: 1277: 1272: 1268: 1264: 1261: 1258: 1253: 1249: 1233: 1222: 1219: 1216: 1211: 1207: 1203: 1200: 1197: 1192: 1188: 1172: 1161: 1158: 1153: 1148: 1144: 1128: 1117: 1114: 1109: 1104: 1100: 1084: 1073: 1070: 1065: 1061: 1049: 1038: 1035: 1030: 1026: 1015: 1004: 1001: 996: 992: 981: 970: 967: 964: 935: 923: 922: 919: 916: 908: 905: 904: 903: 889: 884: 880: 875: 872: 867: 863: 859: 855: 850: 846: 842: 839: 836: 831: 827: 822: 811: 797: 792: 788: 783: 772: 761: 756: 752: 748: 745: 742: 737: 733: 717: 704: 699: 695: 679: 666: 662: 637: 625: 622: 591: 587: 574: 571: 566:analogical set 554:given context. 541: 538: 536: 535:Implementation 533: 495: 494: 492: 491: 484: 477: 469: 466: 465: 454: 453: 450: 449: 444: 439: 434: 432:Prescriptivism 429: 424: 419: 414: 409: 404: 399: 394: 389: 384: 379: 374: 369: 363: 360: 359: 356: 355: 352: 351: 346: 345: 344: 339: 334: 329: 324: 319: 314: 309: 299: 298: 297: 292: 287: 282: 277: 266: 263: 262: 259: 258: 255: 254: 249: 240: 235: 230: 225: 220: 215: 210: 205: 200: 195: 190: 185: 180: 175: 170: 165: 160: 155: 150: 145: 140: 134: 131: 130: 127: 126: 123: 122: 117: 112: 107: 102: 97: 92: 87: 82: 76: 73: 72: 69: 68: 66: 65: 60: 55: 49: 46: 45: 39: 38: 9: 6: 4: 3: 2: 2562: 2551: 2548: 2546: 2543: 2541: 2538: 2537: 2535: 2525: 2521: 2518: 2516: 2513: 2512: 2503: 2502: 2497: 2493: 2491:1-58811-302-7 2487: 2483: 2478: 2474: 2472:0-7923-1935-4 2468: 2464: 2459: 2452: 2448: 2444: 2437: 2432: 2428: 2426:0-7923-0517-5 2422: 2418: 2414: 2413: 2407: 2406: 2398: 2395: 2393: 2390: 2388: 2387:Connectionism 2385: 2383: 2380: 2379: 2373: 2371: 2360: 2358: 2354: 2350: 2346: 2336: 2334: 2330: 2326: 2316: 2314: 2304: 2301: 2298: 2293: 2292: 2288: 2285: 2282: 2277: 2276: 2272: 2269: 2266: 2261: 2260: 2256: 2253: 2250: 2245: 2244: 2240: 2237: 2234: 2229: 2228: 2217: 2215:supracontexts 2210: 2207: 2206: 2203: 2201: 2197: 2193: 2189: 2175: 2172: 2171: 2168: 2165: 2163: 2162: 2154: 2151: 2150: 2147: 2144: 2139: 2138: 2130: 2127: 2126: 2123: 2120: 2115: 2114: 2106: 2103: 2102: 2099: 2096: 2091: 2090: 2082: 2079: 2078: 2075: 2072: 2067: 2066: 2059: 2056: 2054: 2051: 2050: 2047: 2045: 2043: 2042: 2027: 2024: 2021: 2019: 2014: 2007: 1990: 1986: 1977: 1976: 1975: 1971: 1969: 1952: 1949: 1945: 1942: 1941: 1940: 1936: 1932: 1917: 1906: 1905: 1898: 1889: 1888: 1881: 1872: 1871: 1864: 1855: 1854: 1847: 1841: 1840: 1833: 1826: 1825: 1818: 1812: 1811: 1804: 1801: 1800: 1789: 1786: 1782: 1780: 1775: 1724: 1719: 1718: 1699: 1694: 1693: 1674: 1669: 1668: 1650: 1645: 1644: 1636: 1631: 1630: 1622: 1617: 1616: 1609: 1604: 1603: 1599: 1594: 1593: 1589: 1587:Supracontext 1586: 1585: 1582: 1580: 1558: 1554: 1545: 1525: 1522: 1519: 1515: 1511: 1503: 1500: 1486: 1483: 1475: 1470: 1466: 1462: 1457: 1452: 1448: 1441: 1436: 1432: 1428: 1420: 1416: 1412: 1409: 1401: 1397: 1393: 1385: 1381: 1377: 1374: 1366: 1362: 1353: 1339: 1336: 1331: 1326: 1322: 1318: 1313: 1308: 1304: 1295: 1281: 1278: 1270: 1266: 1262: 1259: 1251: 1247: 1238: 1234: 1220: 1217: 1209: 1205: 1201: 1198: 1190: 1186: 1177: 1173: 1159: 1156: 1151: 1146: 1142: 1133: 1129: 1115: 1112: 1107: 1102: 1098: 1089: 1085: 1071: 1068: 1063: 1059: 1050: 1036: 1033: 1028: 1024: 1016: 1002: 999: 994: 990: 982: 968: 965: 962: 955: 954: 953: 950: 947: 945: 940: 934: 932: 928: 920: 917: 914: 913: 912: 887: 882: 878: 873: 870: 865: 861: 857: 848: 844: 840: 837: 829: 825: 820: 812: 795: 790: 786: 781: 773: 754: 750: 746: 743: 735: 731: 722: 718: 702: 697: 693: 684: 680: 664: 660: 651: 650: 649: 635: 621: 618: 614: 610: 606: 589: 585: 570: 567: 563: 559: 558:supracontexts 555: 550: 547: 532: 529: 525: 524:connectionism 521: 517: 513: 512:Royal Skousen 509: 505: 501: 490: 485: 483: 478: 476: 471: 470: 468: 467: 464: 460: 456: 455: 448: 445: 443: 440: 438: 435: 433: 430: 428: 425: 423: 420: 418: 415: 413: 410: 408: 405: 403: 400: 398: 395: 393: 390: 388: 385: 383: 382:Descriptivism 380: 378: 375: 373: 370: 368: 365: 364: 358: 357: 350: 349:Structuralism 347: 343: 340: 338: 335: 333: 332:Prague circle 330: 328: 325: 323: 320: 318: 315: 313: 310: 308: 305: 304: 303: 300: 296: 293: 291: 288: 286: 283: 281: 278: 276: 273: 272: 271: 268: 267: 261: 260: 253: 250: 248: 244: 241: 239: 236: 234: 231: 229: 226: 224: 221: 219: 216: 214: 211: 209: 206: 204: 201: 199: 196: 194: 191: 189: 186: 184: 181: 179: 178:Documentation 176: 174: 171: 169: 166: 164: 161: 159: 156: 154: 153:Computational 151: 149: 146: 144: 141: 139: 136: 135: 129: 128: 121: 118: 116: 113: 111: 108: 106: 103: 101: 98: 96: 93: 91: 88: 86: 83: 81: 78: 77: 71: 70: 64: 61: 59: 56: 54: 51: 50: 48: 47: 44: 41: 40: 36: 32: 31: 26: 19: 2523: 2500: 2481: 2462: 2451:the original 2446: 2442: 2411: 2366: 2342: 2339:Applications 2332: 2329:Noam Chomsky 2322: 2312: 2310: 2199: 2195: 2191: 2187: 2185: 2057: 2052: 2031:supracontext 2022: 2015: 2011: 1972: 1967: 1961: 1947: 1938: 1934: 1926: 1783: 1776: 1773: 1590:Subcontexts 1578: 1543: 1540: 1501: 1236: 1175: 1131: 1087: 951: 948: 941: 938: 930: 926: 924: 910: 720: 682: 627: 619: 615: 611: 607: 576: 565: 562:inconsistent 557: 553: 551: 543: 503: 499: 498: 295:Glossematics 275:Constituency 247:interpreting 85:Lexicography 2353:orthography 2213:homogeneous 2034:Occurrences 2029:Homogeneous 1610:3 1 2, 3 1 1502:uncertainty 520:Provo, Utah 447:Terminology 422:Orthography 342:Usage-based 243:Translating 138:Acquisition 43:Linguistics 2534:Categories 2403:References 2223:Analogical 2208:Occurrence 2097:"2 1 2 r" 1791:Subcontext 648:elements: 417:Orismology 302:Functional 290:Generative 280:Dependency 100:Pragmatics 90:Morphology 80:Diachronic 2417:xii+212pp 2345:phonology 2218:Number of 2211:Number of 2039:pointers 2037:Number of 1987:χ 1679:1 2, 3 1 1651:3 1 2, 3 1623:3 1 2, 3 1442:− 1413:− 1378:− 1263:− 1202:− 874:∑ 871:− 841:− 821:∑ 782:∑ 747:− 569:context. 392:Iconicity 387:Etymology 307:Cognitive 270:Formalist 223:Phonetics 213:Philology 105:Semantics 95:Phonology 2376:See also 2363:Problems 2325:Saussure 2220:pointers 2166:Totals: 1950:outcome. 1918:(empty) 1899:0 3 2 r 1882:2 1 0 r 1865:(empty) 1848:2 1 2 r 1834:(empty) 1805:(empty) 1794:Behavior 624:Formulas 540:Overview 508:exemplar 193:Forensic 173:Distance 120:Typology 35:a series 33:Part of 2550:Analogy 2295:2 1 0 r 2279:0 3 2 r 2263:2 1 2 r 2247:3 1 1 r 2231:3 1 0 e 2225:effect 1733:2, 3 1 1729:1 2, 3 1725:3 1 2, 1704:1 2, 3 1700:3 1 2, 1675:3 1 2, 1655:2, 3 1 1637:3 1 2, 907:Example 148:Applied 58:History 53:Outline 2488:  2469:  2423:  2357:syntax 2289:15.4% 2273:23.1% 2257:30.8% 2241:30.8% 1802:3 1 2 1600:3 1 2 463:Portal 361:Topics 110:Syntax 2454:(PDF) 2439:(PDF) 2305:0.0% 2141:- - 2 2117:3 - - 2093:- 1 2 2069:3 1 - 1721:- - - 1696:- - 2 1671:- 1 - 1647:3 - - 1633:- 1 2 1619:3 - 2 1606:3 1 - 1596:3 1 2 63:Index 2486:ISBN 2467:ISBN 2421:ISBN 2355:and 2347:and 1845:1 2 1813:3 1 1752:, 3 1659:, 3 1641:1 2 929:and 526:and 245:and 238:Text 2522:of 2351:to 1744:2, 1708:2, 1526:.32 518:in 2536:: 2447:17 2445:. 2441:. 2419:. 2359:. 2327:. 2302:0 2299:0 2286:2 2283:1 2270:3 2267:2 2254:4 2251:2 2238:4 2235:2 2176:9 2173:4 2155:4 2152:0 2131:2 2128:2 2107:1 2104:0 2083:2 2080:2 1921:0 1902:0 1896:2 1885:0 1876:1 1868:0 1856:3 1851:0 1837:0 1831:2 1827:3 1822:2 1808:0 1759:, 1748:1 1737:, 1715:2 1687:1 1683:, 1627:2 1520:25 1340:17 1239:: 1178:: 1134:: 1116:16 1090:: 1072:25 723:: 685:: 504:AM 37:on 2494:. 2475:. 2429:. 2200:r 2196:e 2192:r 2188:e 2058:r 2053:e 1991:2 1968:r 1964:2 1956:2 1948:r 1929:2 1914:2 1911:1 1908:3 1894:1 1891:3 1878:2 1874:3 1861:2 1858:1 1843:3 1829:1 1815:2 1767:2 1764:1 1761:3 1757:2 1754:1 1750:2 1746:3 1742:1 1739:3 1735:2 1731:1 1727:3 1713:1 1710:3 1706:1 1702:3 1689:2 1685:3 1681:2 1677:3 1664:2 1661:1 1657:2 1653:1 1639:3 1625:1 1612:2 1579:- 1575:x 1559:m 1555:2 1544:m 1523:= 1516:/ 1512:8 1487:8 1484:= 1481:) 1476:2 1471:e 1467:n 1463:+ 1458:2 1453:r 1449:n 1445:( 1437:2 1433:n 1429:= 1426:) 1421:e 1417:n 1410:n 1407:( 1402:e 1398:n 1394:+ 1391:) 1386:r 1382:n 1375:n 1372:( 1367:r 1363:n 1337:= 1332:2 1327:e 1323:n 1319:+ 1314:2 1309:r 1305:n 1282:4 1279:= 1276:) 1271:e 1267:n 1260:n 1257:( 1252:e 1248:n 1237:e 1221:4 1218:= 1215:) 1210:r 1206:n 1199:n 1196:( 1191:r 1187:n 1176:r 1160:1 1157:= 1152:2 1147:e 1143:n 1132:e 1113:= 1108:2 1103:r 1099:n 1088:r 1069:= 1064:2 1060:n 1037:1 1034:= 1029:e 1025:n 1003:4 1000:= 995:r 991:n 969:5 966:= 963:n 931:r 927:e 888:2 883:i 879:n 866:2 862:n 858:= 854:) 849:i 845:n 838:n 835:( 830:i 826:n 796:2 791:i 787:n 760:) 755:i 751:n 744:n 741:( 736:i 732:n 721:i 703:2 698:i 694:n 683:i 665:2 661:n 636:n 590:n 586:2 502:( 488:e 481:t 474:v 27:. 20:.

Index

Analogical model
Analogy (disambiguation)
a series
Linguistics
Outline
History
Index
Diachronic
Lexicography
Morphology
Phonology
Pragmatics
Semantics
Syntax
Syntax–semantics interface
Typology
Acquisition
Anthropological
Applied
Computational
Conversation analysis
Corpus linguistics
Discourse analysis
Distance
Documentation
Ethnography of communication
Ethnomethodology
Forensic
History of linguistics
Interlinguistics

Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.

↑