Knowledge

Error-driven learning

Source đź“ť

1710: 1431:
In the context of error-driven learning, the computer vision model learns from the mistakes it makes during the interpretation process. When an error is encountered, the model updates its internal parameters to avoid making the same mistake in the future. This repeated process of learning from errors
1435:
For NLP to do well at computer vision, it employs deep learning techniques. This form of computer vision is sometimes called neural computer vision (NCV), since it makes use of neural networks. NCV therefore interprets visual data based on a statistical, trial and error approach and can deal with
1589:
In the context of error-driven learning, the dialogue system learns from the mistakes it makes during the dialogue process. When an error is encountered, the model updates its internal parameters to avoid making the same mistake in the future. This iterative process of learning from errors helps
1556:
Machine translation is a complex task that involves converting text from one language to another. In the context of error-driven learning, the machine translation model learns from the mistakes it makes during the translation process. When an error is encountered, the model updates its internal
1499:
In the context of error-driven learning, the parser learns from the mistakes it makes during the parsing process. When an error is encountered, the parser updates its internal model to avoid making the same mistake in the future. This iterative process of learning from errors helps improve the
1571:
Speech recognition is a complex task that involves converting spoken language into written text. In the context of error-driven learning, the speech recognition model learns from the mistakes it makes during the recognition process. When an error is encountered, the model updates its internal
1518:
NER is the task of identifying and classifying entities (such as persons, locations, organizations, etc.) in a text. Error-driven learning can help the model learn from its false positives and false negatives and improve its recall and precision on (NER).
1461:
Part-of-speech (POS) tagging is a crucial component in Natural Language Processing (NLP). It helps resolve human language ambiguity at different analysis levels. In addition, its output (tagged data) can be used in various applications of NLP such as
1332:
and the expected output of a system to regulate the system's parameters. Typically applied in supervised learning, these algorithms are provided with a collection of input-output pairs to facilitate the process of generalization.
1537:
This is where the role of NER becomes crucial in error-driven learning. By accurately recognizing and classifying entities, it can help minimize these errors and improve the overall accuracy of the learning process. Furthermore,
1674:
expensive and time-consuming, especially for nonlinear and deep models, as they require multiple iterations(repetitions) and calculations to update the weights of the system. This can be alleviated by using parallel and
1371:
Simpler error-driven learning models effectively capture complex human cognitive phenomena and anticipate elusive behaviors. They provide a flexible mechanism for modeling the brain's learning process, encompassing
1503:
In conclusion, error-driven learning plays a crucial role in improving the accuracy and efficiency of NLP parsers by allowing them to learn from their mistakes and adapt their internal models accordingly.
1022:
Error-driven learning models are ones that rely on the feedback of prediction errors to adjust the expectations or parameters of a model. The key components of error-driven learning include the following:
963:
based on the difference between the proposed and actual results. These models stand out as they depend on environmental feedback instead of explicit labels or categories. They are based on the idea that
1712:
Scalability of Networks and Services: Third International Conference on Autonomous Infrastructure, Management and Security, AIMS 2009 Enschede, The Netherlands, June 30 - July 2, 2009, Proceedings
1542:-based NER methods have shown to be more accurate as they are capable of assembling words, enabling them to understand the semantic and syntactic relationship between various words better. 1388:. By using errors as guiding signals, these algorithms adeptly adapt to changing environmental demands and objectives, capturing statistical regularities and structure. 1586:
Dialogue systems are a popular NLP task as they have promising real-life applications. They are also complicated tasks since many NLP tasks deserving study are involved.
1271: 1190: 1109: 866: 904: 1313: 1292: 1230: 1210: 1149: 1129: 1068: 1045: 1909:
Iosif, Elias; Klasinas, Ioannis; Athanasopoulou, Georgia; Palogiannidi, Elisavet; Georgiladakis, Spiros; Louka, Katerina; Potamianos, Alexandros (2018-01-01).
1665:
and the quality of the solution. This requires careful tuning and experimentation, or using adaptive methods that adjust the hyperparameters automatically.
1572:
parameters to avoid making the same mistake in the future. This iterative process of learning from errors helps improve the model’s performance over time.
1557:
parameters to avoid making the same mistake in the future. This iterative process of learning from errors helps improve the model’s performance over time.
861: 851: 972:(MPSE). By leveraging these prediction errors, the models consistently refine expectations and decrease computational complexity. Typically, these 692: 1407:, follow the principles and constraints of the brain and nervous system. Their primary aim is to capture the emergent properties and dynamics of 899: 856: 707: 1522:
In the context of error-driven learning, the significance of NER is quite profound. Traditional sequence labeling methods identify nested
1391:
Furthermore, cognitive science has led to the creation of new error-driven learning algorithms that are both biologically acceptable and
438: 1615:
They can achieve high accuracy and performance, as they can learn complex and nonlinear relationships between the input and the output.
939: 742: 1632:, which means that they memorize the training data and fail to generalize to new and unseen data. This can be mitigated by using 818: 1602:
They can learn from feedback and correct their mistakes, which makes them adaptive and robust to noise and changes in the data.
367: 1720: 2250: 1860:" Proceedings of the Eighth Conference on Computational Natural Language Learning (CoNLL-2004) at HLT-NAACL 2004. 2004. APA 876: 639: 174: 1811:"Biologically Plausible Error-Driven Learning Using Local Activation Differences: The Generalized Recirculation Algorithm" 894: 727: 702: 651: 775: 770: 423: 433: 71: 1751:"An exploration of error-driven learning in simple two-layer networks from a discriminative learning perspective" 1662: 2026:"Using error decay prediction to overcome practical issues of deep active learning for named entity recognition" 1947:
Bengio, Y. (2009). Learning deep architectures for AI. Foundations and trends® in Machine Learning, 2(1), 1-127
1894:
Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
932: 828: 592: 413: 2137:
Tan, Zhixing; Wang, Shuo; Yang, Zonghan; Chen, Gang; Huang, Xuancheng; Sun, Maosong; Liu, Yang (2020-01-01).
969: 803: 505: 281: 1633: 760: 697: 607: 585: 428: 418: 1962:
Voulodimos, Athanasios; Doulamis, Nikolaos; Doulamis, Anastasios; Protopapadakis, Eftychios (2018-02-01).
1352:
sequences. Many other error-driven learning algorithms are derived from alternative versions of GeneRec.
1624:
Although error driven learning has its advantages, their algorithms also have the following limitations:
1445: 988: 911: 823: 808: 269: 91: 2188: 871: 798: 548: 443: 231: 164: 124: 1910: 1428:
is a complex task that involves understanding and interpreting visual data, such as images or videos.
2207:"Analysis of error-based machine learning algorithms in network anomaly detection and categorization" 925: 531: 299: 169: 1911:"Speech understanding for spoken dialogue systems: From corpus harvesting to grammar rule induction" 1513: 999: 553: 473: 396: 314: 144: 106: 101: 61: 56: 1612:, as they do not require explicit feature engineering or prior knowledge of the data distribution. 500: 349: 249: 76: 2024:
Chang, Haw-Shiuan; Vembu, Shankar; Mohan, Sunil; Uppaal, Rheeya; McCallum, Andrew (2020-09-01).
1671: 1487: 1463: 1456: 1400: 1325: 992: 956: 680: 656: 558: 319: 294: 254: 66: 1876:." Proceedings of the seventh conference on Natural language learning at HLT-NAACL 2003. 2003. 1598:
Error-driven learning has several advantages over other types of machine learning algorithms:
2193:
2023 10th International Conference on Computing for Sustainable Global Development (INDIACom)
1676: 1530:, it can lead to incorrect identification of the outer entity, leading to a problem known as 1496:) based on grammar rules. If a sentence cannot be parsed, it may contain grammatical errors. 1467: 634: 456: 408: 264: 179: 51: 1241: 1160: 1079: 965: 563: 513: 2080:"Research on Named Entity Recognition Based on Multi-Task Learning and Biaffine Mechanism" 8: 1551: 1404: 1396: 1003: 666: 602: 573: 478: 304: 237: 223: 209: 184: 134: 86: 46: 2150: 2114: 2079: 2037: 1998: 1963: 1783: 1749:
Hoppe, Dorothée B.; Hendriks, Petra; Ramscar, Michael; van Rij, Jacolien (2022-10-01).
1566: 1475: 1471: 1298: 1277: 1215: 1195: 1134: 1114: 1053: 1030: 1007: 644: 568: 354: 149: 2226: 2170: 2119: 2101: 2057: 2003: 1985: 1930: 1830: 1788: 1770: 1716: 1689: 1531: 1366: 980: 737: 580: 493: 289: 259: 204: 199: 154: 96: 2218: 2160: 2109: 2091: 2047: 1993: 1975: 1922: 1822: 1778: 1762: 1581: 765: 518: 468: 378: 362: 332: 194: 189: 139: 129: 27: 1858:
Combining lexical and syntactic features for supervised word sense disambiguation.
2165: 2138: 1425: 1420: 1385: 1337: 1011: 984: 793: 597: 463: 403: 2222: 2052: 2025: 1766: 1658: 1650: 1408: 1392: 1047:
of states representing the different situations that the learner can encounter.
813: 344: 81: 2078:
Gao, Wenchao; Li, Yu; Guan, Xiaole; Chen, Shiyu; Zhao, Shanshan (2022-08-25).
1926: 1826: 2244: 2230: 2206: 2174: 2105: 2061: 1989: 1934: 1834: 1774: 1750: 1654: 1637: 1539: 732: 661: 543: 274: 159: 1810: 1111:
that gives the learner’s current prediction of the outcome of taking action
2123: 2096: 2007: 1980: 1792: 1329: 1629: 538: 32: 1961: 1889: 1873: 1857: 2139:"Neural machine translation: A review of methods, resources, and tools" 1908: 1641: 1478:, text-to-speech conversion, partial parsing, and grammar correction. 1373: 687: 383: 309: 1606: 1377: 973: 960: 846: 627: 2155: 2042: 1609: 1890:
Grammatical error correction: Machine translation and classifiers.
1526:
layer by layer. If an error occurs in the recognition of an inner
1492:
Parsing in NLP involves breaking down a text into smaller pieces (
1341: 996: 622: 2205:
Ajila, Samuel A.; Lung, Chung-Horng; Das, Anurag (2022-06-01).
1527: 1523: 1493: 1381: 1344:, a generalized recirculation algorithm primarily employed for 373: 617: 612: 339: 1748: 1345: 987:. These methods have also found successful application in 1349: 1328:
algorithms that leverage the disparity between the real
1324:
Error-driven learning algorithms refer to a category of
905:
List of datasets in computer vision and image processing
1874:
Named entity recognition through classifier combination
2189:
NLP & AI Speech Recognition: An Analytical Review
2023: 1679:, or using specialized hardware such as GPUs or TPUs. 1301: 1280: 1244: 1218: 1198: 1163: 1137: 1117: 1082: 1056: 1033: 979:
Error-driven learning has widespread applications in
1964:"Deep Learning for Computer Vision: A Brief Review" 1070:
of actions that the learner can take in each state.
1307: 1286: 1265: 1224: 1204: 1184: 1143: 1123: 1103: 1062: 1039: 1636:techniques, such as adding a penalty term to the 1507: 1432:helps improve the model’s performance over time. 2242: 2187:A. Thakur, L. Ahuja, R. Vashisth and R. Simon, " 2136: 1657:, the initialization of the weights, and other 1439: 900:List of datasets for machine-learning research 2077: 1436:context and other subtleties of visual data. 933: 2204: 1808: 2084:Computational Intelligence and Neuroscience 1968:Computational Intelligence and Neuroscience 1649:They can be sensitive to the choice of the 1590:improve the model’s performance over time. 940: 926: 2164: 2154: 2113: 2095: 2051: 2041: 1997: 1979: 1782: 1708: 1450: 2195:, New Delhi, India, 2023, pp. 1390-1396. 1709:Sadre, Ramin; Pras, Aiko (2009-06-19). 976:are operated by the GeneRec algorithm. 2243: 1545: 2073: 2071: 2019: 2017: 1560: 959:method. This method tweaks a model’s 1957: 1955: 1953: 1904: 1902: 1884: 1882: 1868: 1866: 1852: 1850: 1848: 1846: 1844: 1804: 1802: 1744: 1742: 1740: 1738: 1736: 1734: 1732: 1360: 1017: 1856:Mohammad, Saif, and Ted Pedersen. " 1809:O'Reilly, Randall C. (1996-07-01). 1575: 895:Glossary of artificial intelligence 13: 2068: 2014: 1414: 14: 2262: 1950: 1899: 1888:Rozovskaya, Alla, and Dan Roth. " 1879: 1863: 1841: 1799: 1729: 1702: 1192:that compares the actual outcome 968:involves the minimization of the 1500:parser’s performance over time. 2198: 2181: 2130: 1355: 1340:learning algorithm is known as 1941: 1915:Computer Speech & Language 1619: 1508:Named entity recognition (NER) 1395:. These algorithms, including 1260: 1248: 1179: 1167: 1098: 1086: 315:Relevance vector machine (RVM) 1: 1695: 1593: 1319: 804:Computational learning theory 368:Expectation–maximization (EM) 2211:Annals of Telecommunications 2166:10.1016/j.aiopen.2020.11.001 1273:that adjusts the prediction 1232:and produces an error value. 991:(NLP), including areas like 761:Coefficient of determination 608:Convolutional neural network 320:Support vector machine (SVM) 16:Subfield of machine learning 7: 2251:Machine learning algorithms 1683: 1446:Natural language processing 1440:Natural Language Processing 989:natural language processing 912:Outline of machine learning 809:Empirical risk minimization 10: 2267: 2223:10.1007/s12243-021-00836-0 2053:10.1007/s10994-020-05897-1 1767:10.3758/s13428-021-01711-5 1605:They can handle large and 1579: 1564: 1549: 1511: 1485: 1481: 1454: 1443: 1418: 1364: 1336:The widely utilized error 549:Feedforward neural network 300:Artificial neural networks 1927:10.1016/j.csl.2017.08.002 1827:10.1162/neco.1996.8.5.895 1755:Behavior Research Methods 1393:computationally efficient 532:Artificial neural network 1514:Named-entity recognition 1000:named entity recognition 841:Journals and conferences 788:Mathematical foundations 698:Temporal difference (TD) 554:Recurrent neural network 474:Conditional random field 397:Dimensionality reduction 145:Dimensionality reduction 107:Quantum machine learning 102:Neuromorphic engineering 62:Self-supervised learning 57:Semi-supervised learning 1872:Florian, Radu, et al. " 1661:, which can affect the 1401:spiking neural networks 250:Apprenticeship learning 1488:Part-of-speech tagging 1464:information extraction 1457:Part-of-speech tagging 1451:Part-of-speech tagging 1326:reinforcement learning 1309: 1295:in light of the error 1288: 1267: 1266:{\displaystyle U(p,e)} 1226: 1206: 1186: 1185:{\displaystyle E(o,p)} 1145: 1125: 1105: 1104:{\displaystyle P(s,a)} 1064: 1041: 993:part-of-speech tagging 957:reinforcement learning 799:Bias–variance tradeoff 681:Reinforcement learning 657:Spiking neural network 67:Reinforcement learning 1677:distributed computing 1628:They can suffer from 1468:information retrieval 1310: 1289: 1268: 1227: 1207: 1187: 1146: 1126: 1106: 1065: 1042: 953:Error-driven learning 635:Neural radiance field 457:Structured prediction 180:Structured prediction 52:Unsupervised learning 2097:10.1155/2022/2687615 1981:10.1155/2018/7068349 1534:of nested entities. 1397:deep belief networks 1299: 1278: 1242: 1216: 1212:with the prediction 1196: 1161: 1135: 1115: 1080: 1054: 1031: 966:language acquisition 824:Statistical learning 722:Learning with humans 514:Local outlier factor 1552:Machine translation 1546:Machine translation 1405:reservoir computing 1075:prediction function 1004:machine translation 667:Electrochemical RAM 574:reservoir computing 305:Logistic regression 224:Supervised learning 210:Multimodal learning 185:Feature engineering 130:Generative modeling 92:Rule-based learning 87:Curriculum learning 47:Supervised learning 22:Part of a series on 1815:Neural Computation 1640:, or reducing the 1567:Speech recognition 1561:Speech recognition 1476:speech eecognition 1472:question Answering 1305: 1284: 1263: 1222: 1202: 1182: 1141: 1121: 1101: 1060: 1037: 1008:speech recognition 981:cognitive sciences 235: • 150:Density estimation 1722:978-3-642-02627-0 1690:Predictive coding 1532:error propagation 1367:Cognitive science 1361:Cognitive science 1308:{\displaystyle e} 1287:{\displaystyle p} 1225:{\displaystyle p} 1205:{\displaystyle o} 1144:{\displaystyle s} 1124:{\displaystyle a} 1063:{\displaystyle A} 1040:{\displaystyle S} 1018:Formal Definition 950: 949: 755:Model diagnostics 738:Human-in-the-loop 581:Boltzmann machine 494:Anomaly detection 290:Linear regression 205:Ontology learning 200:Grammar induction 175:Semantic analysis 170:Association rules 155:Anomaly detection 97:Neuro-symbolic AI 2258: 2235: 2234: 2202: 2196: 2185: 2179: 2178: 2168: 2158: 2134: 2128: 2127: 2117: 2099: 2075: 2066: 2065: 2055: 2045: 2036:(9): 1749–1778. 2030:Machine Learning 2021: 2012: 2011: 2001: 1983: 1959: 1948: 1945: 1939: 1938: 1906: 1897: 1886: 1877: 1870: 1861: 1854: 1839: 1838: 1806: 1797: 1796: 1786: 1761:(5): 2221–2251. 1746: 1727: 1726: 1706: 1607:high-dimensional 1582:Dialogue systems 1576:Dialogue systems 1314: 1312: 1311: 1306: 1293: 1291: 1290: 1285: 1272: 1270: 1269: 1264: 1231: 1229: 1228: 1223: 1211: 1209: 1208: 1203: 1191: 1189: 1188: 1183: 1150: 1148: 1147: 1142: 1130: 1128: 1127: 1122: 1110: 1108: 1107: 1102: 1069: 1067: 1066: 1061: 1046: 1044: 1043: 1038: 1012:dialogue systems 970:prediction error 942: 935: 928: 889:Related articles 766:Confusion matrix 519:Isolation forest 464:Graphical models 243: 242: 195:Learning to rank 190:Feature learning 28:Machine learning 19: 18: 2266: 2265: 2261: 2260: 2259: 2257: 2256: 2255: 2241: 2240: 2239: 2238: 2203: 2199: 2186: 2182: 2135: 2131: 2076: 2069: 2022: 2015: 1960: 1951: 1946: 1942: 1907: 1900: 1887: 1880: 1871: 1864: 1855: 1842: 1807: 1800: 1747: 1730: 1723: 1707: 1703: 1698: 1686: 1672:computationally 1659:hyperparameters 1622: 1596: 1584: 1578: 1569: 1563: 1554: 1548: 1516: 1510: 1490: 1484: 1459: 1453: 1448: 1442: 1426:Computer vision 1423: 1421:Computer vision 1417: 1415:Computer vision 1409:neural circuits 1386:decision-making 1369: 1363: 1358: 1338:backpropagation 1322: 1300: 1297: 1296: 1279: 1276: 1275: 1243: 1240: 1239: 1217: 1214: 1213: 1197: 1194: 1193: 1162: 1159: 1158: 1136: 1133: 1132: 1116: 1113: 1112: 1081: 1078: 1077: 1055: 1052: 1051: 1032: 1029: 1028: 1020: 985:computer vision 946: 917: 916: 890: 882: 881: 842: 834: 833: 794:Kernel machines 789: 781: 780: 756: 748: 747: 728:Active learning 723: 715: 714: 683: 673: 672: 598:Diffusion model 534: 524: 523: 496: 486: 485: 459: 449: 448: 404:Factor analysis 399: 389: 388: 372: 335: 325: 324: 245: 244: 228: 227: 226: 215: 214: 120: 112: 111: 77:Online learning 42: 30: 17: 12: 11: 5: 2264: 2254: 2253: 2237: 2236: 2217:(5): 359–370. 2197: 2180: 2129: 2067: 2013: 1949: 1940: 1898: 1878: 1862: 1840: 1821:(5): 895–938. 1798: 1728: 1721: 1700: 1699: 1697: 1694: 1693: 1692: 1685: 1682: 1681: 1680: 1667: 1666: 1651:error function 1646: 1645: 1634:regularization 1621: 1618: 1617: 1616: 1613: 1603: 1595: 1592: 1577: 1574: 1562: 1559: 1547: 1544: 1509: 1506: 1483: 1480: 1452: 1449: 1441: 1438: 1416: 1413: 1362: 1359: 1357: 1354: 1348:prediction in 1321: 1318: 1317: 1316: 1304: 1283: 1262: 1259: 1256: 1253: 1250: 1247: 1233: 1221: 1201: 1181: 1178: 1175: 1172: 1169: 1166: 1156:error function 1152: 1140: 1120: 1100: 1097: 1094: 1091: 1088: 1085: 1071: 1059: 1048: 1036: 1019: 1016: 948: 947: 945: 944: 937: 930: 922: 919: 918: 915: 914: 909: 908: 907: 897: 891: 888: 887: 884: 883: 880: 879: 874: 869: 864: 859: 854: 849: 843: 840: 839: 836: 835: 832: 831: 826: 821: 816: 814:Occam learning 811: 806: 801: 796: 790: 787: 786: 783: 782: 779: 778: 773: 771:Learning curve 768: 763: 757: 754: 753: 750: 749: 746: 745: 740: 735: 730: 724: 721: 720: 717: 716: 713: 712: 711: 710: 700: 695: 690: 684: 679: 678: 675: 674: 671: 670: 664: 659: 654: 649: 648: 647: 637: 632: 631: 630: 625: 620: 615: 605: 600: 595: 590: 589: 588: 578: 577: 576: 571: 566: 561: 551: 546: 541: 535: 530: 529: 526: 525: 522: 521: 516: 511: 503: 497: 492: 491: 488: 487: 484: 483: 482: 481: 476: 471: 460: 455: 454: 451: 450: 447: 446: 441: 436: 431: 426: 421: 416: 411: 406: 400: 395: 394: 391: 390: 387: 386: 381: 376: 370: 365: 360: 352: 347: 342: 336: 331: 330: 327: 326: 323: 322: 317: 312: 307: 302: 297: 292: 287: 279: 278: 277: 272: 267: 257: 255:Decision trees 252: 246: 232:classification 222: 221: 220: 217: 216: 213: 212: 207: 202: 197: 192: 187: 182: 177: 172: 167: 162: 157: 152: 147: 142: 137: 132: 127: 125:Classification 121: 118: 117: 114: 113: 110: 109: 104: 99: 94: 89: 84: 82:Batch learning 79: 74: 69: 64: 59: 54: 49: 43: 40: 39: 36: 35: 24: 23: 15: 9: 6: 4: 3: 2: 2263: 2252: 2249: 2248: 2246: 2232: 2228: 2224: 2220: 2216: 2212: 2208: 2201: 2194: 2190: 2184: 2176: 2172: 2167: 2162: 2157: 2152: 2148: 2144: 2140: 2133: 2125: 2121: 2116: 2111: 2107: 2103: 2098: 2093: 2089: 2085: 2081: 2074: 2072: 2063: 2059: 2054: 2049: 2044: 2039: 2035: 2031: 2027: 2020: 2018: 2009: 2005: 2000: 1995: 1991: 1987: 1982: 1977: 1973: 1969: 1965: 1958: 1956: 1954: 1944: 1936: 1932: 1928: 1924: 1920: 1916: 1912: 1905: 1903: 1895: 1891: 1885: 1883: 1875: 1869: 1867: 1859: 1853: 1851: 1849: 1847: 1845: 1836: 1832: 1828: 1824: 1820: 1816: 1812: 1805: 1803: 1794: 1790: 1785: 1780: 1776: 1772: 1768: 1764: 1760: 1756: 1752: 1745: 1743: 1741: 1739: 1737: 1735: 1733: 1724: 1718: 1714: 1713: 1705: 1701: 1691: 1688: 1687: 1678: 1673: 1669: 1668: 1664: 1660: 1656: 1655:learning rate 1652: 1648: 1647: 1644:of the model. 1643: 1639: 1638:loss function 1635: 1631: 1627: 1626: 1625: 1614: 1611: 1608: 1604: 1601: 1600: 1599: 1591: 1587: 1583: 1573: 1568: 1558: 1553: 1543: 1541: 1540:deep learning 1535: 1533: 1529: 1525: 1520: 1515: 1505: 1501: 1497: 1495: 1489: 1479: 1477: 1473: 1469: 1465: 1458: 1447: 1437: 1433: 1429: 1427: 1422: 1412: 1411:and systems. 1410: 1406: 1402: 1398: 1394: 1389: 1387: 1383: 1379: 1375: 1368: 1353: 1351: 1347: 1343: 1339: 1334: 1331: 1327: 1302: 1294: 1281: 1257: 1254: 1251: 1245: 1238: 1234: 1219: 1199: 1176: 1173: 1170: 1164: 1157: 1153: 1138: 1118: 1095: 1092: 1089: 1083: 1076: 1072: 1057: 1049: 1034: 1026: 1025: 1024: 1015: 1013: 1009: 1005: 1001: 998: 994: 990: 986: 982: 977: 975: 971: 967: 962: 958: 955:is a type of 954: 943: 938: 936: 931: 929: 924: 923: 921: 920: 913: 910: 906: 903: 902: 901: 898: 896: 893: 892: 886: 885: 878: 875: 873: 870: 868: 865: 863: 860: 858: 855: 853: 850: 848: 845: 844: 838: 837: 830: 827: 825: 822: 820: 817: 815: 812: 810: 807: 805: 802: 800: 797: 795: 792: 791: 785: 784: 777: 774: 772: 769: 767: 764: 762: 759: 758: 752: 751: 744: 741: 739: 736: 734: 733:Crowdsourcing 731: 729: 726: 725: 719: 718: 709: 706: 705: 704: 701: 699: 696: 694: 691: 689: 686: 685: 682: 677: 676: 668: 665: 663: 662:Memtransistor 660: 658: 655: 653: 650: 646: 643: 642: 641: 638: 636: 633: 629: 626: 624: 621: 619: 616: 614: 611: 610: 609: 606: 604: 601: 599: 596: 594: 591: 587: 584: 583: 582: 579: 575: 572: 570: 567: 565: 562: 560: 557: 556: 555: 552: 550: 547: 545: 544:Deep learning 542: 540: 537: 536: 533: 528: 527: 520: 517: 515: 512: 510: 508: 504: 502: 499: 498: 495: 490: 489: 480: 479:Hidden Markov 477: 475: 472: 470: 467: 466: 465: 462: 461: 458: 453: 452: 445: 442: 440: 437: 435: 432: 430: 427: 425: 422: 420: 417: 415: 412: 410: 407: 405: 402: 401: 398: 393: 392: 385: 382: 380: 377: 375: 371: 369: 366: 364: 361: 359: 357: 353: 351: 348: 346: 343: 341: 338: 337: 334: 329: 328: 321: 318: 316: 313: 311: 308: 306: 303: 301: 298: 296: 293: 291: 288: 286: 284: 280: 276: 275:Random forest 273: 271: 268: 266: 263: 262: 261: 258: 256: 253: 251: 248: 247: 240: 239: 234: 233: 225: 219: 218: 211: 208: 206: 203: 201: 198: 196: 193: 191: 188: 186: 183: 181: 178: 176: 173: 171: 168: 166: 163: 161: 160:Data cleaning 158: 156: 153: 151: 148: 146: 143: 141: 138: 136: 133: 131: 128: 126: 123: 122: 116: 115: 108: 105: 103: 100: 98: 95: 93: 90: 88: 85: 83: 80: 78: 75: 73: 72:Meta-learning 70: 68: 65: 63: 60: 58: 55: 53: 50: 48: 45: 44: 38: 37: 34: 29: 26: 25: 21: 20: 2214: 2210: 2200: 2192: 2183: 2146: 2142: 2132: 2090:: e2687615. 2087: 2083: 2033: 2029: 1974:: e7068349. 1971: 1967: 1943: 1918: 1914: 1893: 1818: 1814: 1758: 1754: 1715:. Springer. 1711: 1704: 1670:They can be 1623: 1597: 1588: 1585: 1570: 1555: 1536: 1521: 1517: 1502: 1498: 1491: 1460: 1434: 1430: 1424: 1390: 1370: 1356:Applications 1335: 1323: 1274: 1236: 1155: 1074: 1021: 978: 952: 951: 819:PAC learning 506: 355: 350:Hierarchical 282: 236: 230: 1921:: 272–297. 1663:convergence 1630:overfitting 1620:Limitations 1237:update rule 703:Multi-agent 640:Transformer 539:Autoencoder 295:Naive Bayes 33:data mining 2156:2012.15515 2043:1911.07335 1696:References 1642:complexity 1594:Advantages 1580:See also: 1565:See also: 1550:See also: 1512:See also: 1486:See also: 1455:See also: 1444:See also: 1419:See also: 1374:perception 1365:See also: 1320:Algorithms 974:algorithms 961:parameters 688:Q-learning 586:Restricted 384:Mean shift 333:Clustering 310:Perceptron 238:regression 140:Clustering 135:Regression 2231:1958-9395 2175:2666-6510 2106:1687-5265 2062:1573-0565 1990:1687-5265 1935:0885-2308 1835:0899-7667 1775:1554-3528 1610:data sets 1378:attention 1131:in state 1010:(SR) and 847:ECML PKDD 829:VC theory 776:ROC curve 708:Self-play 628:DeepDream 469:Bayes net 260:Ensembles 41:Paradigms 2245:Category 2149:: 5–21. 2124:36059424 2008:29487619 1793:35032022 1684:See also 1524:entities 270:Boosting 119:Problems 2143:AI Open 2115:9436550 1999:5816885 1896:. 2016. 1784:9579095 1494:phrases 1482:Parsing 1342:GeneRec 1002:(NER), 997:parsing 852:NeurIPS 669:(ECRAM) 623:AlexNet 265:Bagging 2229:  2173:  2122:  2112:  2104:  2060:  2006:  1996:  1988:  1933:  1833:  1791:  1781:  1773:  1719:  1653:, the 1528:entity 1403:, and 1384:, and 1382:memory 1330:output 1050:A set 1027:A set 1006:(MT), 645:Vision 501:RANSAC 379:OPTICS 374:DBSCAN 358:-means 165:AutoML 2151:arXiv 2038:arXiv 1014:. 867:IJCAI 693:SARSA 652:Mamba 618:LeNet 613:U-Net 439:t-SNE 363:Fuzzy 340:BIRCH 2227:ISSN 2171:ISSN 2120:PMID 2102:ISSN 2088:2022 2058:ISSN 2004:PMID 1986:ISSN 1972:2018 1931:ISSN 1831:ISSN 1789:PMID 1771:ISSN 1717:ISBN 1346:gene 983:and 877:JMLR 862:ICLR 857:ICML 743:RLHF 559:LSTM 345:CURE 31:and 2219:doi 2191:," 2161:doi 2110:PMC 2092:doi 2048:doi 2034:109 1994:PMC 1976:doi 1923:doi 1823:doi 1779:PMC 1763:doi 1350:DNA 1235:An 1154:An 603:SOM 593:GAN 569:ESN 564:GRU 509:-NN 444:SDL 434:PGD 429:PCA 424:NMF 419:LDA 414:ICA 409:CCA 285:-NN 2247:: 2225:. 2215:77 2213:. 2209:. 2169:. 2159:. 2145:. 2141:. 2118:. 2108:. 2100:. 2086:. 2082:. 2070:^ 2056:. 2046:. 2032:. 2028:. 2016:^ 2002:. 1992:. 1984:. 1970:. 1966:. 1952:^ 1929:. 1919:47 1917:. 1913:. 1901:^ 1892:" 1881:^ 1865:^ 1843:^ 1829:. 1817:. 1813:. 1801:^ 1787:. 1777:. 1769:. 1759:54 1757:. 1753:. 1731:^ 1474:, 1470:, 1466:, 1399:, 1380:, 1376:, 1073:A 995:, 872:ML 2233:. 2221:: 2177:. 2163:: 2153:: 2147:1 2126:. 2094:: 2064:. 2050:: 2040:: 2010:. 1978:: 1937:. 1925:: 1837:. 1825:: 1819:8 1795:. 1765:: 1725:. 1315:. 1303:e 1282:p 1261:) 1258:e 1255:, 1252:p 1249:( 1246:U 1220:p 1200:o 1180:) 1177:p 1174:, 1171:o 1168:( 1165:E 1151:. 1139:s 1119:a 1099:) 1096:a 1093:, 1090:s 1087:( 1084:P 1058:A 1035:S 941:e 934:t 927:v 507:k 356:k 283:k 241:) 229:(

Index

Machine learning
data mining
Supervised learning
Unsupervised learning
Semi-supervised learning
Self-supervised learning
Reinforcement learning
Meta-learning
Online learning
Batch learning
Curriculum learning
Rule-based learning
Neuro-symbolic AI
Neuromorphic engineering
Quantum machine learning
Classification
Generative modeling
Regression
Clustering
Dimensionality reduction
Density estimation
Anomaly detection
Data cleaning
AutoML
Association rules
Semantic analysis
Structured prediction
Feature engineering
Feature learning
Learning to rank

Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.

↑