Knowledge (XXG)

Averaged one-dependence estimators

Source 📝

507: 81: 1301: 502:{\displaystyle {\hat {P}}(y\mid x_{1},\ldots x_{n})={\frac {\sum _{i:1\leq i\leq n\wedge F(x_{i})\geq m}{\hat {P}}(y,x_{i})\prod _{j=1}^{n}{\hat {P}}(x_{j}\mid y,x_{i})}{\sum _{y^{\prime }\in Y}\sum _{i:1\leq i\leq n\wedge F(x_{i})\geq m}{\hat {P}}(y^{\prime },x_{i})\prod _{j=1}^{n}{\hat {P}}(x_{j}\mid y^{\prime },x_{i})}}} 1160:
that makes the above independence assumption that is weaker (and hence potentially less harmful) than the naive Bayes' independence assumption. In consequence, each ODE should create a less biased estimator than naive Bayes. However, because the base probability estimates are each conditioned by two
1173:
whereby the classifier can be updated efficiently with information from new examples as they become available. It predicts class probabilities rather than simply predicting a single class, allowing the user to determine the confidence with which each classification can be made. Its probabilistic
1267:
is the number of classes. This makes it infeasible for application to high-dimensional data. However, within that limitation, it is linear with respect to the number of training examples and hence can efficiently process large numbers of training examples.
791: 1151: 1161:
variables rather than one, they are formed from less data (the training examples that satisfy both variables) and hence are likely to have more variance. AODE reduces this variance by averaging the estimates of all such ODEs.
972: 548: 1318: 826: 648: 1253: 1214: 606: 577: 1008: 834: 31:. It frequently develops substantially more accurate classifiers than naive Bayes at the cost of a modest increase in the amount of computation. 1365: 1337: 1169:
Like naive Bayes, AODE does not perform model selection and does not use tuneable parameters. As a result, it has low variance. It supports
1344: 1451: 1441: 1384: 1351: 612:
is a user specified minimum frequency with which a term must appear in order to be used in the outer summation. In recent practice
1333: 1322: 1446: 1311: 1289: 1358: 24: 1277: 1157: 28: 515: 1170: 786:{\displaystyle P(y\mid x_{1},\ldots x_{n})={\frac {P(y,x_{1},\ldots x_{n})}{P(x_{1},\ldots x_{n})}}.} 799: 1219: 1180: 582: 553: 1419: 27:
technique. It was developed to address the attribute-independence problem of the popular
1156:
This formula defines a special form of One Dependence Estimator (ODE), a variant of the
1146:{\displaystyle P(y,x_{1},\ldots x_{n})=P(y,x_{i})\prod _{j=1}^{n}P(x_{j}\mid y,x_{i}).} 1435: 967:{\displaystyle P(y,x_{1},\ldots x_{n})=P(y,x_{i})P(x_{1},\ldots x_{n}\mid y,x_{i}).} 1300: 1423: 1411: 608:
is the frequency with which the argument appears in the sample data and
1174:
model can directly handle situations where some data are missing.
1294: 475: 400: 311: 1412:"Not So Naive Bayes: Aggregating One-Dependence Estimators" 1280:
machine learning suite includes an implementation of AODE.
1222: 1183: 1011: 837: 802: 651: 585: 556: 518: 84: 39:
AODE seeks to estimate the probability of each class
1325:. Unsourced material may be challenged and removed. 1247: 1208: 1145: 966: 820: 785: 600: 571: 542: 501: 642:). By the definition of conditional probability 1410:Webb, G. I., J. Boughton, and Z. Wang (2005). 8: 1385:Learn how and when to remove this message 1236: 1221: 1197: 1182: 1131: 1112: 1096: 1085: 1072: 1044: 1028: 1010: 952: 933: 917: 898: 870: 854: 836: 801: 768: 752: 731: 715: 696: 684: 668: 650: 584: 555: 520: 519: 517: 487: 474: 461: 443: 442: 436: 425: 412: 399: 381: 380: 363: 328: 310: 305: 290: 271: 253: 252: 246: 235: 222: 198: 197: 180: 145: 138: 126: 110: 86: 85: 83: 1403: 1263:is the number of training examples and 7: 1334:"Averaged one-dependence estimators" 1323:adding citations to reliable sources 1177:AODE has computational complexity 543:{\displaystyle {\hat {P}}(\cdot )} 43:given a specified set of features 17:Averaged one-dependence estimators 14: 620:Derivation of the AODE classifier 75:). To do so it uses the formula 1299: 1310:needs additional citations for 1165:Features of the AODE classifier 1255:at classification time, where 1242: 1226: 1203: 1187: 1137: 1105: 1078: 1059: 1050: 1015: 958: 910: 904: 885: 876: 841: 774: 745: 737: 702: 690: 655: 595: 589: 566: 560: 537: 531: 525: 493: 454: 448: 418: 392: 386: 369: 356: 296: 264: 258: 228: 209: 203: 186: 173: 132: 97: 91: 1: 821:{\displaystyle 1\leq i\leq n} 1259:is the number of features, 1468: 1452:Statistical classification 1442:Classification algorithms 1424:10.1007/s10994-005-4258-6 1290:Cluster-weighted modeling 1248:{\displaystyle O(kn^{2})} 1209:{\displaystyle O(ln^{2})} 977:Under an assumption that 601:{\displaystyle F(\cdot )} 572:{\displaystyle P(\cdot )} 550:denotes an estimate of 25:classification learning 1249: 1210: 1158:naive Bayes classifier 1147: 1101: 991:are independent given 968: 822: 787: 624:We seek to estimate P( 602: 573: 544: 503: 441: 251: 29:naive Bayes classifier 1250: 1216:at training time and 1211: 1148: 1081: 969: 823: 788: 616:is usually set at 1. 603: 574: 545: 504: 421: 231: 23:) is a probabilistic 1319:improve this article 1220: 1181: 1171:incremental learning 1009: 835: 800: 649: 583: 554: 516: 82: 1447:Bayesian estimation 35:The AODE classifier 1245: 1206: 1143: 1002:, it follows that 964: 818: 783: 598: 569: 540: 499: 379: 323: 196: 1395: 1394: 1387: 1369: 778: 528: 497: 451: 389: 324: 301: 261: 206: 141: 94: 1459: 1426: 1416:Machine Learning 1408: 1390: 1383: 1379: 1376: 1370: 1368: 1327: 1303: 1295: 1254: 1252: 1251: 1246: 1241: 1240: 1215: 1213: 1212: 1207: 1202: 1201: 1152: 1150: 1149: 1144: 1136: 1135: 1117: 1116: 1100: 1095: 1077: 1076: 1049: 1048: 1033: 1032: 973: 971: 970: 965: 957: 956: 938: 937: 922: 921: 903: 902: 875: 874: 859: 858: 827: 825: 824: 819: 792: 790: 789: 784: 779: 777: 773: 772: 757: 756: 740: 736: 735: 720: 719: 697: 689: 688: 673: 672: 607: 605: 604: 599: 578: 576: 575: 570: 549: 547: 546: 541: 530: 529: 521: 508: 506: 505: 500: 498: 496: 492: 491: 479: 478: 466: 465: 453: 452: 444: 440: 435: 417: 416: 404: 403: 391: 390: 382: 378: 368: 367: 322: 315: 314: 299: 295: 294: 276: 275: 263: 262: 254: 250: 245: 227: 226: 208: 207: 199: 195: 185: 184: 139: 131: 130: 115: 114: 96: 95: 87: 1467: 1466: 1462: 1461: 1460: 1458: 1457: 1456: 1432: 1431: 1430: 1429: 1418:, 58(1), 5–24. 1409: 1405: 1400: 1391: 1380: 1374: 1371: 1328: 1326: 1316: 1304: 1286: 1274: 1272:Implementations 1232: 1218: 1217: 1193: 1179: 1178: 1167: 1127: 1108: 1068: 1040: 1024: 1007: 1006: 1000: 990: 983: 948: 929: 913: 894: 866: 850: 833: 832: 798: 797: 764: 748: 741: 727: 711: 698: 680: 664: 647: 646: 641: 634: 622: 581: 580: 552: 551: 514: 513: 483: 470: 457: 408: 395: 359: 306: 300: 286: 267: 218: 176: 140: 122: 106: 80: 79: 74: 67: 56: 49: 37: 12: 11: 5: 1465: 1463: 1455: 1454: 1449: 1444: 1434: 1433: 1428: 1427: 1402: 1401: 1399: 1396: 1393: 1392: 1307: 1305: 1298: 1293: 1292: 1285: 1282: 1273: 1270: 1244: 1239: 1235: 1231: 1228: 1225: 1205: 1200: 1196: 1192: 1189: 1186: 1166: 1163: 1154: 1153: 1142: 1139: 1134: 1130: 1126: 1123: 1120: 1115: 1111: 1107: 1104: 1099: 1094: 1091: 1088: 1084: 1080: 1075: 1071: 1067: 1064: 1061: 1058: 1055: 1052: 1047: 1043: 1039: 1036: 1031: 1027: 1023: 1020: 1017: 1014: 998: 988: 981: 975: 974: 963: 960: 955: 951: 947: 944: 941: 936: 932: 928: 925: 920: 916: 912: 909: 906: 901: 897: 893: 890: 887: 884: 881: 878: 873: 869: 865: 862: 857: 853: 849: 846: 843: 840: 817: 814: 811: 808: 805: 794: 793: 782: 776: 771: 767: 763: 760: 755: 751: 747: 744: 739: 734: 730: 726: 723: 718: 714: 710: 707: 704: 701: 695: 692: 687: 683: 679: 676: 671: 667: 663: 660: 657: 654: 639: 632: 621: 618: 597: 594: 591: 588: 568: 565: 562: 559: 539: 536: 533: 527: 524: 510: 509: 495: 490: 486: 482: 477: 473: 469: 464: 460: 456: 450: 447: 439: 434: 431: 428: 424: 420: 415: 411: 407: 402: 398: 394: 388: 385: 377: 374: 371: 366: 362: 358: 355: 352: 349: 346: 343: 340: 337: 334: 331: 327: 321: 318: 313: 309: 304: 298: 293: 289: 285: 282: 279: 274: 270: 266: 260: 257: 249: 244: 241: 238: 234: 230: 225: 221: 217: 214: 211: 205: 202: 194: 191: 188: 183: 179: 175: 172: 169: 166: 163: 160: 157: 154: 151: 148: 144: 137: 134: 129: 125: 121: 118: 113: 109: 105: 102: 99: 93: 90: 72: 65: 54: 47: 36: 33: 13: 10: 9: 6: 4: 3: 2: 1464: 1453: 1450: 1448: 1445: 1443: 1440: 1439: 1437: 1425: 1421: 1417: 1413: 1407: 1404: 1397: 1389: 1386: 1378: 1367: 1364: 1360: 1357: 1353: 1350: 1346: 1343: 1339: 1336: –  1335: 1331: 1330:Find sources: 1324: 1320: 1314: 1313: 1308:This article 1306: 1302: 1297: 1296: 1291: 1288: 1287: 1283: 1281: 1279: 1271: 1269: 1266: 1262: 1258: 1237: 1233: 1229: 1223: 1198: 1194: 1190: 1184: 1175: 1172: 1164: 1162: 1159: 1140: 1132: 1128: 1124: 1121: 1118: 1113: 1109: 1102: 1097: 1092: 1089: 1086: 1082: 1073: 1069: 1065: 1062: 1056: 1053: 1045: 1041: 1037: 1034: 1029: 1025: 1021: 1018: 1012: 1005: 1004: 1003: 1001: 994: 987: 980: 961: 953: 949: 945: 942: 939: 934: 930: 926: 923: 918: 914: 907: 899: 895: 891: 888: 882: 879: 871: 867: 863: 860: 855: 851: 847: 844: 838: 831: 830: 829: 815: 812: 809: 806: 803: 780: 769: 765: 761: 758: 753: 749: 742: 732: 728: 724: 721: 716: 712: 708: 705: 699: 693: 685: 681: 677: 674: 669: 665: 661: 658: 652: 645: 644: 643: 638: 631: 627: 619: 617: 615: 611: 592: 586: 563: 557: 534: 522: 488: 484: 480: 471: 467: 462: 458: 445: 437: 432: 429: 426: 422: 413: 409: 405: 396: 383: 375: 372: 364: 360: 353: 350: 347: 344: 341: 338: 335: 332: 329: 325: 319: 316: 307: 302: 291: 287: 283: 280: 277: 272: 268: 255: 247: 242: 239: 236: 232: 223: 219: 215: 212: 200: 192: 189: 181: 177: 170: 167: 164: 161: 158: 155: 152: 149: 146: 142: 135: 127: 123: 119: 116: 111: 107: 103: 100: 88: 78: 77: 76: 71: 64: 60: 53: 46: 42: 34: 32: 30: 26: 22: 18: 1415: 1406: 1381: 1372: 1362: 1355: 1348: 1341: 1329: 1317:Please help 1312:verification 1309: 1275: 1264: 1260: 1256: 1176: 1168: 1155: 996: 992: 985: 978: 976: 795: 636: 629: 625: 623: 613: 609: 511: 69: 62: 58: 51: 44: 40: 38: 20: 16: 15: 1436:Categories 1398:References 1375:March 2011 1345:newspapers 1276:The free 1119:∣ 1083:∏ 1038:… 940:∣ 927:… 864:… 813:≤ 807:≤ 762:… 725:… 678:… 662:∣ 593:⋅ 564:⋅ 535:⋅ 526:^ 476:′ 468:∣ 449:^ 423:∏ 401:′ 387:^ 373:≥ 351:∧ 345:≤ 339:≤ 326:∑ 317:∈ 312:′ 303:∑ 278:∣ 259:^ 233:∏ 204:^ 190:≥ 168:∧ 162:≤ 156:≤ 143:∑ 120:… 104:∣ 92:^ 1284:See also 796:For any 1359:scholar 1361:  1354:  1347:  1340:  1332:  984:, ... 635:, ... 512:where 68:, ... 50:, ... 1366:JSTOR 1352:books 1338:news 1278:Weka 995:and 57:, P( 21:AODE 1420:doi 1321:by 1438:: 1414:. 828:, 628:| 579:, 61:| 1422:: 1388:) 1382:( 1377:) 1373:( 1363:· 1356:· 1349:· 1342:· 1315:. 1265:k 1261:l 1257:n 1243:) 1238:2 1234:n 1230:k 1227:( 1224:O 1204:) 1199:2 1195:n 1191:l 1188:( 1185:O 1141:. 1138:) 1133:i 1129:x 1125:, 1122:y 1114:j 1110:x 1106:( 1103:P 1098:n 1093:1 1090:= 1087:j 1079:) 1074:i 1070:x 1066:, 1063:y 1060:( 1057:P 1054:= 1051:) 1046:n 1042:x 1035:, 1030:1 1026:x 1022:, 1019:y 1016:( 1013:P 999:i 997:x 993:y 989:n 986:x 982:1 979:x 962:. 959:) 954:i 950:x 946:, 943:y 935:n 931:x 924:, 919:1 915:x 911:( 908:P 905:) 900:i 896:x 892:, 889:y 886:( 883:P 880:= 877:) 872:n 868:x 861:, 856:1 852:x 848:, 845:y 842:( 839:P 816:n 810:i 804:1 781:. 775:) 770:n 766:x 759:, 754:1 750:x 746:( 743:P 738:) 733:n 729:x 722:, 717:1 713:x 709:, 706:y 703:( 700:P 694:= 691:) 686:n 682:x 675:, 670:1 666:x 659:y 656:( 653:P 640:n 637:x 633:1 630:x 626:y 614:m 610:m 596:) 590:( 587:F 567:) 561:( 558:P 538:) 532:( 523:P 494:) 489:i 485:x 481:, 472:y 463:j 459:x 455:( 446:P 438:n 433:1 430:= 427:j 419:) 414:i 410:x 406:, 397:y 393:( 384:P 376:m 370:) 365:i 361:x 357:( 354:F 348:n 342:i 336:1 333:: 330:i 320:Y 308:y 297:) 292:i 288:x 284:, 281:y 273:j 269:x 265:( 256:P 248:n 243:1 240:= 237:j 229:) 224:i 220:x 216:, 213:y 210:( 201:P 193:m 187:) 182:i 178:x 174:( 171:F 165:n 159:i 153:1 150:: 147:i 136:= 133:) 128:n 124:x 117:, 112:1 108:x 101:y 98:( 89:P 73:n 70:x 66:1 63:x 59:y 55:n 52:x 48:1 45:x 41:y 19:(

Index

classification learning
naive Bayes classifier
naive Bayes classifier
incremental learning
Weka
Cluster-weighted modeling

verification
improve this article
adding citations to reliable sources
"Averaged one-dependence estimators"
news
newspapers
books
scholar
JSTOR
Learn how and when to remove this message
"Not So Naive Bayes: Aggregating One-Dependence Estimators"
doi
10.1007/s10994-005-4258-6
Categories
Classification algorithms
Bayesian estimation
Statistical classification

Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.