Knowledge

Dependency network (graphical model)

Source 📝

63: 22: 125: 244:
Dependency networks have advantages and disadvantages with respect to Bayesian networks. In particular, they are easier to parameterize from data, as there are efficient algorithms for learning both the structure and probabilities of a dependency network from data. Such algorithms are not available
1585:
has many variables. So, the law of total probability along with the independencies encoded in a dependency network can be used to decompose the inference task into a set of inference tasks on single variables. This approach comes with the advantage that some terms may be obtained by direct lookup,
223:
of a node is the set of parents and children of that node, together with the children's parents. The values of the parents and children of a node evidently give information about that node. However, its children's parents also have to be included in the Markov blanket, because they can be used to
901:
Two important tasks in a dependency network are to learn its structure and probabilities from data. Essentially, the learning algorithm consists of independently performing a probabilistic regression or classification for each variable in the domain. It comes from observation that the local
825: 1037:, we independently estimate its local distribution from data using a classification algorithm, even though it is a distinct method for each variable. Here, we will briefly show how probabilistic decision trees are used to estimate the local distributions. For each variable 253:
Consistent dependency networks and Markov networks have the same representational power. Nonetheless, it is possible to construct non-consistent dependency networks, i.e., dependency networks for which there is no compatible valid
622: 861:. Dependency networks learned using large data sets with large sample sizes will almost always be consistent. A non-consistent network is a network for which there is no joint probability distribution compatible with the pair 245:
for Bayesian networks, for which the problem of determining the optimal structure is NP-hard. Nonetheless, a dependency network may be more difficult to construct using a knowledge-based approach driven by expert-knowledge.
990:, which can be estimated by any number of classification or regression techniques, such as methods using a probabilistic decision tree, a neural network or a probabilistic support-vector machine. Hence, for each variable 2119: 630: 2435: 324: 988: 2364:(CF), which is the task of predicting preferences. Dependency networks are a natural model class on which to base CF predictions, once an algorithm for this task only needs estimation of 1686: 1656: 1881: 2350: 2300: 2258: 2216: 2172: 1919: 1477: 1409: 1280: 35: 1235: 1146: 1626: 508: 1828: 1798: 1768: 1508: 1440: 859: 355: 1536: 2030: 1850: 1730: 1708: 1583: 1561: 1368: 1346: 1324: 1302: 1084: 429: 143: 1946: 1200: 1173: 1111: 1062: 1015: 927: 891: 476: 387: 207:, DNs may contain cycles. Each node is associated to a conditional probability table, which determines the realization of the random variable given its parents. 2008: 1986: 1966: 1035: 449: 407: 513: 1175:, the search algorithm begins with a singleton root node without children. Then, each leaf node in the tree is replaced with a binary split on some variable 41: 2454:
Another class of useful applications for dependency networks is related to data visualization, that is, visualization of predictive relationships.
2450:
Predicting what product a person will buy based on products he or she has already purchased and/or dropped into his or her shopping basket.
820:{\displaystyle p(x_{i}\mid \mathbf {pa_{i}} )=p(x_{i}\mid x_{1},\ldots ,x_{i-1},x_{i+1},\ldots ,x_{n})=p(x_{i}\mid \mathbf {x} -{x_{i}}).} 2036: 179: 161: 106: 84: 49: 893:. In that case, there is no joint probability distribution that satisfies the independence relationships subsumed by that pair. 255: 203:, wherein each vertex (node) corresponds to a random variable and each edge captures dependencies among variables. Unlike 830:
The dependency network is consistent in the sense that each local distribution can be obtained from the joint distribution
2463: 2437:
to produce recommendations. In particular, these estimates may be obtained by a direct lookup in a dependency network.
2367: 269: 932: 1661: 1631: 1442:
is small, then many iterations are required for an accurate probability estimate. Another approach for estimating
2560: 1858: 77: 71: 2360:
In addition to the applications to probabilistic inference, the following applications are in the category of
1374:. A naive approach for this uses an ordered Gibbs sampler, an important difficulty of which is that if either 2361: 2309: 2264: 2222: 2180: 2131: 1889: 1445: 1377: 1248: 88: 2483:
HECKERMAN, David; MAXWELL C., David; MEEK, Christopher; ROUNTHWAITE, Robert; KADIE, Carl (October 2000).
2527: 1205: 1116: 1592: 481: 2506: 1805: 1775: 1745: 1482: 1414: 833: 329: 1513: 2565: 2484: 1245:
A probabilistic inference is the task in which we wish to answer probabilistic queries of the form
225: 2013: 1833: 1713: 1691: 1566: 1544: 1351: 1329: 1307: 1285: 1067: 412: 2513: 2447:
Predicting what news stories a person is interested in based on other stories he or she read;
216: 204: 1924: 1178: 1151: 1089: 1040: 993: 905: 864: 454: 360: 2540: 196: 617:{\displaystyle \mathbf {Pa_{i}} \subseteq (X_{1},\ldots ,X_{i-1},X_{i+1},\ldots ,X_{n})} 232:
for a node is simply its adjacent (or neighboring) nodes. In a dependency network, the
2444:
Predicting what web pages a person will access based on his or her history on the site;
1993: 1971: 1951: 1371: 1020: 434: 392: 233: 229: 220: 200: 2554: 2441:
Predicting what movies a person will like based on his or her ratings of movies seen;
2485:"Dependency Networks for Inference, Collaborative Filtering, and Data Visualization" 409:
is a cyclic directed graph, where each of its nodes corresponds to a variable in
1370:. One of the alternatives for performing probabilistic inference is using 451:
is a set of conditional probability distributions. The parents of node
2518: 2114:{\displaystyle p(x_{i}|\mathbf {p} ):=p(x_{i}|\mathbf {pa_{i}} )} 1148:
are the input variables. To learn a decision tree structure for
266:
A consistent dependency network for a set of random variables
118: 56: 15: 1237:, until no more replacements increase the score of the tree. 1589:
You can see below an algorithm that can be used for obtain
1510:
is small is to use modified ordered Gibbs sampler, where
929:
in a dependency network is the conditional distribution
2507:"Large-Sample Learning of Bayesian Networks is NP-Hard" 258:. Markov networks, in contrast, are always consistent. 139: 624:
that satisfy the following independence relationships
272: 2370: 2312: 2267: 2225: 2183: 2134: 2039: 2016: 1996: 1974: 1954: 1927: 1892: 1861: 1836: 1808: 1778: 1748: 1716: 1694: 1664: 1634: 1595: 1569: 1547: 1516: 1485: 1448: 1417: 1380: 1354: 1332: 1310: 1288: 1251: 1208: 1181: 1154: 1119: 1092: 1070: 1043: 1023: 996: 935: 908: 867: 836: 633: 516: 484: 457: 437: 415: 395: 363: 332: 134:
may be too technical for most readers to understand
2429: 2344: 2294: 2252: 2210: 2166: 2128:Use a modified ordered Gibbs sampler to determine 2113: 2024: 2002: 1980: 1960: 1940: 1913: 1875: 1844: 1822: 1792: 1762: 1724: 1702: 1680: 1650: 1620: 1577: 1555: 1530: 1502: 1471: 1434: 1403: 1362: 1340: 1318: 1296: 1274: 1229: 1194: 1167: 1140: 1105: 1078: 1056: 1029: 1009: 982: 921: 885: 853: 819: 616: 502: 470: 443: 423: 401: 381: 349: 318: 2430:{\displaystyle p(x_{i}=1|\mathbf {x} -{x_{i}}=0)} 1607: 1086:, a probabilistic decision tree is learned where 1348:(the 'input' variables) are disjoint subsets of 319:{\textstyle \mathbf {X} =(X_{1},\ldots ,X_{n})} 1800:(* the processed and conditioning variables *) 983:{\displaystyle p(x_{i}|\mathbf {x} -{x_{i}})} 236:for a node is simply the set of its parents. 8: 1681:{\displaystyle \mathbf {z} \in \mathbf {Z} } 1651:{\displaystyle \mathbf {y} \in \mathbf {Y} } 1876:{\displaystyle \mathbf {U} \neq \emptyset } 240:Dependency network versus Bayesian networks 50:Learn how and when to remove these messages 249:Dependency networks versus Markov networks 2517: 2411: 2406: 2398: 2393: 2381: 2369: 2334: 2329: 2323: 2311: 2286: 2268: 2266: 2244: 2226: 2224: 2202: 2184: 2182: 2156: 2151: 2145: 2133: 2101: 2093: 2088: 2082: 2061: 2056: 2050: 2038: 2017: 2015: 1995: 1973: 1953: 1932: 1926: 1906: 1897: 1891: 1862: 1860: 1837: 1835: 1809: 1807: 1779: 1777: 1749: 1747: 1717: 1715: 1695: 1693: 1673: 1665: 1663: 1643: 1635: 1633: 1606: 1602: 1594: 1570: 1568: 1548: 1546: 1517: 1515: 1492: 1484: 1455: 1447: 1424: 1416: 1387: 1379: 1355: 1353: 1333: 1331: 1311: 1309: 1289: 1287: 1258: 1250: 1221: 1209: 1207: 1186: 1180: 1159: 1153: 1132: 1120: 1118: 1097: 1091: 1071: 1069: 1048: 1042: 1022: 1001: 995: 970: 965: 957: 952: 946: 934: 913: 907: 866: 843: 835: 804: 799: 791: 782: 760: 735: 716: 697: 684: 661: 653: 644: 632: 605: 580: 561: 542: 525: 517: 515: 493: 485: 483: 462: 456: 436: 416: 414: 394: 362: 339: 331: 307: 288: 273: 271: 180:Learn how and when to remove this message 162:Learn how and when to remove this message 146:, without removing the technical details. 107:Learn how and when to remove this message 2306:Returns the product of the conditionals 224:explain away the node in question. In a 70:This article includes a list of general 2475: 2536: 2525: 1586:thereby avoiding some Gibbs sampling. 2345:{\displaystyle p(x_{i}|\mathbf {p} )} 2295:{\displaystyle \mathbf {p:=p} +x_{i}} 2253:{\displaystyle \mathbf {P:=P} +X_{i}} 2211:{\displaystyle \mathbf {U:=U} -X_{i}} 2167:{\displaystyle p(x_{i}|\mathbf {p} )} 1914:{\displaystyle X_{i}\in \mathbf {U} } 1472:{\displaystyle p(\mathbf {y\mid z} )} 1404:{\displaystyle p(\mathbf {y\mid z} )} 1275:{\displaystyle p(\mathbf {y\mid z} )} 144:make it understandable to non-experts 7: 2492:Journal of Machine Learning Research 1870: 1230:{\displaystyle \mathbf {X} -X_{i}} 1141:{\displaystyle \mathbf {X} -X_{i}} 76:it lacks sufficient corresponding 14: 1621:{\displaystyle p(\mathbf {y|z} )} 897:Structure and parameters learning 503:{\displaystyle \mathbf {Pa_{i}} } 31:This article has multiple issues. 2399: 2335: 2275: 2272: 2269: 2233: 2230: 2227: 2191: 2188: 2185: 2157: 2102: 2098: 2094: 2062: 2018: 1907: 1863: 1838: 1816: 1813: 1810: 1786: 1783: 1780: 1756: 1753: 1750: 1718: 1696: 1674: 1666: 1644: 1636: 1611: 1603: 1571: 1549: 1538:is fixed during Gibbs sampling. 1524: 1521: 1518: 1493: 1462: 1456: 1425: 1394: 1388: 1356: 1334: 1312: 1290: 1265: 1259: 1210: 1121: 1072: 958: 844: 792: 662: 658: 654: 526: 522: 518: 510:, correspond to those variables 494: 490: 486: 417: 340: 274: 123: 61: 20: 1823:{\displaystyle \mathbf {p:=z} } 1793:{\displaystyle \mathbf {P:=Z} } 1770:(* the unprocessed variables *) 1763:{\displaystyle \mathbf {U:=Y} } 1503:{\displaystyle p(\mathbf {z} )} 1435:{\displaystyle p(\mathbf {z} )} 854:{\displaystyle p(\mathbf {x} )} 350:{\displaystyle p(\mathbf {x} )} 39:or discuss these issues on the 2424: 2394: 2374: 2339: 2330: 2316: 2161: 2152: 2138: 2108: 2089: 2075: 2066: 2057: 2043: 1615: 1599: 1531:{\displaystyle \mathbf {Z=z} } 1497: 1489: 1466: 1452: 1429: 1421: 1398: 1384: 1282:, given a graphical model for 1269: 1255: 977: 953: 939: 880: 868: 848: 840: 811: 775: 766: 677: 668: 637: 611: 535: 376: 364: 344: 336: 313: 281: 256:joint probability distribution 1: 2464:Relational dependency network 1628:for a particular instance of 2025:{\displaystyle \mathbf {P} } 1845:{\displaystyle \mathbf {P} } 1725:{\displaystyle \mathbf {Z} } 1703:{\displaystyle \mathbf {Y} } 1578:{\displaystyle \mathbf {Y} } 1556:{\displaystyle \mathbf {y} } 1363:{\displaystyle \mathbf {X} } 1341:{\displaystyle \mathbf {Z} } 1319:{\displaystyle \mathbf {Y} } 1297:{\displaystyle \mathbf {X} } 1079:{\displaystyle \mathbf {X} } 424:{\displaystyle \mathbf {X} } 1113:is the target variable and 2582: 902:distribution for variable 2505:HECKERMAN, David (2012). 1326:(the 'target' variables) 193:Dependency networks (DNs) 1541:It may also happen that 326:with joint distribution 2362:Collaborative Filtering 1948:has no more parents in 1241:Probabilistic Inference 91:more precise citations. 2535:Cite journal requires 2431: 2346: 2296: 2254: 2212: 2168: 2115: 2026: 2004: 1990:If all the parents of 1982: 1962: 1942: 1915: 1877: 1846: 1824: 1794: 1764: 1732:are disjoint subsets. 1726: 1704: 1682: 1652: 1622: 1579: 1557: 1532: 1504: 1473: 1436: 1405: 1364: 1342: 1320: 1298: 1276: 1231: 1196: 1169: 1142: 1107: 1080: 1058: 1031: 1011: 984: 923: 887: 855: 821: 618: 504: 472: 445: 425: 403: 383: 351: 320: 2432: 2347: 2297: 2255: 2213: 2169: 2116: 2027: 2005: 1983: 1968:than any variable in 1963: 1943: 1941:{\displaystyle X_{i}} 1916: 1878: 1847: 1825: 1795: 1765: 1727: 1705: 1683: 1653: 1623: 1580: 1558: 1533: 1505: 1474: 1437: 1406: 1365: 1343: 1321: 1299: 1277: 1232: 1197: 1195:{\displaystyle X_{j}} 1170: 1168:{\displaystyle X_{i}} 1143: 1108: 1106:{\displaystyle X_{i}} 1081: 1059: 1057:{\displaystyle X_{i}} 1032: 1012: 1010:{\displaystyle X_{i}} 985: 924: 922:{\displaystyle X_{i}} 888: 886:{\displaystyle (G,P)} 856: 822: 619: 505: 473: 471:{\displaystyle X_{i}} 446: 426: 404: 384: 382:{\displaystyle (G,P)} 352: 321: 2368: 2310: 2265: 2223: 2181: 2132: 2037: 2014: 1994: 1972: 1952: 1925: 1890: 1859: 1834: 1806: 1776: 1746: 1714: 1692: 1662: 1632: 1593: 1567: 1545: 1514: 1483: 1446: 1415: 1378: 1352: 1330: 1308: 1286: 1249: 1206: 1179: 1152: 1117: 1090: 1068: 1041: 1021: 994: 933: 906: 865: 834: 631: 514: 482: 455: 435: 413: 393: 361: 330: 270: 1563:is rare, e.g. when 226:Markov random field 2427: 2342: 2292: 2250: 2208: 2164: 2111: 2022: 2000: 1978: 1958: 1938: 1911: 1873: 1842: 1830:(* the values for 1820: 1790: 1760: 1722: 1700: 1678: 1648: 1618: 1575: 1553: 1528: 1500: 1469: 1432: 1401: 1360: 1338: 1316: 1294: 1272: 1227: 1192: 1165: 1138: 1103: 1076: 1054: 1027: 1007: 980: 919: 883: 851: 817: 614: 500: 468: 441: 421: 399: 379: 347: 316: 2003:{\displaystyle X} 1981:{\displaystyle U} 1961:{\displaystyle U} 1030:{\displaystyle X} 444:{\displaystyle P} 402:{\displaystyle G} 205:Bayesian networks 190: 189: 182: 172: 171: 164: 117: 116: 109: 54: 2573: 2561:Graphical models 2545: 2544: 2538: 2533: 2531: 2523: 2521: 2511: 2502: 2496: 2495: 2489: 2480: 2436: 2434: 2433: 2428: 2417: 2416: 2415: 2402: 2397: 2386: 2385: 2351: 2349: 2348: 2343: 2338: 2333: 2328: 2327: 2301: 2299: 2298: 2293: 2291: 2290: 2278: 2259: 2257: 2256: 2251: 2249: 2248: 2236: 2217: 2215: 2214: 2209: 2207: 2206: 2194: 2173: 2171: 2170: 2165: 2160: 2155: 2150: 2149: 2120: 2118: 2117: 2112: 2107: 2106: 2105: 2092: 2087: 2086: 2065: 2060: 2055: 2054: 2031: 2029: 2028: 2023: 2021: 2009: 2007: 2006: 2001: 1987: 1985: 1984: 1979: 1967: 1965: 1964: 1959: 1947: 1945: 1944: 1939: 1937: 1936: 1920: 1918: 1917: 1912: 1910: 1902: 1901: 1882: 1880: 1879: 1874: 1866: 1851: 1849: 1848: 1843: 1841: 1829: 1827: 1826: 1821: 1819: 1799: 1797: 1796: 1791: 1789: 1769: 1767: 1766: 1761: 1759: 1731: 1729: 1728: 1723: 1721: 1709: 1707: 1706: 1701: 1699: 1687: 1685: 1684: 1679: 1677: 1669: 1657: 1655: 1654: 1649: 1647: 1639: 1627: 1625: 1624: 1619: 1614: 1610: 1584: 1582: 1581: 1576: 1574: 1562: 1560: 1559: 1554: 1552: 1537: 1535: 1534: 1529: 1527: 1509: 1507: 1506: 1501: 1496: 1478: 1476: 1475: 1470: 1465: 1441: 1439: 1438: 1433: 1428: 1410: 1408: 1407: 1402: 1397: 1369: 1367: 1366: 1361: 1359: 1347: 1345: 1344: 1339: 1337: 1325: 1323: 1322: 1317: 1315: 1303: 1301: 1300: 1295: 1293: 1281: 1279: 1278: 1273: 1268: 1236: 1234: 1233: 1228: 1226: 1225: 1213: 1201: 1199: 1198: 1193: 1191: 1190: 1174: 1172: 1171: 1166: 1164: 1163: 1147: 1145: 1144: 1139: 1137: 1136: 1124: 1112: 1110: 1109: 1104: 1102: 1101: 1085: 1083: 1082: 1077: 1075: 1063: 1061: 1060: 1055: 1053: 1052: 1036: 1034: 1033: 1028: 1016: 1014: 1013: 1008: 1006: 1005: 989: 987: 986: 981: 976: 975: 974: 961: 956: 951: 950: 928: 926: 925: 920: 918: 917: 892: 890: 889: 884: 860: 858: 857: 852: 847: 826: 824: 823: 818: 810: 809: 808: 795: 787: 786: 765: 764: 746: 745: 727: 726: 702: 701: 689: 688: 667: 666: 665: 649: 648: 623: 621: 620: 615: 610: 609: 591: 590: 572: 571: 547: 546: 531: 530: 529: 509: 507: 506: 501: 499: 498: 497: 477: 475: 474: 469: 467: 466: 450: 448: 447: 442: 430: 428: 427: 422: 420: 408: 406: 405: 400: 388: 386: 385: 380: 356: 354: 353: 348: 343: 325: 323: 322: 317: 312: 311: 293: 292: 277: 217:Bayesian network 197:graphical models 185: 178: 167: 160: 156: 153: 147: 127: 126: 119: 112: 105: 101: 98: 92: 87:this article by 78:inline citations 65: 64: 57: 46: 24: 23: 16: 2581: 2580: 2576: 2575: 2574: 2572: 2571: 2570: 2551: 2550: 2549: 2548: 2534: 2524: 2509: 2504: 2503: 2499: 2487: 2482: 2481: 2477: 2472: 2460: 2407: 2377: 2366: 2365: 2358: 2319: 2308: 2307: 2282: 2263: 2262: 2240: 2221: 2220: 2198: 2179: 2178: 2141: 2130: 2129: 2097: 2078: 2046: 2035: 2034: 2012: 2011: 1992: 1991: 1970: 1969: 1950: 1949: 1928: 1923: 1922: 1893: 1888: 1887: 1857: 1856: 1832: 1831: 1804: 1803: 1774: 1773: 1744: 1743: 1712: 1711: 1690: 1689: 1660: 1659: 1630: 1629: 1591: 1590: 1565: 1564: 1543: 1542: 1512: 1511: 1481: 1480: 1444: 1443: 1413: 1412: 1376: 1375: 1350: 1349: 1328: 1327: 1306: 1305: 1284: 1283: 1247: 1246: 1243: 1217: 1204: 1203: 1182: 1177: 1176: 1155: 1150: 1149: 1128: 1115: 1114: 1093: 1088: 1087: 1066: 1065: 1044: 1039: 1038: 1019: 1018: 997: 992: 991: 966: 942: 931: 930: 909: 904: 903: 899: 863: 862: 832: 831: 800: 778: 756: 731: 712: 693: 680: 657: 640: 629: 628: 601: 576: 557: 538: 521: 512: 511: 489: 480: 479: 458: 453: 452: 433: 432: 411: 410: 391: 390: 359: 358: 328: 327: 303: 284: 268: 267: 264: 251: 242: 213: 201:Markov networks 186: 175: 174: 173: 168: 157: 151: 148: 140:help improve it 137: 128: 124: 113: 102: 96: 93: 83:Please help to 82: 66: 62: 25: 21: 12: 11: 5: 2579: 2577: 2569: 2568: 2563: 2553: 2552: 2547: 2546: 2537:|journal= 2497: 2474: 2473: 2471: 2468: 2467: 2466: 2459: 2456: 2452: 2451: 2448: 2445: 2442: 2426: 2423: 2420: 2414: 2410: 2405: 2401: 2396: 2392: 2389: 2384: 2380: 2376: 2373: 2357: 2354: 2353: 2352: 2341: 2337: 2332: 2326: 2322: 2318: 2315: 2304: 2303: 2302: 2289: 2285: 2281: 2277: 2274: 2271: 2260: 2247: 2243: 2239: 2235: 2232: 2229: 2218: 2205: 2201: 2197: 2193: 2190: 2187: 2176: 2175: 2174: 2163: 2159: 2154: 2148: 2144: 2140: 2137: 2123: 2122: 2121: 2110: 2104: 2100: 2096: 2091: 2085: 2081: 2077: 2074: 2071: 2068: 2064: 2059: 2053: 2049: 2045: 2042: 2020: 1999: 1988: 1977: 1957: 1935: 1931: 1909: 1905: 1900: 1896: 1872: 1869: 1865: 1853: 1840: 1818: 1815: 1812: 1801: 1788: 1785: 1782: 1771: 1758: 1755: 1752: 1740: 1739: 1720: 1698: 1676: 1672: 1668: 1646: 1642: 1638: 1617: 1613: 1609: 1605: 1601: 1598: 1573: 1551: 1526: 1523: 1520: 1499: 1495: 1491: 1488: 1468: 1464: 1461: 1458: 1454: 1451: 1431: 1427: 1423: 1420: 1400: 1396: 1393: 1390: 1386: 1383: 1372:Gibbs sampling 1358: 1336: 1314: 1292: 1271: 1267: 1264: 1261: 1257: 1254: 1242: 1239: 1224: 1220: 1216: 1212: 1189: 1185: 1162: 1158: 1135: 1131: 1127: 1123: 1100: 1096: 1074: 1051: 1047: 1026: 1004: 1000: 979: 973: 969: 964: 960: 955: 949: 945: 941: 938: 916: 912: 898: 895: 882: 879: 876: 873: 870: 850: 846: 842: 839: 828: 827: 816: 813: 807: 803: 798: 794: 790: 785: 781: 777: 774: 771: 768: 763: 759: 755: 752: 749: 744: 741: 738: 734: 730: 725: 722: 719: 715: 711: 708: 705: 700: 696: 692: 687: 683: 679: 676: 673: 670: 664: 660: 656: 652: 647: 643: 639: 636: 613: 608: 604: 600: 597: 594: 589: 586: 583: 579: 575: 570: 567: 564: 560: 556: 553: 550: 545: 541: 537: 534: 528: 524: 520: 496: 492: 488: 465: 461: 440: 419: 398: 378: 375: 372: 369: 366: 346: 342: 338: 335: 315: 310: 306: 302: 299: 296: 291: 287: 283: 280: 276: 263: 260: 250: 247: 241: 238: 234:Markov blanket 230:Markov blanket 221:Markov blanket 212: 211:Markov blanket 209: 188: 187: 170: 169: 131: 129: 122: 115: 114: 69: 67: 60: 55: 29: 28: 26: 19: 13: 10: 9: 6: 4: 3: 2: 2578: 2567: 2564: 2562: 2559: 2558: 2556: 2542: 2529: 2520: 2515: 2508: 2501: 2498: 2493: 2486: 2479: 2476: 2469: 2465: 2462: 2461: 2457: 2455: 2449: 2446: 2443: 2440: 2439: 2438: 2421: 2418: 2412: 2408: 2403: 2390: 2387: 2382: 2378: 2371: 2363: 2355: 2324: 2320: 2313: 2305: 2287: 2283: 2279: 2261: 2245: 2241: 2237: 2219: 2203: 2199: 2195: 2177: 2146: 2142: 2135: 2127: 2126: 2124: 2083: 2079: 2072: 2069: 2051: 2047: 2040: 2033: 2032: 1997: 1989: 1975: 1955: 1933: 1929: 1903: 1898: 1894: 1885: 1884: 1867: 1854: 1802: 1772: 1742: 1741: 1738: 1735: 1734: 1733: 1670: 1640: 1596: 1587: 1539: 1486: 1459: 1449: 1418: 1391: 1381: 1373: 1262: 1252: 1240: 1238: 1222: 1218: 1214: 1187: 1183: 1160: 1156: 1133: 1129: 1125: 1098: 1094: 1049: 1045: 1024: 1002: 998: 971: 967: 962: 947: 943: 936: 914: 910: 896: 894: 877: 874: 871: 837: 814: 805: 801: 796: 788: 783: 779: 772: 769: 761: 757: 753: 750: 747: 742: 739: 736: 732: 728: 723: 720: 717: 713: 709: 706: 703: 698: 694: 690: 685: 681: 674: 671: 650: 645: 641: 634: 627: 626: 625: 606: 602: 598: 595: 592: 587: 584: 581: 577: 573: 568: 565: 562: 558: 554: 551: 548: 543: 539: 532: 463: 459: 438: 396: 373: 370: 367: 333: 308: 304: 300: 297: 294: 289: 285: 278: 261: 259: 257: 248: 246: 239: 237: 235: 231: 227: 222: 218: 210: 208: 206: 202: 199:, similar to 198: 194: 184: 181: 166: 163: 155: 145: 141: 135: 132:This article 130: 121: 120: 111: 108: 100: 90: 86: 80: 79: 73: 68: 59: 58: 53: 51: 44: 43: 38: 37: 32: 27: 18: 17: 2528:cite journal 2500: 2491: 2478: 2453: 2359: 2356:Applications 1737:Algorithm 1: 1736: 1588: 1540: 1244: 900: 829: 265: 252: 243: 214: 192: 191: 176: 158: 152:January 2020 149: 133: 103: 97:January 2020 94: 75: 47: 40: 34: 33:Please help 30: 89:introducing 2566:Algorithms 2555:Categories 2470:References 1921:such that 1017:in domain 478:, denoted 357:is a pair 262:Definition 72:references 36:improve it 2519:1212.2468 2404:− 2196:− 1904:∈ 1871:∅ 1868:≠ 1671:∈ 1641:∈ 1460:∣ 1392:∣ 1263:∣ 1215:− 1126:− 963:− 797:− 789:∣ 751:… 721:− 707:… 691:∣ 651:∣ 596:… 566:− 552:… 533:⊆ 298:… 42:talk page 2458:See also 1688:, where 1304:, where 2010:are in 1886:Choose 219:, the 138:Please 85:improve 1855:While 431:, and 389:where 228:, the 74:, but 2514:arXiv 2510:(PDF) 2488:(PDF) 2125:Else 1479:when 215:In a 2541:help 1710:and 1658:and 195:are 1411:or 1202:in 1064:in 142:to 2557:: 2532:: 2530:}} 2526:{{ 2512:. 2490:. 2273::= 2231::= 2189::= 2070::= 1883:: 1852:*) 1814::= 1784::= 1754::= 45:. 2543:) 2539:( 2522:. 2516:: 2494:. 2425:) 2422:0 2419:= 2413:i 2409:x 2400:x 2395:| 2391:1 2388:= 2383:i 2379:x 2375:( 2372:p 2340:) 2336:p 2331:| 2325:i 2321:x 2317:( 2314:p 2288:i 2284:x 2280:+ 2276:p 2270:p 2246:i 2242:X 2238:+ 2234:P 2228:P 2204:i 2200:X 2192:U 2186:U 2162:) 2158:p 2153:| 2147:i 2143:x 2139:( 2136:p 2109:) 2103:i 2099:a 2095:p 2090:| 2084:i 2080:x 2076:( 2073:p 2067:) 2063:p 2058:| 2052:i 2048:x 2044:( 2041:p 2019:P 1998:X 1976:U 1956:U 1934:i 1930:X 1908:U 1899:i 1895:X 1864:U 1839:P 1817:z 1811:p 1787:Z 1781:P 1757:Y 1751:U 1719:Z 1697:Y 1675:Z 1667:z 1645:Y 1637:y 1616:) 1612:z 1608:| 1604:y 1600:( 1597:p 1572:Y 1550:y 1525:z 1522:= 1519:Z 1498:) 1494:z 1490:( 1487:p 1467:) 1463:z 1457:y 1453:( 1450:p 1430:) 1426:z 1422:( 1419:p 1399:) 1395:z 1389:y 1385:( 1382:p 1357:X 1335:Z 1313:Y 1291:X 1270:) 1266:z 1260:y 1256:( 1253:p 1223:i 1219:X 1211:X 1188:j 1184:X 1161:i 1157:X 1134:i 1130:X 1122:X 1099:i 1095:X 1073:X 1050:i 1046:X 1025:X 1003:i 999:X 978:) 972:i 968:x 959:x 954:| 948:i 944:x 940:( 937:p 915:i 911:X 881:) 878:P 875:, 872:G 869:( 849:) 845:x 841:( 838:p 815:. 812:) 806:i 802:x 793:x 784:i 780:x 776:( 773:p 770:= 767:) 762:n 758:x 754:, 748:, 743:1 740:+ 737:i 733:x 729:, 724:1 718:i 714:x 710:, 704:, 699:1 695:x 686:i 682:x 678:( 675:p 672:= 669:) 663:i 659:a 655:p 646:i 642:x 638:( 635:p 612:) 607:n 603:X 599:, 593:, 588:1 585:+ 582:i 578:X 574:, 569:1 563:i 559:X 555:, 549:, 544:1 540:X 536:( 527:i 523:a 519:P 495:i 491:a 487:P 464:i 460:X 439:P 418:X 397:G 377:) 374:P 371:, 368:G 365:( 345:) 341:x 337:( 334:p 314:) 309:n 305:X 301:, 295:, 290:1 286:X 282:( 279:= 275:X 183:) 177:( 165:) 159:( 154:) 150:( 136:. 110:) 104:( 99:) 95:( 81:. 52:) 48:(

Index

improve it
talk page
Learn how and when to remove these messages
references
inline citations
improve
introducing
Learn how and when to remove this message
help improve it
make it understandable to non-experts
Learn how and when to remove this message
Learn how and when to remove this message
graphical models
Markov networks
Bayesian networks
Bayesian network
Markov blanket
Markov random field
Markov blanket
Markov blanket
joint probability distribution
Gibbs sampling
Collaborative Filtering
Relational dependency network
"Dependency Networks for Inference, Collaborative Filtering, and Data Visualization"
"Large-Sample Learning of Bayesian Networks is NP-Hard"
arXiv
1212.2468
cite journal
help

Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.