Knowledge

Recursive Bayesian estimation

Source 📝

170: 1352: 749: 1032: 1021: 1548: 101:
In a simple example, a robot moving throughout a grid may have several different sensors that provide it with information about its surroundings. The robot may start out with certainty that it is at position (0,0). However, as it moves farther and farther from its original position, the robot has
89:
to infer its position and orientation. Essentially, Bayes filters allow robots to continuously update their most likely position within a coordinate system, based on the most recently acquired sensor data. This is a recursive algorithm. It consists of two parts: prediction and innovation. If the
511: 340: 500: 1347:{\displaystyle p({\textbf {x}}_{k}|{\textbf {z}}_{1:k})={\frac {p({\textbf {z}}_{k}|{\textbf {x}}_{k})p({\textbf {x}}_{k}|{\textbf {z}}_{1:k-1})}{p({\textbf {z}}_{k}|{\textbf {z}}_{1:k-1})}}\propto p({\textbf {z}}_{k}|{\textbf {x}}_{k})p({\textbf {x}}_{k}|{\textbf {z}}_{1:k-1})} 758:, the probability distribution of interest is associated with the current states conditioned on the measurements up to the current timestep. (This is achieved by marginalising out the previous states and dividing by the probability of the measurement set.) 817: 102:
continuously less certainty about its position; using a Bayes filter, a probability can be assigned to the robot's belief about its current position, and that probability can be continuously updated from additional sensor information.
1363: 69:) recursively over time using incoming measurements and a mathematical process model. The process relies heavily upon mathematical concepts and models that are theorized within a study of prior and posterior probabilities known as 769:
steps of the Kalman filter written probabilistically. The probability distribution associated with the predicted state is the sum (integral) of the products of the probability distribution associated with the transition from the
744:{\displaystyle p({\textbf {x}}_{0},\dots ,{\textbf {x}}_{k},{\textbf {z}}_{1},\dots ,{\textbf {z}}_{k})=p({\textbf {x}}_{0})\prod _{i=1}^{k}p({\textbf {z}}_{i}|{\textbf {x}}_{i})p({\textbf {x}}_{i}|{\textbf {x}}_{i-1}).} 1633:
Sequential Bayesian filtering is the extension of the Bayesian estimation for the case when the observed value changes in time. It is a method to estimate the real value of an observed variable that evolves in time.
183: 355: 1016:{\displaystyle p({\textbf {x}}_{k}|{\textbf {z}}_{1:k-1})=\int p({\textbf {x}}_{k}|{\textbf {x}}_{k-1})p({\textbf {x}}_{k-1}|{\textbf {z}}_{1:k-1})\,d{\textbf {x}}_{k-1}} 177:
Because of the Markov assumption, the probability of the current true state given the immediately previous one is conditionally independent of the other earlier states.
1591: 809: 1543:{\displaystyle p({\textbf {z}}_{k}|{\textbf {z}}_{1:k-1})=\int p({\textbf {z}}_{k}|{\textbf {x}}_{k})p({\textbf {x}}_{k}|{\textbf {z}}_{1:k-1})d{\textbf {x}}_{k}} 1571: 156: 128: 1692:
Arulampalam, M. Sanjeev; Maskell, Simon; Gordon, Neil (2002). "A Tutorial on Particle Filters for On-line Non-linear/Non-Gaussian Bayesian Tracking".
335:{\displaystyle p({\textbf {x}}_{k}|{\textbf {x}}_{k-1},{\textbf {x}}_{k-2},\dots ,{\textbf {x}}_{0})=p({\textbf {x}}_{k}|{\textbf {x}}_{k-1})} 1593:, which can usually be ignored in practice. The numerator can be calculated and then simply normalized, since its integral must be unity. 495:{\displaystyle p({\textbf {z}}_{k}|{\textbf {x}}_{k},{\textbf {x}}_{k-1},\dots ,{\textbf {x}}_{0})=p({\textbf {z}}_{k}|{\textbf {x}}_{k})} 349:-th timestep is dependent only upon the current state, so is conditionally independent of all other states given the current state. 1606: 1026:
The probability distribution of update is proportional to the product of the measurement likelihood and the predicted state.
169: 1809: 1615: 66: 62: 1824: 1814: 31: 30:
This article is about Bayes filter, a general probabilistic approach. For the spam filter with a similar name, see
1819: 1701: 505:
Using these assumptions the probability distribution over all states of the HMM can be written simply as
1755:"A survey of probabilistic models, using the Bayesian Programming methodology as a unifying framework" 1779:
Volkov, Alexander (2015). "Accuracy bounds of non-Gaussian Bayesian tracking in a NLOS environment".
1706: 1766: 135: 91: 70: 1740:
Chen, Zhe Sage (2003). "Bayesian Filtering: From Kalman Filters to Particle Filters, and Beyond".
58: 38: 1651: 131: 1576: 781: 1788: 1728: 1711: 163: 82: 46: 778:-th and the probability distribution associated with the previous state, over all possible 1725:
A Discriminative Approach to Bayesian Filtering with Applications to Human Neural Decoding
1611: 1675: 1556: 159: 141: 113: 1803: 1602: 95: 1792: 42: 1723:
Burkhart, Michael C. (2019). "Chapter 1. An Overview of Bayesian Filtering".
17: 1754: 1679: 1732: 1715: 94:
and the transitions are linear, the Bayes filter becomes equal to the
1674:
The notion of Sequential Bayesian filtering is extensively used in
1614:, a sequential Monte Carlo (SMC) based technique, which models the 86: 85:
for calculating the probabilities of multiple beliefs to allow a
1624:, which subdivide the PDF into a deterministic discrete grid 754:
However, when using the Kalman filter to estimate the state
168: 1742:
Statistics: A Journal of Theoretical and Applied Statistics
1753:
Diard, Julien; Bessière, Pierre; Mazer, Emmanuel (2003).
1579: 1559: 1366: 1035: 820: 784: 514: 358: 186: 144: 116: 27:
Process for estimating a probability density function
1573:, so we can always substitute it for a coefficient 1585: 1565: 1542: 1346: 1015: 803: 743: 494: 334: 150: 122: 1660:values given past and current observations, and 8: 1670:value given past and current observations. 1648:value given past and current observations, 57:, is a general probabilistic approach for 1727:. Providence, RI, USA: Brown University. 1705: 1578: 1558: 1534: 1528: 1527: 1502: 1496: 1495: 1489: 1483: 1477: 1476: 1460: 1454: 1453: 1447: 1441: 1435: 1434: 1400: 1394: 1393: 1387: 1381: 1375: 1374: 1365: 1323: 1317: 1316: 1310: 1304: 1298: 1297: 1281: 1275: 1274: 1268: 1262: 1256: 1255: 1221: 1215: 1214: 1208: 1202: 1196: 1195: 1165: 1159: 1158: 1152: 1146: 1140: 1139: 1123: 1117: 1116: 1110: 1104: 1098: 1097: 1087: 1069: 1063: 1062: 1056: 1050: 1044: 1043: 1034: 1001: 995: 994: 989: 968: 962: 961: 955: 943: 937: 936: 914: 908: 907: 901: 895: 889: 888: 854: 848: 847: 841: 835: 829: 828: 819: 789: 783: 723: 717: 716: 710: 704: 698: 697: 681: 675: 674: 668: 662: 656: 655: 642: 631: 618: 612: 611: 592: 586: 585: 569: 563: 562: 552: 546: 545: 529: 523: 522: 513: 483: 477: 476: 470: 464: 458: 457: 438: 432: 431: 409: 403: 402: 392: 386: 385: 379: 373: 367: 366: 357: 317: 311: 310: 304: 298: 292: 291: 272: 266: 265: 243: 237: 236: 220: 214: 213: 207: 201: 195: 194: 185: 143: 115: 81:A Bayes filter is an algorithm used in 1694:IEEE Transactions on Signal Processing 7: 1529: 1497: 1478: 1455: 1436: 1395: 1376: 1318: 1299: 1276: 1257: 1216: 1197: 1160: 1141: 1118: 1099: 1064: 1045: 996: 963: 938: 909: 890: 849: 830: 718: 699: 676: 657: 613: 587: 564: 547: 524: 478: 459: 433: 404: 387: 368: 312: 293: 267: 238: 215: 196: 162:. The following picture presents a 1605:, a recursive Bayesian filter for 345:Similarly, the measurement at the 138:(HMM), which means the true state 25: 1607:multivariate normal distributions 1768:Bayesian Filtering and Smoothing 158:is assumed to be an unobserved 1637:There are several variations: 1618:using a set of discrete points 1520: 1490: 1472: 1466: 1448: 1430: 1418: 1388: 1370: 1341: 1311: 1293: 1287: 1269: 1251: 1239: 1209: 1191: 1183: 1153: 1135: 1129: 1111: 1093: 1081: 1057: 1039: 986: 956: 932: 926: 902: 884: 872: 842: 824: 735: 711: 693: 687: 669: 651: 624: 607: 598: 518: 489: 471: 453: 444: 380: 362: 329: 305: 287: 278: 208: 190: 1: 1774:. Cambridge University Press. 1629:Sequential Bayesian filtering 51:recursive Bayesian estimation 1793:10.1016/j.sigpro.2014.10.025 63:probability density function 1666:when estimating a probable 1841: 32:Naive Bayes spam filtering 29: 1553:is constant relative to 774:- 1)-th timestep to the 1586:{\displaystyle \alpha } 804:{\displaystyle x_{k-1}} 1587: 1567: 1544: 1348: 1017: 805: 745: 647: 496: 336: 174: 152: 124: 1765:Särkkä, Simo (2013). 1622:Grid-based estimators 1588: 1568: 1545: 1349: 1018: 806: 746: 627: 497: 337: 172: 153: 125: 1644:when estimating the 1577: 1557: 1364: 1033: 818: 782: 512: 356: 184: 142: 114: 92:normally distributed 1810:Bayesian estimation 173:Hidden Markov model 136:hidden Markov model 71:Bayesian statistics 1733:10.26300/nhfp-xv22 1583: 1563: 1540: 1344: 1013: 801: 761:This leads to the 741: 492: 332: 175: 148: 120: 53:, also known as a 39:probability theory 1825:Signal estimation 1815:Nonlinear filters 1781:Signal Processing 1716:10.1109/78.978374 1566:{\displaystyle x} 1531: 1499: 1480: 1457: 1438: 1397: 1378: 1320: 1301: 1278: 1259: 1243: 1218: 1199: 1162: 1143: 1120: 1101: 1066: 1047: 998: 965: 940: 911: 892: 851: 832: 720: 701: 678: 659: 615: 589: 566: 549: 526: 480: 461: 435: 406: 389: 370: 314: 295: 269: 240: 217: 198: 151:{\displaystyle x} 123:{\displaystyle z} 110:The measurements 16:(Redirected from 1832: 1796: 1775: 1773: 1761: 1760:. cogprints.org. 1759: 1749: 1736: 1719: 1709: 1656:when estimating 1592: 1590: 1589: 1584: 1572: 1570: 1569: 1564: 1549: 1547: 1546: 1541: 1539: 1538: 1533: 1532: 1519: 1518: 1501: 1500: 1493: 1488: 1487: 1482: 1481: 1465: 1464: 1459: 1458: 1451: 1446: 1445: 1440: 1439: 1417: 1416: 1399: 1398: 1391: 1386: 1385: 1380: 1379: 1357:The denominator 1353: 1351: 1350: 1345: 1340: 1339: 1322: 1321: 1314: 1309: 1308: 1303: 1302: 1286: 1285: 1280: 1279: 1272: 1267: 1266: 1261: 1260: 1244: 1242: 1238: 1237: 1220: 1219: 1212: 1207: 1206: 1201: 1200: 1186: 1182: 1181: 1164: 1163: 1156: 1151: 1150: 1145: 1144: 1128: 1127: 1122: 1121: 1114: 1109: 1108: 1103: 1102: 1088: 1080: 1079: 1068: 1067: 1060: 1055: 1054: 1049: 1048: 1022: 1020: 1019: 1014: 1012: 1011: 1000: 999: 985: 984: 967: 966: 959: 954: 953: 942: 941: 925: 924: 913: 912: 905: 900: 899: 894: 893: 871: 870: 853: 852: 845: 840: 839: 834: 833: 810: 808: 807: 802: 800: 799: 750: 748: 747: 742: 734: 733: 722: 721: 714: 709: 708: 703: 702: 686: 685: 680: 679: 672: 667: 666: 661: 660: 646: 641: 623: 622: 617: 616: 597: 596: 591: 590: 574: 573: 568: 567: 557: 556: 551: 550: 534: 533: 528: 527: 501: 499: 498: 493: 488: 487: 482: 481: 474: 469: 468: 463: 462: 443: 442: 437: 436: 420: 419: 408: 407: 397: 396: 391: 390: 383: 378: 377: 372: 371: 341: 339: 338: 333: 328: 327: 316: 315: 308: 303: 302: 297: 296: 277: 276: 271: 270: 254: 253: 242: 241: 231: 230: 219: 218: 211: 206: 205: 200: 199: 164:Bayesian network 157: 155: 154: 149: 129: 127: 126: 121: 83:computer science 47:machine learning 21: 1840: 1839: 1835: 1834: 1833: 1831: 1830: 1829: 1800: 1799: 1778: 1771: 1764: 1757: 1752: 1739: 1722: 1707:10.1.1.117.1144 1691: 1688: 1686:Further reading 1631: 1612:Particle filter 1599: 1575: 1574: 1555: 1554: 1526: 1494: 1475: 1452: 1433: 1392: 1373: 1362: 1361: 1315: 1296: 1273: 1254: 1213: 1194: 1187: 1157: 1138: 1115: 1096: 1089: 1061: 1042: 1031: 1030: 993: 960: 935: 906: 887: 846: 827: 816: 815: 785: 780: 779: 715: 696: 673: 654: 610: 584: 561: 544: 521: 510: 509: 475: 456: 430: 401: 384: 365: 354: 353: 309: 290: 264: 235: 212: 193: 182: 181: 140: 139: 112: 111: 108: 79: 35: 28: 23: 22: 15: 12: 11: 5: 1838: 1836: 1828: 1827: 1822: 1820:Linear filters 1817: 1812: 1802: 1801: 1798: 1797: 1776: 1762: 1750: 1737: 1720: 1700:(2): 174–188. 1687: 1684: 1672: 1671: 1664: 1661: 1654: 1649: 1642: 1630: 1627: 1626: 1625: 1619: 1609: 1598: 1595: 1582: 1562: 1551: 1550: 1537: 1525: 1522: 1517: 1514: 1511: 1508: 1505: 1492: 1486: 1474: 1471: 1468: 1463: 1450: 1444: 1432: 1429: 1426: 1423: 1420: 1415: 1412: 1409: 1406: 1403: 1390: 1384: 1372: 1369: 1355: 1354: 1343: 1338: 1335: 1332: 1329: 1326: 1313: 1307: 1295: 1292: 1289: 1284: 1271: 1265: 1253: 1250: 1247: 1241: 1236: 1233: 1230: 1227: 1224: 1211: 1205: 1193: 1190: 1185: 1180: 1177: 1174: 1171: 1168: 1155: 1149: 1137: 1134: 1131: 1126: 1113: 1107: 1095: 1092: 1086: 1083: 1078: 1075: 1072: 1059: 1053: 1041: 1038: 1024: 1023: 1010: 1007: 1004: 992: 988: 983: 980: 977: 974: 971: 958: 952: 949: 946: 934: 931: 928: 923: 920: 917: 904: 898: 886: 883: 880: 877: 874: 869: 866: 863: 860: 857: 844: 838: 826: 823: 798: 795: 792: 788: 752: 751: 740: 737: 732: 729: 726: 713: 707: 695: 692: 689: 684: 671: 665: 653: 650: 645: 640: 637: 634: 630: 626: 621: 609: 606: 603: 600: 595: 583: 580: 577: 572: 560: 555: 543: 540: 537: 532: 520: 517: 503: 502: 491: 486: 473: 467: 455: 452: 449: 446: 441: 429: 426: 423: 418: 415: 412: 400: 395: 382: 376: 364: 361: 343: 342: 331: 326: 323: 320: 307: 301: 289: 286: 283: 280: 275: 263: 260: 257: 252: 249: 246: 234: 229: 226: 223: 210: 204: 192: 189: 160:Markov process 147: 132:manifestations 119: 107: 104: 90:variables are 78: 75: 26: 24: 14: 13: 10: 9: 6: 4: 3: 2: 1837: 1826: 1823: 1821: 1818: 1816: 1813: 1811: 1808: 1807: 1805: 1794: 1790: 1786: 1782: 1777: 1770: 1769: 1763: 1756: 1751: 1747: 1743: 1738: 1734: 1730: 1726: 1721: 1717: 1713: 1708: 1703: 1699: 1695: 1690: 1689: 1685: 1683: 1681: 1677: 1669: 1665: 1662: 1659: 1655: 1653: 1650: 1647: 1643: 1640: 1639: 1638: 1635: 1628: 1623: 1620: 1617: 1613: 1610: 1608: 1604: 1603:Kalman filter 1601: 1600: 1596: 1594: 1580: 1560: 1535: 1523: 1515: 1512: 1509: 1506: 1503: 1484: 1469: 1461: 1442: 1427: 1424: 1421: 1413: 1410: 1407: 1404: 1401: 1382: 1367: 1360: 1359: 1358: 1336: 1333: 1330: 1327: 1324: 1305: 1290: 1282: 1263: 1248: 1245: 1234: 1231: 1228: 1225: 1222: 1203: 1188: 1178: 1175: 1172: 1169: 1166: 1147: 1132: 1124: 1105: 1090: 1084: 1076: 1073: 1070: 1051: 1036: 1029: 1028: 1027: 1008: 1005: 1002: 990: 981: 978: 975: 972: 969: 950: 947: 944: 929: 921: 918: 915: 896: 881: 878: 875: 867: 864: 861: 858: 855: 836: 821: 814: 813: 812: 796: 793: 790: 786: 777: 773: 768: 764: 759: 757: 738: 730: 727: 724: 705: 690: 682: 663: 648: 643: 638: 635: 632: 628: 619: 604: 601: 593: 581: 578: 575: 570: 558: 553: 541: 538: 535: 530: 515: 508: 507: 506: 484: 465: 450: 447: 439: 427: 424: 421: 416: 413: 410: 398: 393: 374: 359: 352: 351: 350: 348: 324: 321: 318: 299: 284: 281: 273: 261: 258: 255: 250: 247: 244: 232: 227: 224: 221: 202: 187: 180: 179: 178: 171: 167: 165: 161: 145: 137: 133: 117: 105: 103: 99: 97: 96:Kalman filter 93: 88: 84: 76: 74: 72: 68: 64: 60: 56: 52: 48: 44: 40: 33: 19: 1784: 1780: 1767: 1745: 1741: 1724: 1697: 1693: 1673: 1667: 1657: 1645: 1636: 1632: 1621: 1597:Applications 1552: 1356: 1025: 775: 771: 766: 762: 760: 755: 753: 504: 346: 344: 176: 109: 100: 80: 55:Bayes filter 54: 50: 36: 18:Bayes filter 1787:: 498–508. 77:In robotics 61:an unknown 1804:Categories 1748:(1): 1–69. 1663:prediction 166:of a HMM. 59:estimating 43:statistics 1702:CiteSeerX 1652:smoothing 1641:filtering 1581:α 1513:− 1425:∫ 1411:− 1334:− 1246:∝ 1232:− 1176:− 1006:− 979:− 948:− 919:− 879:∫ 865:− 794:− 728:− 629:∏ 579:… 539:… 425:… 414:− 322:− 259:… 248:− 225:− 1680:robotics 130:are the 1676:control 1646:current 763:predict 1704:  1668:future 767:update 45:, and 1772:(PDF) 1758:(PDF) 134:of a 106:Model 87:robot 1678:and 1658:past 765:and 1789:doi 1785:108 1746:182 1729:doi 1712:doi 1616:PDF 67:PDF 37:In 1806:: 1783:. 1744:. 1710:. 1698:50 1696:. 1682:. 811:. 98:. 73:. 49:, 41:, 1795:. 1791:: 1735:. 1731:: 1718:. 1714:: 1561:x 1536:k 1530:x 1524:d 1521:) 1516:1 1510:k 1507:: 1504:1 1498:z 1491:| 1485:k 1479:x 1473:( 1470:p 1467:) 1462:k 1456:x 1449:| 1443:k 1437:z 1431:( 1428:p 1422:= 1419:) 1414:1 1408:k 1405:: 1402:1 1396:z 1389:| 1383:k 1377:z 1371:( 1368:p 1342:) 1337:1 1331:k 1328:: 1325:1 1319:z 1312:| 1306:k 1300:x 1294:( 1291:p 1288:) 1283:k 1277:x 1270:| 1264:k 1258:z 1252:( 1249:p 1240:) 1235:1 1229:k 1226:: 1223:1 1217:z 1210:| 1204:k 1198:z 1192:( 1189:p 1184:) 1179:1 1173:k 1170:: 1167:1 1161:z 1154:| 1148:k 1142:x 1136:( 1133:p 1130:) 1125:k 1119:x 1112:| 1106:k 1100:z 1094:( 1091:p 1085:= 1082:) 1077:k 1074:: 1071:1 1065:z 1058:| 1052:k 1046:x 1040:( 1037:p 1009:1 1003:k 997:x 991:d 987:) 982:1 976:k 973:: 970:1 964:z 957:| 951:1 945:k 939:x 933:( 930:p 927:) 922:1 916:k 910:x 903:| 897:k 891:x 885:( 882:p 876:= 873:) 868:1 862:k 859:: 856:1 850:z 843:| 837:k 831:x 825:( 822:p 797:1 791:k 787:x 776:k 772:k 770:( 756:x 739:. 736:) 731:1 725:i 719:x 712:| 706:i 700:x 694:( 691:p 688:) 683:i 677:x 670:| 664:i 658:z 652:( 649:p 644:k 639:1 636:= 633:i 625:) 620:0 614:x 608:( 605:p 602:= 599:) 594:k 588:z 582:, 576:, 571:1 565:z 559:, 554:k 548:x 542:, 536:, 531:0 525:x 519:( 516:p 490:) 485:k 479:x 472:| 466:k 460:z 454:( 451:p 448:= 445:) 440:0 434:x 428:, 422:, 417:1 411:k 405:x 399:, 394:k 388:x 381:| 375:k 369:z 363:( 360:p 347:k 330:) 325:1 319:k 313:x 306:| 300:k 294:x 288:( 285:p 282:= 279:) 274:0 268:x 262:, 256:, 251:2 245:k 239:x 233:, 228:1 222:k 216:x 209:| 203:k 197:x 191:( 188:p 146:x 118:z 65:( 34:. 20:)

Index

Bayes filter
Naive Bayes spam filtering
probability theory
statistics
machine learning
estimating
probability density function
PDF
Bayesian statistics
computer science
robot
normally distributed
Kalman filter
manifestations
hidden Markov model
Markov process
Bayesian network
Hidden Markov model
Kalman filter
multivariate normal distributions
Particle filter
PDF
smoothing
control
robotics
CiteSeerX
10.1.1.117.1144
doi
10.1109/78.978374
doi

Text is available under the Creative Commons Attribution-ShareAlike License. Additional terms may apply.