Abstract
Recently, O'Sullivan introduced roughness penalties for use in stochastic problems where the use of Markov random fields may not arise naturally. In this study, these penalties are used for the deterministic problem. θ ε Rp is considered to be the vector of parameters to be estimated. The available data are ym = Σn=1N Hmn Xn, 1 ≤m ≤ M where y ε R+M, Hmn ≥ 0 and x ε R+N depends on θ. The manner in which x depends on θ yields slightly different algorithms. The matrix H is assumed to have at least one positive entry in each column. It is shown that x may be considered as the complete data for θ. The incomplete data I-divergence is shown to equal an averaged complete data I-divergence plus an additional term. The deterministic FM algorithm then consists of minimizing the averaged complete data I-divergence. Lastly, a maximum entropy penalty and a roughness penalty are incorporated into the problem.
| Original language | English |
|---|---|
| Pages | 177 |
| Number of pages | 1 |
| State | Published - 1995 |
| Event | Proceedings of the 1995 IEEE International Symposium on Information Theory - Whistler, BC, Can Duration: Sep 17 1995 → Sep 22 1995 |
Conference
| Conference | Proceedings of the 1995 IEEE International Symposium on Information Theory |
|---|---|
| City | Whistler, BC, Can |
| Period | 09/17/95 → 09/22/95 |