Results 1 
6 of
6
Information Theory and Statistics
, 1968
"... Entropy and relative entropy are proposed as features extracted from symbol sequences. Firstly, a proper Iterated Function System is driven by the sequence, producing a fractaMike representation (CSR) with a low computational cost. Then, two entropic measures are applied to the CSR histogram of th ..."
Abstract

Cited by 1317 (0 self)
 Add to MetaCart
Entropy and relative entropy are proposed as features extracted from symbol sequences. Firstly, a proper Iterated Function System is driven by the sequence, producing a fractaMike representation (CSR) with a low computational cost. Then, two entropic measures are applied to the CSR histogram of the CSR and theoretically justified. Examples are included.
Deterministic Builtin Pattern Generation for Sequential Circuits
, 1999
"... We present a new pattern generation approach for deterministic builtin self testing (BIST) of sequential circuits. Our approach is based on precomputed test sequences, and is especially suited to sequential circuits that contain a large number of flipflops but relatively few controllable primary i ..."
Abstract

Cited by 17 (8 self)
 Add to MetaCart
We present a new pattern generation approach for deterministic builtin self testing (BIST) of sequential circuits. Our approach is based on precomputed test sequences, and is especially suited to sequential circuits that contain a large number of flipflops but relatively few controllable primary inputs. Such circuits, often encountered as embedded cores and as filters for digital signal processing, are difficult to test and require long test sequences. We show that statistical encoding of precomputed test sequences can be combined with lowcost pattern decoding to provide deterministic BIST with practical levels of overhead. Optimal Huffman codes and nearoptimal Comma codes are especially useful for test set encoding. This approach exploits recent advances in automatic test pattern generation for sequential circuits and, unlike other BIST schemes, does not require access to a gatelevel model of the circuit under test. It can be easily automated and integrated with design automation tools. Experimental results for the ISCAS 89 benchmark circuits show that the proposed method provides higher fault coverage than pseudorandom testing with shorter test application time and low to moderate hardware overhead.
A Computational Theory of the Firm
 Journal of Economic Behavior and Organization
, 2002
"... This paper proposes using computational learning theory (CLT) as a framework for analyzing the information processing behavior of firms; we argue that firms can be viewed as learning algorithms. The costs and benefits of processing information are linked to the structure of the firm and its relation ..."
Abstract

Cited by 9 (6 self)
 Add to MetaCart
This paper proposes using computational learning theory (CLT) as a framework for analyzing the information processing behavior of firms; we argue that firms can be viewed as learning algorithms. The costs and benefits of processing information are linked to the structure of the firm and its relationship with the environment. Wemodel the firm as a type of artificial neural network (ANN).By a simulation experiment, we show which types of networks maximize the net return to computation given different environments. 2002 Elsevier Science B.V. All rights reserved.
Waveletbased medical image compression
"... In view of the increasingly important role played by digital medical imaging in modern health care and the consequent blow up in the amount of image data that have to be economically stored and/or transmitted, the need for the development of image compression systems that combine high compression pe ..."
Abstract

Cited by 1 (0 self)
 Add to MetaCart
In view of the increasingly important role played by digital medical imaging in modern health care and the consequent blow up in the amount of image data that have to be economically stored and/or transmitted, the need for the development of image compression systems that combine high compression performance and preservation of critical information is ever growing. A powerful compression scheme that is based on the stateoftheart in waveletbased compression is presented in this paper. Compression is achieved via efficient encoding of wavelet zerotrees (with the embedded zerotree wavelet (EZW) algorithm) and subsequent entropy coding. The performance of the basic version of EZW is improved upon by a simple, yet effective, way of a more accurate estimation of the centroids of the quantization intervals, at a negligible cost in side information. Regarding the entropy coding stage, a novel RLEbased coder is proposed that proves to be much simpler and faster yet only slightly worse than contextdependent adaptive arithmetic coding. A useful and flexible compromise between the need for high compression and the requirement for preservation of selected regions of interest is provided through two intelligent, yet simple, ways of achieving the socalled selective compression. The use of the lifting scheme in achieving compression that is guaranteed to be lossless in the presence of numerical inaccuracies is being investigated with interesting preliminary results. Experimental results are presented that verify the superiority of our scheme over conventional block transform coding techniques (JPEG) with respect to both objective and subjective criteria. The high potential of our scheme for progressive transmission, where the regions of interest are given the highest priority, is also demonstrated. c○1999 Elsevier Science B.V.
Information and Distinguishability of Ensembles of Identical Quantum States
, 2008
"... We consider a fixed quantum measurement performed over n identical copies of quantum states. Using a rigorous notion of distinguishability based on Shannon’s 12th theorem, we show that in the case of a single qubit the number of distinguishable states is W(α1, α2, n) = α1 − α2  √ 2n, where (α1, α ..."
Abstract
 Add to MetaCart
We consider a fixed quantum measurement performed over n identical copies of quantum states. Using a rigorous notion of distinguishability based on Shannon’s 12th theorem, we show that in the case of a single qubit the number of distinguishable states is W(α1, α2, n) = α1 − α2  √ 2n, where (α1, α2) is the angle interval from which the states πe are chosen. In the general case of an Ndimensional Hilbert space and an area Ω of the domain on the unit sphere from which the states are chosen, the number of distinguishable states is W(N, n,Ω) = Ω ( 2n πe N −1
Mean field annealing Density estimation
"... Independent factor analysis Kullback–Leibler (KL) divergence of independent components. By using Potts encoding, we express the KL divergence in an approximating form, which is shown to be tractable with respect to the WPW parameters as well as the demixing matrix and can be minimized by two interac ..."
Abstract
 Add to MetaCart
Independent factor analysis Kullback–Leibler (KL) divergence of independent components. By using Potts encoding, we express the KL divergence in an approximating form, which is shown to be tractable with respect to the WPW parameters as well as the demixing matrix and can be minimized by two interactive dynamic modules derived by the annealed expectationmaximization method and the natural gradient descent method, respectively. By numerical simulations, we test the proposed ICA algorithm with observations separately sampled from linear mixtures of independent sources and real world signals, including fetal electrocardiograms, mixed facial images and eventrelated potentials, extensively showing its accuracy and reliability for independent component analysis in comparison with some other popular ICA algorithms. © 2008 Elsevier Ltd. All rights reserved. 1.