Linear algebra is one of the most basic foundations of a wide range of scientific domains, and most textbooks of linear algebra are written by mathematicians.
Real-time or applied digital signal processing courses are offered as follow-ups to conventional or theory-oriented digital signal processing courses in many engineering programs for the purpose of teaching students the technical know-how for putting signal processing algorithms or theory into practical use.
This book focuses on a class of single-channel noise reduction methods that are performed in the frequency domain via the short-time Fourier transform (STFT).
Immediately following the Second World War, between 1947 and 1955, several classic papers quantified the fundamentals of human speech information processing and recognition.
The Kalman filter is the Bayesian optimum solution to the problem of sequentially estimating the states of a dynamical system in which the state evolution and measurement processes are both linear and Gaussian.
Real-time or applied digital signal processing courses are offered as follow-ups to conventional or theory-oriented digital signal processing courses in many engineering programs for the purpose of teaching students the technical know-how for putting signal processing algorithms or theory into practical use.
In these notes, we introduce particle filtering as a recursive importance sampling method that approximates the minimum-mean-square-error (MMSE) estimate of a sequence of hidden state vectors in scenarios where the joint probability distribution of the states and the observations is non-Gaussian and, therefore, closed-form analytical expressions for the MMSE estimate are generally unavailable.
This book is intended to fill the gap between the "e;"e;ideal precision"e;"e; digital signal processing (DSP) that is widely taught, and the limited precision implementation skills that are commonly required in fixed-point processors and field programmable gate arrays (FPGAs).
The purpose of this lecture book is to present the state of the art in nonlinear blind source separation, in a form appropriate for students, researchers and developers.
Spectral estimation is important in many fields including astronomy, meteorology, seismology, communications, economics, speech analysis, medical imaging, radar, sonar, and underwater acoustics.
Image fusion in remote sensing or pansharpening involves fusing spatial (panchromatic) and spectral (multispectral) images that are captured by different sensors on satellites.
The book is based on the material originally developed for the course on Virtual Reality, which the author was teaching at Tampere University of Technology, as well as course on Virtual Environments that the author had prepared for the University for Advancing Studies at Tempe, Arizona.
The last few years have witnessed fast development on dictionary learning approaches for a set of visual computing tasks, largely due to their utilization in developing new techniques based on sparse representation.
Video context analysis is an active and vibrant research area, which provides means for extracting, analyzing and understanding behavior of a single target and multiple targets.
This book explains the stages necessary to create a wavelet compression system for images and describes state-of-the-art systems used in image compression standards and current research.
Earth observation is the field of science concerned with the problem of monitoring and modeling the processes on the Earth surface and their interaction with the atmosphere.
The sequel to the popular lecture book entitled Biomedical Image Analysis: Tracking, this book on Biomedical Image Analysis: Segmentation tackles the challenging task of segmenting biological and medical images.
This lecture presents research on a general framework for perceptual organization that was conducted mainly at the Institute for Robotics and Intelligent Systems of the University of Southern California.
This book presents an overview of the guidelines and strategies for transitioning an image or video processing algorithm from a research environment into a real-time constrained environment.
An important merit of the MPEG-4 video standard is that it not only provided tools and algorithms for enhancing the compression efficiency of existing MPEG-2 and H.
This Lecture book is about objective image quality assessment-where the aim is to provide computational models that can automatically predict perceptual image quality.
The recognition of humans and their activities from video sequences is currently a very active area of research because of its applications in video surveillance, design of realistic entertainment systems, multimedia communications, and medical diagnosis.
The adaptive configuration of nodes in a sensor network has the potential to improve sequential estimation performance by intelligently allocating limited sensor network resources.
While it may be attractive to view sensors as simple transducers which convert physical quantities into electrical signals, the truth of the matter is more complex.
The availability of inexpensive, custom, highly integrated circuits is enabling some very powerful systems that bring together sensors, smart phones, wearables, cloud computing, and other technologies.