

Website deprecated and outdated. Click here for the new site.  
Dept. of Computer Sc. » Pattern Recognition » Research » Groups » Computer Vision » Color and Reflectance
Color and ReflectanceThe color of an illuminant heavily influences the color appearance of digitally captured objects. Color features for high level computer vision task like recognition, tracking and detection of objects are therefore inherently dependent on the illumination conditions. The problem of neutralizing the effects of the illumination color is known as color constancy. It is typically assumed that the knowledge of the color of the dominant light source suffices to correct the image for the most disturbing illumination effects. This procedure is also known as recovering the socalled intrinsic image of a scene. In the 1980's, Shafer introduced the Dichromatic Reflectance Model, which is the most widely used physicsbased reflectance model for computer vision tasks. Algorithms for various challenges like image segmentation, specularity removal and illumination color estimation use this model to express the relationship between light and surface material and geometry. A common critique with physicsbased methodologies is that they are "cute", but tend to break in realworld scenes due to their simplifying assumptions and the constraints they impose on the scene. In our group, we examine ways to relax these constraints in order to make physicsbased methodologies applicable to natural images. As a first result, we present a novel illumination estimation method that takes different votes on the illuminant color and makes a consensus decision afterwards.
CPCV 2011 (ICCV Workshop)
Illuminant Color Estimation for RealWorld MixedIlluminant ScenesWe present a physicsbased approach for illuminant color estimation of arbitrary images, which is explicitly designed for handling images with multiple illuminants [1]. An illuminant color estimate is obtained independently from distinct image miniregions. From these miniregions a robust local illumination color is computed by consensus. These local estimates are then used in deriving the chromaticity of the dominant illuminants. Experiments on an established benchmark database of realworld images show that our technique performs comparable to uniformilluminant estimation methods. Furthermore, extensive tests on realworld images show that we can reliably process mixedilluminant scenes. The source code is available here!
References:
CPCV 2011 (ICCV Workshop)
Color Constancy and NonUniform Illumination: Can Existing Algorithms Work?A typical assumption of todays color constancy algorithms is globally uniform illumination. However, more often than not, this assumption is an insufficient approximation of realworld illumination conditions. Thus, illumination should be locally determined, taking under consideration that multiple illuminants may be present. In this paper [1] we investigate the suitability of adapting five stateoftheart color constancy methods so that they can be used for local illuminant estimation. Each of the methods is applied independently on local pathes. We then evaluated different fusion methodologies. Our experiments indicate that the best performance is obtained by fusion strategies that combine the outputs of the estimators using regression. See the project page for additional details, source code and data.
References:
ICIP 2009
PhysicsBased Illuminant Color Estimation as an Image Semantics ClueWe investigated the behavior of the inverseintensity chromaticity (IIC) illuminant estimation by Tan et al. [1] on realworld images downloaded from the web. The method consists of two parts: first, a specularity segmentation, and second, the estimation of the illuminant color from the segmented pixels. We found that the specularity segmentation might complicate the estimation. Furthermore, we did experiments towards a confidence measure for the illuminant color estimation. In IICspace, the illuminant color is determined by finding the prevalent yaxis intercept of connecting lines of the points in this space. Tan et al. proposed to do this via Hough transform. Our idea was to examine the distributions in Hough space as a way towards automated selfassessment. Indeed, the distinctiveness peak in this distribution appears to give a hint on how well the specularitysegmented pixels fit the underlying physicsbased model [2]. In future work, we aim to expand on the analysis of these distributions towards a more robust confidence measure for this illumination estimation in IIC space.
References:
CRICV 2009 (ICCV Workshop)
A Common Framework for Ambient Illumination in the Dichromatic Reflectance ModelThe dichromatic reflectance model, originally proposed by Shafer [1], is the core of most physicsbased algorithms for illumination color estimation. Maxwell et al. [2] recently proposed an extension of this model, the BiIlluminant Dichromatic Reflectance (BIDR) model. This extension contains a term for a second illuminant. Interestingly, this extension can be seen as a theoretical fundament of several algorithms that deal with two illuminant cases. Most prominently, shadow detection algorithms often model ambient illumination as a second global light source. In our paper [3], we show connections between such algorithms via the BiIlluminant Dichromatic Reflectance Model. Our ultimate goal is to find a way to exploit the BIDR model directly, in order to obtain more general methods for handling twoilluminant scenarios.
References:
