This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Name Your Colour For the Task: Artificially Discover Colour Naming via Colour Quantisation Transformer

Shenghan Su1, Lin Gu3,2111Corresponding author. , Yue Yang1,4, Zenghui Zhang1, Tatsuya Harada2,3
1Shanghai Jiao Tong University, 2The University of Tokyo, 3RIKEN AIP, 4Shanghai AI Laboratory
{su2564468850, yang-yue, zenghui.zhang}@sjtu.edu.cn, [email protected], [email protected]
Abstract

The long-standing theory that a colour-naming system evolves under dual pressure of efficient communication and perceptual mechanism is supported by more and more linguistic studies, including analysing four decades of diachronic data from the Nafaanra language. This inspires us to explore whether machine learning could evolve and discover a similar colour-naming system via optimising the communication efficiency represented by high-level recognition performance. Here, we propose a novel colour quantisation transformer, CQFormer, that quantises colour space while maintaining the accuracy of machine recognition on the quantised images. Given an RGB image, Annotation Branch maps it into an index map before generating the quantised image with a colour palette; meanwhile the Palette Branch utilises a key-point detection way to find proper colours in the palette among the whole colour space. By interacting with colour annotation, CQFormer is able to balance both the machine vision accuracy and colour perceptual structure such as distinct and stable colour distribution for discovered colour system. Very interestingly, we even observe the consistent evolution pattern between our artificial colour system and basic colour terms across human languages. Besides, our colour quantisation method also offers an efficient quantisation method that effectively compresses the image storage while maintaining high performance in high-level recognition tasks such as classification and detection. Extensive experiments demonstrate the superior performance of our method with extremely low bit-rate colours, showing potential to integrate into quantisation network to quantities from image to network activation. The source code is available at https://github.com/ryeocthiv/CQFormer

Refer to caption
Figure 1: (a) The theoretical limit of efficiency for colour naming (black curve) and cases of the WCS probability map of human colour language copied from [49]. (b) The colour size (from 1-bit to 6-bit)-accuracy curve on the tiny-imagenet-200 [29] dataset. The WCS probability maps generated by our CQFormer are also shown along the curve. (c) The colour naming stimulus grid used in the WCS [26]. (d) The three-term WCS probability map of CQFormer after embedding 1978 Nafaanra three-colour system ((light (‘fiNge’), dark (‘wOO’), and warm or red-like (‘nyiE’)) into the latent representation. (e) The four-term WCS probability map of CQFormer evolved from (d). The evolved fourth colour, yellow-green, is consistent with the prediction of basic colour term theory [2].

1 Introduction

Hath not a Jew eyes? Hath not a Jew hands, organs,

dimensions, senses, affections, passions?

 

William Shakespeare “The Merchant of Venice”

Does artificial intelligence share the same perceptual mechanism for colours as human beings? We aim to explore this intriguing problem from machine learning perspective.

Colour involves the visual reception and neural registering of light stimulants when the spectrum of light interacts with cone cells in the eyes. Physical specifications of colour also include the reflective properties of the physical objects, geometry incident illumination, etc. By defining a colour space [18], people could identify colours directly according to these quantifiable coordinates.

Compared to the pure physiological nature of hue categorisation, the complex phenomenon of colour naming or colour categorisation spans multiple disciplines, from cognitive science to anthropology. Solid diachronic research [2] also suggests that human languages constantly evolve to acquire new colour names, resulting in an increasingly fine-grained colour naming system. This evolutionary process is hypothesised to be under the pressure of communication efficiency and perceptual structure. Communication efficiency requires shared colour partitioning to be communicated accurately with a lexicon as simple and economical as possible. Colour perceptual structure is relevant to human perception in nature. For example, the colour space distance between nearby colours should correspond to their perceptual dissimilarity. This structure of perceptual colour space has long been used to explain colour naming patterns across languages. A recent analysis of human colour naming systems, especially in Nafaanra, contributes very direct evidence to support this hypothesis through the employment of Shannon’s communication model [41]. Interestingly, this echos the research on colour quantisation, which quantises colour space to reduce the number of distinct colours in an image.

Traditional colour quantisation methods [23, 19, 17] are perception-centred and generate a new image that is as visually perceptually similar as possible to the original image. These methods group similar colours in the colour space and represent each group with a new colour, thus naturally preserving the perceptual structure. Instead of prioritising the perceptual quality, Hou et al. [25] proposed a task-centred/machine-centred colour quantisation method, ColorCNN, focusing on maintaining network classification accuracy in the restricted colour spaces. While achieving an impressive classification accuracy on even a few-bit image, ColorCNN only identifies and preserves machine-centred structure without directly clustering similar colours in the colour space. Therefore, this pure machine-centred strategy sacrifices perceptual structure and often associates nearby colours with different quantised indices.

Zaslavsky et al. [49] measure the communication efficiency in colour naming by analysing the informational complexity based on the information bottleneck (IB) principle. Here, we argue that the network recognition accuracy also reflects the communication efficiency when the number of colours is restricted. Since the human colour naming is shaped by both perception structure and communication efficiency [51], we integrate the need for both perception and machine to propose a novel end-to-end colour quantisation transformer, CQFormer, to discover the artificial colour naming systems.

As illustrated in Fig. 1 (b), the recognition accuracy increases with the number of colours in our discovered colour naming system. Surprisingly, the complexity-accuracy trade-offs are similar to the numerical results (Fig. 1 (a)) independently derived from linguistic research  [49]. What is more, after embedding 1978 Nafaanra three-colour system (Nafaanra-1978) into the latent representation of CQFormer (Fig. 1 (d)), our method automatically evolves the fourth colour closed to yellow-green, matching the basic colour terms theory [2] summarised in different languages. Berlin and Kay found universal restrictions on colour naming across cultures and claimed languages acquire new basic colour categories in a strict chronological sequence. For example, if a culture has three colours (light (‘fiNge’), dark (‘wOO’), and warm or red-like (‘nyiE’) in Nafaanra), the fourth colour it evolves should be yellow or green, exactly the one (Fig. 1 (e)) discovered by our CQFormer.

The pipeline of CQFormer, shown in Fig. 2, comprises two branches: Annotation Branch and Palette Branch. Annotation Branch annotates each pixel of the input RGB image with the proper quantised colour index before painting the index map with the corresponding colour in the colour palette. We localise the colour palette in the whole RGB colour space with a novel Palette Branch, which detects the key-point with explicit attention queries of transformer. During the training stage, as illustrated in the red line and black line of Fig. 2 (a), Palette Branch interacts with an input image and Reference Palette Queries to maintain the perceptual structure by reducing the perceptual structure loss. This perception-centred design groups similar colours and ensures the colour palette sufficiently represents the colour naming system defined by the World Color Survey (WCS) colour naming stimulus grids. As shown in Fig. 2 (b), each item in the colour palette (noted by an asteroid) lies in the middle of the corresponding colour distribution in the WCS colour naming probability map. Finally, the quantised image is passed to a high-level recognition module for machine accuracy tasks such as classification and detection. Through the joint optimisation of CQFormer and consequent high-level module, we can balance both perception and machine. Besides automatically discovering the colour naming system, our CQFormer also offers an effective solution to extremely compress image storage while maintaining high performance in high-level recognition tasks. For example, the CQFormer achieves 50.6% top-1 accuracy on the CIFAR100 [28] dataset with only a 1-bit colour space (i.e., two colours). The extremely low-bit quantisation of our also demonstrates the potential to integrate into quantisation network research [40, 47], allowing the end-to-end optimisation from image to weight and activation.

Our contributions could be summarised as follows:

  • We propose a novel end-to-end colour quantisation transformer, CQFormer, to artificially discover a colour naming system by counting both perception and machine. The discovered colour naming system shows a similar pattern to human language on colour.

  • We propose a novel colour quantisation method taking palette generation as an attention-based key-point detection task. With the input of independent attention queries, it automatically generates 3D coordinates as the selected colour in the whole colour space.

  • Our colour quantisation achieves superior classification and object detection performance with extremely low bit-rate colours.

2 Related Works

Colour Quantisation. Colour quantisation, also known as optimal palette generation, compresses images by remapping original pixel colours to a limited set of entries in a small palette.

Traditional colour quantisation [37, 14, 32, 46] aims to reduce the colour space while preserving visual fidelity. These perception-centred methods usually cluster colours to create a new image as visually perceptually similar as possible to the original image. For example, MedianCut [23] and OCTree [19] are two representative clustering-based algorithms. Dithering [16] eliminates visual artefacts by including a noise pattern. The colour-quantised images can be expressed as indexed colour [39], and encoded with PNG  [3].

There are also much of efforts [7, 8] to generate the optimal colour palette and recolour the image. For example, Bahng et al. [1] proposed the text-to-palette generation networks to generate an appropriate palette according to the semantics of the text input. Yoo et al. [48] leverage the memory networks to retrieve the feature of colour palette for colourisation with limited data. Li et al. [30] develop a self-supervised approach to recolouring images from design-oriented fields by reproducing colour palettes with luminance distributions different from the input.

Recently, Hou et al. [25] propose a pure machine-centred CNN-based colour quantisation method, ColorCNN, which effectively maintains the informative structures under an extremely constrained colour space. In addition to colour quantisation, Camposeco et al. [4] also design a task-centred image compression method for localisation in 3D map data. Since human colour naming reflects both perceptual structure and communicative need, our CQFormer also considers both perception and machine to discover the colour naming system artificially.

World Color Survey. The World Color Survey (WCS) [26] comprises colour name data from 110 languages of non-industrialised societies [52], w.r.t. the stimulus grid shown in Fig. 1 (c). There are 320 colour chips in colour naming stimulus grids, and each chip is at its maximum saturation for that hue/lightness combination, while columns correspond to equally spaced Munsell hues and rows to equally spaced Munsell values. Participants were asked to name the colour of each chip to record the colour naming system, generating the WCS probability map for human language (e.g. the Nafaanra-1978 in Fig. 3 (c)).

Colour Categorisation/Naming. Van De Weijer et al. [45] gathered datasets from Google and Ebay with the explicit aim of learning colour names from real-world images. Parraga and Akbarinia[38] then recruited 17 paid subjects and employed a psychophysical experiment to bridge the gap between the physiology of the visual system and colour categorisation. Gibson et al. [20] indicated that warm colours are communicated more efficiently than cool colours in general, and crucially, categorical variations between languages are attributed to differences in the practical value of colour. Siuda-Krzywicka et al.[42] indicates the independence of colour categorisation and naming, while also offering a plausible neural foundation for the process of colour naming. Chaabouni et al.[6] treated communicative concepts as the primary driving force behind the formation of colour categories and demonstrated that continuous message passing increased system complexity and reduced efficiency. De Vries et al. [12] demonstrated that an artificial visual system develops colour categorical boundaries resembling those of humans by learning to recognise objects in images.

Refer to caption
Figure 2: (a) Overview of our CQFormer (taking 3-colour quantisation as an example). Red lines are only for training stage, purple lines are only for testing stage, and black lines are both used in both stage. (b) WCS colour probability map counted across all pixels of the input image. (c) Detailed structure of PAM. (d) Detailed structure of FFN.

3 Methodology

3.1 Problem Formulation

For a dataset 𝒟\mathcal{D} containing image-label pairs (𝒙,y)(\boldsymbol{x},y), the recognition network fθ()f_{\theta}(\cdot) takes the input image 𝒙\boldsymbol{x} and predicts its label y^=fθ(x)\hat{y}=f_{\theta}(x) (e.g., the class probability in a classification task). The network’s parameters θ\theta can be optimised by minimising the loss between the predicted label y^\hat{y} and the ground truth label yy, defined as the machine-centred loss M\mathcal{L}_{M}. This process allows us to find the optimal parameter set θ\theta^{\star}.

θ=argmin𝜃(𝒙,y)𝒟M(y,fθ(𝒙)).\theta^{\star}=\underset{\theta}{\arg\min}\sum_{(\boldsymbol{x},y)\in\mathcal{D}}\mathcal{L}_{M}\left(y,f_{\theta}(\boldsymbol{x})\right). (1)

We aim to discover the artificial colour naming system that balances the need for both machine accuracy and human perception. To achieve this, CQFormer focuses not only on recognition accuracy but also on preserving the perceptual structure of the image. The function in Eq. 1 can be rewritten as follows:

ψ,θ=argminψ,θ(𝒙,y)𝒟M(y,fθ(gψ(𝒙)))+P,\psi^{\star},\theta^{\star}=\underset{\psi,\theta}{\arg\min}\sum_{(\boldsymbol{x},y)\in\mathcal{D}}\mathcal{L}_{M}\left(y,f_{\theta}\left(g_{\psi}(\boldsymbol{x})\right)\right)+\mathcal{L}_{P}, (2)

where ψ\psi and θ\theta represent the parameters of the CQFormer gg and the high-level recognition network ff, respectively. We jointly optimise gg and ff to find the optimal parameters ψ\psi^{\star} and θ\theta^{\star}. Additionally, P\mathcal{L}_{P} is a perceptual structure loss that is focused on preserving the perceptual structure of the image and will be further explained in detail in Sec. 3.4.

3.2 CQFormer Architecture

Overall Architecture. An overview of the CQFormer is illustrated in Fig. 2 (a). It consists of two main branches: (1) Annotation Branch, which assigns a quantised colour index to each pixel of the input RGB image, and (2) Palette Branch, which is responsible for generating a suitable colour palette.

Annotation Branch of CQFormer takes an original image 𝒙H×W×3\boldsymbol{x}\in\mathbb{R}^{H\times W\times 3} as input, where HH and WW are the height and width of the image, respectively. During the training stage, it generates a probability map mτ(𝒙)H×W×Cm_{\tau}(\boldsymbol{x})\in\mathbb{R}^{H\times W\times C}, where CC is the number of quantised colours, and τ>0\tau>0 is the temperature parameter of the Softmax function [22]. During the testing stage, it produces a one-hot colour index map Onehot(M(𝒙))H×W×C\operatorname{One-hot}(M(\boldsymbol{x}))\in\mathbb{R}^{H\times W\times C}, where each pixel of the image is assigned a single colour index among the CC quantised colours.

Palette Branch of CQFormer takes the original image 𝒙\boldsymbol{x} and Reference Palette Queries QC×d\textbf{Q}\in\mathbb{R}^{C\times d} as input. These queries are composed of CC learnable vectors of dimension dd, each representing an automatically mined colour. The queries Q interact with the keys K(HW16)×d\textbf{K}\in\mathbb{R}^{(\frac{HW}{16})\times d} and values V(HW16)×d\textbf{V}\in\mathbb{R}^{(\frac{HW}{16})\times d}, generated from the input image 𝒙\boldsymbol{x} to produce the colour palette P(𝒙)C×3P(\boldsymbol{x})\in\mathbb{R}^{C\times 3}. This palette consists of CC triples (R,G,B)(R,G,B), each representing one of the machine-discovered CC colours.

Finally, CQFormer produces the quantised image by performing a matrix multiplication between mτ(𝒙)m_{\tau}(\boldsymbol{x}) and P(𝒙)P(\boldsymbol{x}) during the training stage. During the testing stage, it is obtained from Onehot(M(𝒙))\operatorname{One-hot}(M(\boldsymbol{x})) and P(𝒙)P(\boldsymbol{x}). The quantised image is then fed into the high-level recognition module for object detection and classification tasks.

Annotation Branch. The first component of Annotation Branch is a UNeXt encoder [44] that generates per-pixel categories. Given the input image 𝒙\boldsymbol{x}, the encoder produces a class activation map 𝒙activationH×W×C\boldsymbol{x}_{activation}\in\mathbb{R}^{H\times W\times C}, which contains crucial and semantically rich features. Then, different approaches are used to process the class activation map during testing and training.

(1) Testing stage. As shown by the purple lines in Fig.2 (a), we use the 𝒙activation\boldsymbol{x}_{activation} as the input to a Softmax function over CC channels, and then apply an argmax\arg\max function to produce a colour index map M(𝒙)H×WM(\boldsymbol{x})\in\mathbb{R}^{H\times W}:

M(𝒙)=argmax𝐶(Softmax(𝒙activation)).\displaystyle M(\boldsymbol{x})=\underset{C}{\arg\max}(\operatorname{Softmax}(\boldsymbol{x}_{activation})). (3)

Subsequently, the colour index map is transformed into a one-hot encoding, denoted as Onehot(M(𝒙))H×W×C\operatorname{One-hot}(M(\boldsymbol{x}))\in\mathbb{R}^{H\times W\times C}, which is then combined with the colour palette P(𝒙)P(\boldsymbol{x}) through matrix multiplication, resulting in the generation of the test-time colour-quantised image 𝒙¯\bar{\boldsymbol{x}}.

𝒙¯=Onehot(M(𝒙))P(𝒙),\displaystyle\bar{\boldsymbol{x}}=\operatorname{One-hot}(M(\boldsymbol{x}))\otimes P(\boldsymbol{x}), (4)

where \otimes represents matrix multiplication.

(2) Training stage. As depicted by the red lines in Fig.2 (a), since the argmax\arg\max function is not differentiable, we use the Softmax function as a substitute during the training stage. To prevent overfitting, we incorporate a temperature parameter τ\tau [22] into the Softmax function, pushing the probability map distribution closer to a one-hot vector. This results in the probability map mτ(𝒙)m_{\tau}(\boldsymbol{x}) with temperature parameter τ\tau, computed as:

mτ(𝒙)=Softmax(𝒙activationτ).\displaystyle m_{\tau}(\boldsymbol{x})=\operatorname{Softmax}(\frac{\boldsymbol{x}_{activation}}{\tau}). (5)

As extensively discussed in [22], the output is similar to a one-hot vector with large diversity when the temperature is low (0<τ<10<\tau<1) and a uniform distribution with small diversity otherwise (τ>1\tau>1). Therefore, we set 0<τ<10<\tau<1 and the train-time colour-quantised image 𝒙~\widetilde{\boldsymbol{x}} is generated as:

𝒙~=mτ(𝒙)P(𝒙).\displaystyle\widetilde{\boldsymbol{x}}=m_{\tau}(\boldsymbol{x})\otimes P(\boldsymbol{x}). (6)

Palette Branch. We locate the representative colours by using an attention-based key-point detection strategy, which is originally designed to use transformer queries to automatically find the key-point location, such as bounding-boxes [5], human poses [31] and colour matrix with gamma [11]. In other words, we reformulate the problem of colour quantisation into a 3D spatial key-point localisation task within the whole RGB colour space.

Given the input image 𝒙\boldsymbol{x}, we first extract a high-dimensional lower resolution feature F0H4×W4×dF_{0}\in\mathbb{R}^{\frac{H}{4}\times\frac{W}{4}\times d} using two stacked convolution layers. The F0F_{0} is then passed to the Palette Acquisition Module (PAM) to acquire the colour palette P(𝒙)P(\boldsymbol{x}). As shown in Fig. 2 (c), different from DETR [5], our QC×d\textbf{Q}\in\mathbb{R}^{C\times d}, called Reference Palette Queries, are explicit learnable embeddings without extra multi-head self-attention, which attends K and V generated from F0F_{0}. The positional encoding mechanism is a depth-wise convolution [9], which is suitable for different input resolutions. After that, the position-encoded feature is flattened into sequences before being passed into our transformer block. Here, the K and V are generated by two fully-connection (FC) layers separately, and the cross-attention is estimated as:

Attention(Q,K,V)=Softmax(QKd)V.\displaystyle\operatorname{Attention}(\textbf{Q},\textbf{K},\textbf{V})=\operatorname{Softmax}(\frac{\textbf{QK}^{\top}}{\sqrt{d}})\textbf{V}. (7)

After a feed forward network (FFN) [15], including two FC layers and a GELU [24] activation function, then with a residual connection [21], the Reference Palette Queries Q are transformed into output embeddings. We then decode embeddings into 3D coordinates with another FFN and a Sigmoid function, resulting in the final localisation of colour palette P(𝒙)C×3P(\boldsymbol{x})\in\mathbb{R}^{C\times 3} within the whole RGB colour space.

High-Level Recognition Module. Taking the quantised image (𝒙¯\bar{\boldsymbol{x}} or 𝒙~\widetilde{\boldsymbol{x}}) as input, the high-level recognition module predicts the results of object detection or classification.

3.3 WCS Colour Probability Map

As introduced in WCS of Sec. 2, the WCS probability map mWCS8×40×Cm_{\operatorname{WCS}}\in\mathbb{R}^{8\times 40\times C} for human language is collected by the participants’ colour perception in WCS. We also define the WCS colour probability map for machine (see in Fig. 1 (d) (e)) and generate it from the datasets. First, the colour index of each pixel is mapped to the grid with the closest (hue, value) coordinate as the pixel. Second, we count the frequency of occurrence of each colour index in each grid, and select the pixel’s colour index with the highest frequency as the grid’s colour index. Therefore, CC clusters are formed according to the CC grid’s colour indexes in WCS colour probability map. Finally, each cluster is shown in the colour corresponding to the centre of mass of its colour category.

The WCS colour probability map for machine is then used to measure the similarity of colours within the same cluster in Sec.3.4. More importantly, it creates a strong correlation between human language and machine vision since they can be represented in the same format. Therefore, we explore the colour evolution using the WCS colour probability map in Sec.3.5.

3.4 Perceptual Structure Loss

To ensure that CQFormer also produces visually pleasing and structurally sound results, an additional perceptual structure loss P\mathcal{L}_{P} is added to the training process:

P=αRColour+βRDiversity+γPerceptual.\displaystyle\mathcal{L}_{P}=\alpha R_{\operatorname{Colour}}+\beta R_{\operatorname{Diversity}}+\gamma\mathcal{L}_{\operatorname{Perceptual}}. (8)

α\alpha , β\beta and γ\gamma are three non-negative parameters, which combine intra-cluster colour similarity regularisation RColourR_{\operatorname{Colour}}, diversity regularisation RDiversityR_{\operatorname{Diversity}} and perceptual similarity loss Perceptual\mathcal{L}_{\operatorname{Perceptual}}, respectively. Therefore, the objective of training the CQFormer is to minimise the total loss total\mathcal{L_{\operatorname{total}}}:

total=M+P.\displaystyle\mathcal{L_{\operatorname{total}}}=\mathcal{L}_{M}+\mathcal{L}_{P}. (9)

Intra-cluster Colour Similarity Regularisation. The CQFormer associates each pixel of the input image with a colour index and forms CC clusters covering different parts of the WCS colour probability map. To ensure that the pixels within the same cluster are perceptually similar in colour, we propose the intra-cluster colour similarity regularisation RColourR_{\operatorname{Colour}}, which is the mean of the colour variance within the CC clusters. At first, the centroid colour values of the CC clusters {μ1,μ2,,μC}\{\mu_{1},\mu_{2},\dots,\mu_{C}\} are calculated. Then, we compute CC squared colour distances dHSV2\operatorname{d}^{2}_{\operatorname{HSV}} between all pixels and their centroid colour value in each cluster. Here, the dHSV2\operatorname{d}^{2}_{\operatorname{HSV}} is calculated in Munsell HSV colour space as:

dHSV2(h1,s1,v1,h2,s2,v2)=(v2v1)2+s12v12+s22v222s1s2v1v2cos(h2h1).\displaystyle\small\begin{split}&\operatorname{d}^{2}_{\operatorname{HSV}}(h_{1},s_{1},v_{1},h_{2},s_{2},v_{2})\\ &=\left(v_{2}-v_{1}\right)^{2}+s_{1}^{2}v_{1}^{2}+s_{2}^{2}v_{2}^{2}-2s_{1}s_{2}v_{1}v_{2}\cos\left(h_{2}-h_{1}\right).\end{split} (10)

Finally, we take the mean value of the CC squared distances as the RColourR_{\operatorname{Colour}}:

RColour=1C×c=1C1Nci=1NcdHSV2(𝒙c[i],μc),\displaystyle R_{\operatorname{Colour}}=\frac{1}{C}\times\sum_{c=1}^{C}\frac{1}{N_{c}}\sum_{i=1}^{N_{c}}\operatorname{d}^{2}_{\operatorname{HSV}}(\boldsymbol{x}_{c}[i],\mu_{c}), (11)

where NcN_{c} is the number of all pixels in ClustercCluster_{c}, and 𝒙c[i]\boldsymbol{x}_{c}[i] represents the Munsell HSV value of the ii-th pixel.

Diversity Regularisation. To encourage the CQFormer to select at least one pixel of all CC colours, we adopt the diversity regularisation term RDiversityR_{\operatorname{Diversity}} proposed by Hou et al. [25]. Diversity is a simple yet efficient metric and serves as an implicit entropy of the assignment distribution. The RDiversityR_{\operatorname{Diversity}} is calculated as:

RDiversity=log2C×(11C×cmax(u,v)[mτ(𝒙)]u,v).\displaystyle\small R_{\operatorname{Diversity}}=\log_{2}C\times\left(1-\frac{1}{C}\times\sum_{c}\max_{(u,v)}[m_{\tau}(\boldsymbol{x})]_{u,v}\right). (12)

Perceptual Similarity Loss. The perceptual similarity loss, denoted by Perceptual\mathcal{L}_{\operatorname{Perceptual}}, is a mean squared error (MSE) loss between the quantised image 𝒙~\widetilde{\boldsymbol{x}} and the input image 𝒙\boldsymbol{x}. It ensures that each item of the colour palette P(𝒙)P(\boldsymbol{x}) lies in the centre of corresponding colour distribution in the WCS colour probability map (noted by asteroids in Fig. 2 (b)). The Perceptual\mathcal{L}_{\operatorname{Perceptual}} is calculated as:

Perceptual=MSE(𝒙~,𝒙).\displaystyle\mathcal{L}_{\operatorname{Perceptual}}=\mathcal{L}_{\operatorname{MSE}}(\widetilde{\boldsymbol{x}},\boldsymbol{x}). (13)

3.5 Colour Evolution

With CQFormer, we explore the colour evolution based on the classification task, involving two successive stages with different loss functions. Since the CQFormer initially has no prior knowledge of colour naming systems associated with corresponding human languages, the first embedding stage aims to embed the colour perceptual knowledge of a certain language into the latent representation of the CQFormer. For example, the CQFormer first learns and matches the 1978 Nafaanra three-colour system by forcing the CQFormer to output a similar WCS colour probability map to that for Nafaanra. Here, we design two embedding solutions and loss functions, i.e. FullEmbedding\mathcal{L_{\operatorname{Full-Embedding}}} and CentralEmbedding\mathcal{L_{\operatorname{Central-Embedding}}}, to distil either full colour probability map embedding or only representative colours to CQFormer. The second evolution stage then lets CQFormer evolve more colours, i.e. splitting the fourth colour from the learned three-colour system under the pressure of both accuracy and perceptual structure.

Embedding Stage.

Refer to caption
Figure 3: The procedure of colour probability map embedding. (a) The origin RGB image. (b) The WCS colour naming stimulus grids. (c) The mWCSm_{\operatorname{WCS}} of Nafaanra-1978 copied from [49]. (d) The human language probability map after argmax\arg\max function for Nafaanra-1978: argmax(mHuman(𝒙))\arg\max(m_{\operatorname{Human}}(\boldsymbol{x})).

(1) Colour Probability Map Embedding. As illustrated in Fig. 3, this embedding forces our CQFormer to match the identical WCS colour probability map of a certain language. At first, we ensure that the CQFormer outputs the same number of quantised colours as the certain human colour system. For each pixel in the input image, we collect its Munsell (hue, value) coordinate and spatial position coordinate (i,j)(i,j) in the input image. Then we locate this (hue, value) coordinate on the WCS colour probability map of a certain language (Fig. 3 (a)\rightarrow(b)\rightarrow(c)) to find its corresponding CC probability values pHuman(i,j)Cp_{\operatorname{Human}}(i,j)\in\mathbb{R}^{C} of human colour categories,e.g. 9% for ‘fiNge’, 4% for ‘wOO’, and 87% for ‘nyiE’. After performing the above operations on all pixels of the input image, a set of probability values {pHuman(i,j)}(i,j)=(1,1)(H,W)\{p_{\operatorname{Human}}(i,j)\}_{(i,j)=(1,1)}^{(H,W)} are generated. We arrange each of them according to their spatial position coordinates (i,j)(i,j) to obtain a new matrix regarded as human language probability map mHuman(𝒙)H×W×Cm_{\operatorname{Human}}(\boldsymbol{x})\in\mathbb{R}^{H\times W\times C}. Then we use a cross-entropy loss CE\mathcal{L}_{CE} instead of P\mathcal{L}_{P}. Finally, for this full embedding solution, we minimise the loss function as:

FullEmbedding=M+CE(mτ(𝒙),mHuman(𝒙)).\displaystyle\mathcal{L_{\operatorname{Full-Embedding}}}=\mathcal{L}_{M}+\mathcal{L}_{CE}(m_{\tau}(\boldsymbol{x}),m_{\operatorname{Human}}(\boldsymbol{x})). (14)

By minimising the above Eq. 14, our CQFormer inherits the full colour naming system of human language.

(2) Central Colour Embedding. Alternatively, we could distil less colour naming information. Here, we only embed representative colours and their CC central colour (hue, value) coordinates {μHuman,c}c=1C\{\mu_{\operatorname{Human},c}\}_{c=1}^{C} in the WCS colour probability map of the certain human language (see three asteroids in Fig. 3 (c)). For this embedding solution, we minimise the loss function as:

CentralEmbedding=M+RColour,\displaystyle\mathcal{L_{\operatorname{Central-Embedding}}}=\mathcal{L}_{M}+R_{\operatorname{Colour}}, (15)

where we replace the μc\mu_{c} in RColourR_{\operatorname{Colour}} with aforementioned μHuman,c\mu_{\operatorname{Human},c} and ignore the saturation of Munsell HSV by setting s1=s2=1s_{1}=s_{2}=1.

Table 1: Object detection results on MS COCO dataset [34] with Sparse-RCNN [43] detector, here we report the average precision (AP) value.
Method 1-bit 2-bit 3-bit 4-bit 5-bit 6-bit Full Colour (24-bit)
Upper bound - - - - - - 45.0
Median Cut (w/o D) [23] 11.5 12.7 15.4 17.0 20.4 23.2 -
Median Cut (w D) [17] 12.3 13.8 15.2 19.6 21.8 25.6 -
OCTree [19] 10.7 13.4 13.2 16.7 18.9 22.8 -
CQFormer 13.9 16.5 18.8 21.5 27.5 29.8 -

Colour Evolution Stage. After the distillation of a certain colour naming system of human language, we evolve the CQFormer to split fine-grained new colour. In this stage, we remove the restriction on the number of quantised colours and minimise different combinations of losses to encourage the CQFormer to evolve more colours. Please refer to Sec. 4.4 for more details.

4 Experiments

We evaluate the CQFormer on mainstream benchmark datasets of both object detection task (Sec. 4.2) and image classification task (Sec. 4.3). Additionally, we specifically design a colour evolution experiment (Sec. 4.4) to demonstrate how our CQFormer automatically evolves to increase fine-grained colours. For ablation study, visualisation and other details, please refer to Supplementary Material.

4.1 Datasets and Experiment Settings

Datasets. For object detection, we utilise MS COCO [34] dataset, which contains \sim118k images with bounding box annotation in 80 categories. Here, we use COCO train2017 set as the train set and use COCO val2017 set as the test set. For classification, we use CIFAR10, CIFAR100 [28], STL10 [10] and tiny-imagenet-200 (Tiny200) [29] dataset. Both CIFAR10 and CIFAR100 contain 60000 images, covering 10 and 100 classes of general objects, respectively. The STL10 consists of 13000 images (5000 for training and 8000 for testing) with 10 classes. The Tiny200 is a subset of ImageNet [13] and contains 200 categories with 110k images. We adopt random crop and random flip for data augmentation.

Evaluation Metrics. For object detection, we report average precision (AP) value. For classification, we report top-1 classification accuracy.

Implement Details.

(1) Upper bound. We utilise the performance of the detector/classifier without additional colour quantisation methods in full-colour space (24-bit) as the upper bound. For object detection, we adopt the Sparse-RCNN [43] detector with ResNet-50 [21] backbone and FPN [33] neck. For classification, we adopt Resnet-18 [21] network.

(2) Training Settings. All colour quantisation experiments are finished at quantisation levels from 1-bit to 6-bit, i.e. C{2,4,8,16,32,64}C\in\{2,4,8,16,32,64\}. We set the temperature parameter τ=0.01\tau=0.01 and the α=1\alpha=1, β=0.3\beta=0.3 , γ=1\gamma=1. For detection, only the Resnet-50 backbone of the detector is initialised with Imagenet [13] pre-train weights. We jointly train our CQFormer with Sparse-RCNN for 36 epochs on 4 Tesla V100 GPUs. The batch size is set to 8, and the optimiser is AdamW [36]. The initial learning rate is set to 1.25e61.25e^{-6}, and the learning rate would decay to one-tenth at 24 epochs. For classification, we cascade CQFormer with the Resnet-18 and jointly train them without pre-trained models on a single GeForce RTX A6000 GPU. We employ an SGD optimiser for 60 epochs using the Cosine-Warm-Restart [35] as the learning rate scheduler. A batch size of 128 (STL10 is set to 32), an initial learning rate of 0.05, a momentum of 0.5, and a weight decay of 0.001 are used.

Comparison Methods. We compare with three traditional perception-centred methods: MedianCut [23] , MedianCut+Dither [17] and OCTree [19], and a task-centred CNN-based method ColorCNN [25]. Specifically, for ColorCNN, we adopt the same training strategy as we did for CQFormer. For the traditional methods, we conduct comparative experiments as described in [25].

4.2 Object Detection Task Evaluation

Table. 1 shows the object detection results on MS COCO dataset [34] using Sparse-RCNN [43] detector. Our CQFormer outperforms all other methods in terms of AP value performance under all colour quantisation levels, from 1-bit to 6-bit. This substantial improvement demonstrates the effectiveness of our CQFormer in the object detection task. We also evaluate CQFormer with other popular detectors in Supplementary Material.

4.3 Classification Task Evaluation

Refer to caption
Figure 4: Top-1 classification accuracy of colour-quantised images on four datasets with Resnet-18 [21] networks.
Refer to caption
Figure 5: (a) is the embedding stage result of CQFormer. (b) (c) (d) are the colour evolution stage results of CQFormer with different combinations. (e) is the embedding stage result of ColorCNN [25]. (f) is the colour evolution stage result of ColorCNN [25].

Results. Fig. 4 presents comparisons to the state-of-the-art methods on the four datasets. Our proposed CQFormer (solid blue line) has a consistent and obvious improvement over all other methods in extremely low-bit colour space (less than 3-bit). Moreover, our CQFormer archives a superior performance than the task-centred method ColorCNN [25] under all colour quantisation levels from 1-bit to 6-bit.

Discussion of classification under a large colour space. Similar to the task-centred ColorCNN, our classification performance is also inferior to the traditional method under a large colour space (greater than 4-bit), despite superior performance on object detection task. As extensively discussed in [25], this is a common characteristic for non-clustering based quantisation.

Here we are satisfied with our classification performance under the small colour space for two reasons: First of all, most human languages only use fewer than 3-bit colour terms. This implies that discovering more colours not only compromises the principle of efficiency but also goes contrary to the expectation of better perceptual effects. In other words, the performance of the CQFormer on limited colour categories may hint at the optimal outcome restricted by the unique neurological structure of human vision and cognition, which are reflected in a wide array of languages in turn.

Secondly, larger colour space naturally preserves more visual fidelity. 6-bit colour space could already deliver a vivid image, and the recognition performance on these images is only marginally below the original image. It could neither save much storage space nor reveal special knowledge. Therefore, we only focus on the small colour space.

4.4 Colour Evolution Evaluation

Settings. This experiment is based on the classification task on CIFAR10 [28] dataset using the Resnet-18 [21] classifier. In the embedding stage, we embed the 1978 Nafaanra three-colour system (Fig. 3 (c)) using the colour probability map embedding and set CC as 3. The reason why we choose Nafaanra is analysed in the Supplementary Material. Here we set τ=1.0\tau=1.0, and force the fourth colour to be split from light (‘fiNge’), dark (‘wOO’), and warm or red-like (‘nyiE’), respectively. We optimise the loss function FullEmbedding\mathcal{L_{\operatorname{Full-Embedding}}} in Eq. 14 for the initial 40 epochs. In the colour evolution stage, we design various combinations of loss function, i.e. only M\mathcal{L}_{M}, M+RDiversity\mathcal{L}_{M}+R_{\operatorname{Diversity}} and M+RColour+RDiversity+Perceptual\mathcal{L}_{M}+R_{\operatorname{Colour}}+R_{\operatorname{Diversity}}+\mathcal{L}_{\operatorname{Perceptual}}, and minimise them separately for the subsequent 20 epochs without the restriction of colour size. Additionally, we also conduct a colour evolution experiment on the task-centred ColorCNN [25] for comparison, which is slightly modified to implement the same embedding and evolution details.

Results and Analysis. The results of colour evolution evaluation is shown in Fig. 5. First, during the embedding stage, the WCS colour probability map generated by the CQFormer (Fig. 5 (a)) is similar to Nafaanra (Fig.3 (c)). Therefore, we successfully embed the colour naming system of human language into the latent space of CQFormer. Second, during the colour evolution stage, as shown in Fig. 5 (b)(c)(d), the CQFormer automatically evolves the fourth colour that is split from dark (‘wOO’) and close to yellow-green under all combinations, matching the basic colour terms theory [27]. Third, although the ColorCNN also evolves the fourth colour close to brown (Fig. 5 (f)), it fails to match the basic colour terms theory [27] since the yellow-green and blue are skipped.

We are not able to see the fourth colour split from either light (‘fiNge’) or warm/red-like (‘nyiE’) in the WCS colour probability map, since only 3.7% (if split from light (‘fiNge’)) or 5.9% (if split from warm/red-like (‘nyiE’)) of all pixels are assigned to the fourth colour, compared with 23.5% (split from dark (‘wOO’)). Very interestingly, this phenomenon echoes the evolution of the information bottleneck (IB) colour naming systems [50], where the fourth colour should be spilt from dark in the ”dark-light-red” colour scheme.

Although the CQFormer can also evolve the fourth colour automatically when optimising M\mathcal{L}_{M} alone, it only covers a small portion of the WCS probability map in Fig. 5 (b). As the items of the perceptual structure loss increase, i.e. from Fig. 5 (b) to (d), these four colours have clearer borders, a more logical proportion and are each more internally clustered. This suggests that the colour evolution is not fully complete when considering machine accuracy alone. In other words, thanks to the CQFormer’s ability to integrate the need for both machine accuracy and human perception, the discovered colour naming system evolves more thoroughly and effectively, which is mirrored in colour naming patterns of human language [49].

5 Limitation and Discussion

While the complexity-accuracy trade-off of machine-discovered colour terms, as shown in Fig. 1 (b), is quite similar to the numerical limit of categorical counterparts for human languages, the current work is still preliminary. As shown in Fig. 1, the newly discovered WCS colour probability map is still quite different from the human one. A more accurate language evolution replication needs to consider more complex variables such as environmental idiosyncrasies, cultural peculiarities, functional necessities, technological maturity, learning experience, and intercultural communication.

Another promising direction would be associating the discovered colours with human colour terms. This would involve much research on Nature Language Processing, and we hope to discuss it with experts from different disciplines in future works. Last but not least, the AI simulation outcome contributes to the long-standing universalist-relativist debate of the linguistic community on colour cognition. Though not entirely excluding the cultural specificity of the colour schemes, the machine finding strongly supports the universalist view that an innate, physiological principle constraints, if not determine, the evolutionary sequence and distributional possibilities of basic colour terms in communities of different cultural traditions. The complexity-efficiency principle is confirmed by the finding that the numerical limitation of colour categories could lead to superior performance on colour-specific tasks, contrary to the intuitive expectation that complexity breeds perfection. The independent AI discovery of the ”green-yellow” category on the basis of the fundamental tripartite ”dark-light-red” colour scheme points to the congruence of neural algorithms and human cognition and opens a new frontier to test contested hypothesis in the social sciences through machine simulation. We would be more than delighted if this tentative attempt would prove to be a bridge to link scholars of different disciplines for more collaboration and generate more fruitful results.

Acknowledgement

This work was supported by JST Moonshot R&D Grant Number JPMJMS2011 and JST ACT-X Grant Number JPMJAX190D, Japan.

References

  • [1] Hyojin Bahng, Seungjoo Yoo, Wonwoong Cho, David Keetae Park, Ziming Wu, Xiaojuan Ma, and Jaegul Choo. Coloring with words: Guiding image colorization through text-based palette generation. In Proceedings of the european conference on computer vision (eccv), pages 431–447, 2018.
  • [2] B. Berlin and P. Kay. Basic Color Terms: Their Universality and Evolution. University of California Press, 1969.
  • [3] Thomas Boutell. Png (portable network graphics) specification version 1.0. Technical report, 1997.
  • [4] Federico Camposeco, Andrea Cohen, Marc Pollefeys, and Torsten Sattler. Hybrid scene compression for visual localization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7653–7662, 2019.
  • [5] Nicolas Carion, Francisco Massa, Gabriel Synnaeve, Nicolas Usunier, Alexander Kirillov, and Sergey Zagoruyko. End-to-end object detection with transformers. In European conference on computer vision, pages 213–229. Springer, 2020.
  • [6] Rahma Chaabouni, Eugene Kharitonov, Emmanuel Dupoux, and Marco Baroni. Communicating artificial neural networks develop efficient color-naming systems. Proceedings of the National Academy of Sciences, 118(12):e2016569118, 2021.
  • [7] Huiwen Chang, Ohad Fried, Yiming Liu, Stephen DiVerdi, and Adam Finkelstein. Palette-based photo recoloring. ACM Trans. Graph., 34(4):139–1, 2015.
  • [8] Junho Cho, Sangdoo Yun, Kyoung Mu Lee, and Jin Young Choi. Palettenet: Image recolorization with given color palette. In Proceedings of the ieee conference on computer vision and pattern recognition workshops, pages 62–70, 2017.
  • [9] François Chollet. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1251–1258, 2017.
  • [10] Adam Coates, Andrew Ng, and Honglak Lee. An analysis of single-layer networks in unsupervised feature learning. In Proceedings of the fourteenth international conference on artificial intelligence and statistics, pages 215–223. JMLR Workshop and Conference Proceedings, 2011.
  • [11] Ziteng Cui, Kunchang Li, Lin Gu, Shenghan Su, Peng Gao, ZhengKai Jiang, Yu Qiao, and Tatsuya Harada. You only need 90k parameters to adapt light: a light weight transformer for image enhancement and exposure correction. In 33rd British Machine Vision Conference 2022, BMVC 2022, London, UK, November 21-24, 2022. BMVA Press, 2022.
  • [12] Jelmer P de Vries, Arash Akbarinia, Alban Flachot, and Karl R Gegenfurtner. Emergent color categorization in a neural network trained for object recognition. Elife, 11:e76472, 2022.
  • [13] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. ImageNet: A Large-Scale Hierarchical Image Database. In CVPR09, 2009.
  • [14] Yining Deng, Charles Kenney, Michael S Moore, and BS Manjunath. Peer group filtering and perceptual color image quantization. In 1999 IEEE International Symposium on Circuits and Systems (ISCAS), volume 4, pages 21–24. IEEE, 1999.
  • [15] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. ICLR, 2021.
  • [16] R. W. Floyd and L. Steinberg. An adaptive algorithm for spatial gray scale. Proceedings of the Society for Information Display, 17, 1975.
  • [17] R. W. Floyd and L. Steinberg. An adaptive algorithm for spatial grayscale. Proceedings of the Society for Information Display, 17, 1976.
  • [18] David A Forsyth and Jean Ponce. Computer vision: a modern approach. prentice hall professional technical reference, 2002.
  • [19] Michael Gervautz and Werner Purgathofer. A simple method for color quantization: Octree quantization. In New trends in computer graphics, pages 219–231. Springer, 1988.
  • [20] Edward Gibson, Richard Futrell, Julian Jara-Ettinger, Kyle Mahowald, Leon Bergen, Sivalogeswaran Ratnasingam, Mitchell Gibson, Steven T Piantadosi, and Bevil R Conway. Color naming across languages reflects color use. Proceedings of the National Academy of Sciences, 114(40):10785–10790, 2017.
  • [21] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  • [22] Yu-Lin He, Xiao-Liang Zhang, Wei Ao, and Joshua Zhexue Huang. Determining the optimal temperature parameter for softmax function in reinforcement learning. Applied Soft Computing, 70:80–85, 2018.
  • [23] Paul Heckbert. Color image quantization for frame buffer display. ACM Siggraph Computer Graphics, 16(3):297–307, 1982.
  • [24] Dan Hendrycks and Kevin Gimpel. Gaussian error linear units (gelus). arXiv preprint arXiv:1606.08415, 2016.
  • [25] Yunzhong Hou, Liang Zheng, and Stephen Gould. Learning to structure an image with few colors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10116–10125, 2020.
  • [26] Paul Kay, Brent Berlin, Luisa Maffi, William R Merrifield, and Richard Cook. The world color survey. CSLI Publications Stanford, CA, 2009.
  • [27] Paul Kay and Chad K McDaniel. The linguistic significance of the meanings of basic color terms. Language, 54(3):610–646, 1978.
  • [28] Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009.
  • [29] Ya Le and Xuan Yang. Tiny imagenet visual recognition challenge. CS 231N, 7(7):3, 2015.
  • [30] Boyi Li, Serge Belongie, Ser-nam Lim, and Abe Davis. Neural image recolorization for creative domains. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2226–2230, 2022.
  • [31] Ke Li, Shijie Wang, Xiang Zhang, Yifan Xu, Weijian Xu, and Zhuowen Tu. Pose recognition with cascade transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1944–1953, 2021.
  • [32] DianLong Liang, Lizhi Cheng, and Zenghui Zhang. General construction of wavelet filters via a lifting scheme and its application in image coding. Optical Engineering, 42(7):1949–1955, 2003.
  • [33] Tsung-Yi Lin, Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. Feature pyramid networks for object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2117–2125, 2017.
  • [34] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C. Lawrence Zitnick. Microsoft coco: Common objects in context. In David Fleet, Tomas Pajdla, Bernt Schiele, and Tinne Tuytelaars, editors, Computer Vision – ECCV 2014, pages 740–755, Cham, 2014. Springer International Publishing.
  • [35] Ilya Loshchilov and Frank Hutter. Sgdr: Stochastic gradient descent with warm restarts. arXiv preprint arXiv:1608.03983, 2016.
  • [36] Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. In International Conference on Learning Representations, 2019.
  • [37] M.T. Orchard and C.A. Bouman. Color quantization of images. IEEE Transactions on Signal Processing, 39(12):2677–2690, 1991.
  • [38] C Alejandro Parraga and Arash Akbarinia. Nice: A computational solution to close the gap from colour perception to colour categorization. PloS one, 11(3):e0149538, 2016.
  • [39] Charles Poynton. Digital video and HD: Algorithms and Interfaces. Elsevier, 2012.
  • [40] Mohammad Rastegari, Vicente Ordonez, Joseph Redmon, and Ali Farhadi. Xnor-net: Imagenet classification using binary convolutional neural networks. In Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, editors, Computer Vision – ECCV 2016, pages 525–542, Cham, 2016. Springer International Publishing.
  • [41] Claude Elwood Shannon. A mathematical theory of communication. The Bell system technical journal, 27(3):379–423, 1948.
  • [42] Katarzyna Siuda-Krzywicka, Christoph Witzel, Emma Chabani, Myriam Taga, Cecile Coste, Noella Cools, Sophie Ferrieux, Laurent Cohen, Tal Seidel Malkinson, and Paolo Bartolomeo. Color categorization independent of color naming. Cell reports, 28(10):2471–2479, 2019.
  • [43] Peize Sun, Rufeng Zhang, Yi Jiang, Tao Kong, Chenfeng Xu, Wei Zhan, Masayoshi Tomizuka, Lei Li, Zehuan Yuan, Changhu Wang, and Ping Luo. Sparse r-cnn: End-to-end object detection with learnable proposals. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 14454–14463, 2021.
  • [44] Jeya Maria Jose Valanarasu and Vishal M Patel. Unext: Mlp-based rapid medical image segmentation network. arXiv preprint arXiv:2203.04967, 2022.
  • [45] Joost Van De Weijer, Cordelia Schmid, and Jakob Verbeek. Learning color names from real-world images. In 2007 IEEE conference on computer vision and pattern recognition, pages 1–8. IEEE, 2007.
  • [46] Xiaolin Wu. Color quantization by dynamic programming and principal analysis. ACM Transactions on Graphics (TOG), 11(4):348–372, 1992.
  • [47] Jiwei Yang, Xu Shen, Jun Xing, Xinmei Tian, Houqiang Li, Bing Deng, Jianqiang Huang, and Xian-sheng Hua. Quantization networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7308–7316, 2019.
  • [48] Seungjoo Yoo, Hyojin Bahng, Sunghyo Chung, Junsoo Lee, Jaehyuk Chang, and Jaegul Choo. Coloring with limited data: Few-shot colorization via memory augmented networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11283–11292, 2019.
  • [49] Noga Zaslavsky, Karee Garvin, Charles Kemp, Naftali Tishby, and Terry Regier. The evolution of color naming reflects pressure for efficiency: Evidence from the recent past. Journal of Language Evolution, 2022.
  • [50] Noga Zaslavsky, Charles Kemp, Terry Regier, and Naftali Tishby. Efficient compression in color naming and its evolution. Proceedings of the National Academy of Sciences, 115(31):7937–7942, 2018.
  • [51] Noga Zaslavsky, Charles Kemp, Naftali Tishby, and Terry Regier. Color naming reflects both perceptual structure and communicative need. Topics in Cognitive Science, 11(1):207–219, 2019.
  • [52] Noga Zaslavsky, Charles Kemp, Naftali Tishby, and Terry Regier. Communicative need in color naming. Cognitive Neuropsychology, 2019.