Buckets:

|
download
raw
101 kB

Title: Learning to Compose: Improving Object Centric Learning by Injecting Compositionality

URL Source: https://arxiv.org/html/2405.00646

Published Time: Sat, 25 May 2024 02:23:28 GMT

Markdown Content: Whie Jung, Jaehoon Yoo, Sungjin Ahn, Seunghoon Hong

School of Computing, KAIST

{whieya, wogns98, sungjin.ahn, seunghoon.hong}@kaist.ac.kr

Abstract

Learning compositional representation is a key aspect of object-centric learning as it enables flexible systematic generalization and supports complex visual reasoning. However, most of the existing approaches rely on auto-encoding objective, while the compositionality is implicitly imposed by the architectural or algorithmic bias in the encoder. This misalignment between auto-encoding objective and learning compositionality often results in failure of capturing meaningful object representations. In this study, we propose a novel objective that explicitly encourages compositionality of the representations. Built upon the existing object-centric learning framework (e.g., slot attention), our method incorporates additional constraints that an arbitrary mixture of object representations from two images should be valid by maximizing the likelihood of the composite data. We demonstrate that incorporating our objective to the existing framework consistently improves the objective-centric learning and enhances the robustness to the architectural choices. Codes are available at https://github.com/whieya/Learning-to-compose.

1 Introduction

As the world is highly compositional in nature, relatively few composable units, such as objects or words, can describe infinitely many observations. Consequently, human intelligence has evolved to recognize the environment as a combination of composable units, (e.g., objects) which enables rapid adaptation to unseen situations by recomposing the already learned concepts(Spelke, 1990; Lake et al., 2017). Mimicking human intelligence, perceiving environment with composable abstractions have shown consistent improvement in tasks related to systematic generalization(Kuo et al., 2021; Bogin et al., 2021; Rahaman et al., 2021), and visual reasoning tasks(D’Amario et al., 2021; Assouel et al., 2022) compared to distributed counterparts.

Inheriting this spirit, object-centric learning(Burgess et al., 2019; Greff et al., 2019; Engelcke et al., 2020; Locatello et al., 2020) aims to discover a composable abstraction purely from data without external supervision. Instead of depicting a scene with a distributed representation, it decomposes the scene into a set of latent representations, where each latent is expected to capture a distinct object. To discover such representation in an unsupervised manner, most existing works employed an auto-encoding framework, where the model is trained to encode the scene into a set of representations and decode them back to the original image.

However, the auto-encoding objective is inherently insufficient to learn compositional representation, since maximizing the reconstruction quality does not necessarily requires the object-level disentanglement. To reduce this gap, the existing works incorporate strong inductive biases to further regularize the encoder, such as architectural bias(Locatello et al., 2020) or algorithmic bias(Burgess et al., 2019; Lin et al., 2020; Jiang et al., 2020). However, it has been widely observed that these methods are highly sensitive to the choice of hyper-parameters, such as encoder and decoder architectures, and a number of slots, often resulting in suboptimal decompositions by position or partial attributes(Singh et al., 2022a; Sajjadi et al., 2022; Jiang et al., 2023) instead of objects. Finding the optimal model configuration is also not straightforward in practice due to the missing object labels.

In this work, we present a novel objective that directly optimizes the compositionality of representations. Based upon the auto-encoding framework, our method extracts object representations independently from two distinct images and simulates their composition by the random mixture. The composite representation is rendered to an image by the decoder, whose likelihood is evaluated by the generative prior. The encoder is then jointly optimized to minimize the reconstruction error of the individual images to encode relevant information of the scene (auto-encoding path) while maximizing the likelihood of the composite image to ensure the compositionality of the representation (composition path). Overall, our method can be viewed as extending the conventional auto-encoding approach with an additional regularization on compositionality. We show that directly injecting compositionality this way significantly boosts the overall quality of object-centric representations and robustness in training.

Our contributions are as follows. (1) We introduce a novel objective that explicitly encourages compositionality of representations. To this end, we investigate strategies to simulate the compositional construction of an image and propose a learning objective for maximizing the likelihood of the composite images. (2) We evaluate our framework on four datasets and verify that our model consistently surpasses auto-encoding based baselines by a substantial margin. (3) We show that our objective enhances the robustness of object-centric learning on three major factors, such as number of latents, encoder and decoder architectures.

2 Preliminary

Problem setup

Object-centric learning aims to discover a set of composable representations from an unlabeled image. Formally, given an image π±βˆˆβ„ HΓ—WΓ—C 𝐱 superscript ℝ 𝐻 π‘Š 𝐢\mathbf{x}\in\mathbb{R}^{H\times W\times C}bold_x ∈ blackboard_R start_POSTSUPERSCRIPT italic_H Γ— italic_W Γ— italic_C end_POSTSUPERSCRIPT represented by either RGB pixels or feature from the pre-trained encoder, the objective is to extract the set 𝐒={𝐬 1,…,𝐬 N}=E θ⁒(𝐱)𝐒 subscript 𝐬 1…subscript 𝐬 𝑁 subscript 𝐸 πœƒ 𝐱\mathbf{S}={\mathbf{s}{1},\dots,\mathbf{s}{N}}=E_{\theta}(\mathbf{x})bold_S = { bold_s start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , … , bold_s start_POSTSUBSCRIPT italic_N end_POSTSUBSCRIPT } = italic_E start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT ( bold_x ), where each element 𝐬 iβˆˆβ„ D subscript 𝐬 𝑖 superscript ℝ 𝐷\mathbf{s}{i}\in\mathbb{R}^{D}bold_s start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ∈ blackboard_R start_POSTSUPERSCRIPT italic_D end_POSTSUPERSCRIPT corresponds to the representation of a composable concept (e.g., an object). Since object concepts should emerge from the data without supervision, a typical approach is to use an auto-encoding framework to formulate the learning process. Formally, the object-centric encoder E ΞΈ:ℝ HΓ—WΓ—C→ℝ NΓ—D:subscript 𝐸 πœƒβ†’superscript ℝ 𝐻 π‘Š 𝐢 superscript ℝ 𝑁 𝐷 E{\theta}:\mathbb{R}^{H\times W\times C}\rightarrow\mathbb{R}^{N\times D}italic_E start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT : blackboard_R start_POSTSUPERSCRIPT italic_H Γ— italic_W Γ— italic_C end_POSTSUPERSCRIPT β†’ blackboard_R start_POSTSUPERSCRIPT italic_N Γ— italic_D end_POSTSUPERSCRIPT is trained jointly with a decoder D Ο•:ℝ NΓ—D→ℝ HΓ—WΓ—C:subscript 𝐷 italic-Ο• absentβ†’superscript ℝ 𝑁 𝐷 superscript ℝ 𝐻 π‘Š 𝐢 D_{\phi}:\mathbb{R}^{N\times D}\xrightarrow{}\mathbb{R}^{H\times W\times C}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT : blackboard_R start_POSTSUPERSCRIPT italic_N Γ— italic_D end_POSTSUPERSCRIPT start_ARROW start_OVERACCENT end_OVERACCENT β†’ end_ARROW blackboard_R start_POSTSUPERSCRIPT italic_H Γ— italic_W Γ— italic_C end_POSTSUPERSCRIPT by minimizing the reconstruction loss.

β„’ AE(ΞΈ,Ο•)=𝔼 𝐱[d(𝐱,D Ο•(E ΞΈ(𝐱))],\mathcal{L}{\text{AE}}(\theta,\phi)=\mathbb{E}{\mathbf{x}}\left[d(\mathbf{x}% ,D_{\phi}(E_{\theta}(\mathbf{x}))\right],caligraphic_L start_POSTSUBSCRIPT AE end_POSTSUBSCRIPT ( italic_ΞΈ , italic_Ο• ) = blackboard_E start_POSTSUBSCRIPT bold_x end_POSTSUBSCRIPT [ italic_d ( bold_x , italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT ( italic_E start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT ( bold_x ) ) ] ,(1)

where d 𝑑 d italic_d is a distance metric (e.g., MSE).

Slot Attention Encoder E ΞΈ subscript 𝐸 πœƒ E_{\theta}italic_E start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT

Since the auto-encoding objective is insufficient to learn highly structured representation, the existing approaches incorporate a strong architectural bias in the encoder E ΞΈ subscript 𝐸 πœƒ E_{\theta}italic_E start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT to guide the object-level disentanglement in 𝐒 𝐒\mathbf{S}bold_S. Among many variants, we consider Slot Attention encoder Locatello et al. (2020) due to its popularity and generality. It employs a dot-product attention mechanism between a query (slot) and a key (input), where normalization is applied over the slots by:

𝐀⁒(𝐱,𝐒)=softmax 𝑁⁒(k⁒(𝐳)β‹…q⁒(𝐒)T D)βˆˆβ„ MΓ—N,𝐀 𝐱 𝐒 𝑁 softmaxβ‹…π‘˜ 𝐳 π‘ž superscript 𝐒 𝑇 𝐷 superscript ℝ 𝑀 𝑁\displaystyle\mathbf{A}(\mathbf{x},\mathbf{S})=\underset{N}{\text{softmax}}% \left(\frac{k(\mathbf{z})\cdot q(\mathbf{S})^{T}}{\sqrt{D}}\right)\in\mathbb{R% }^{M\times N},bold_A ( bold_x , bold_S ) = underitalic_N start_ARG softmax end_ARG ( divide start_ARG italic_k ( bold_z ) β‹… italic_q ( bold_S ) start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT end_ARG start_ARG square-root start_ARG italic_D end_ARG end_ARG ) ∈ blackboard_R start_POSTSUPERSCRIPT italic_M Γ— italic_N end_POSTSUPERSCRIPT ,(2)

where 𝐳=f θ⁒(𝐱)βˆˆβ„ MΓ—D′𝐳 subscript 𝑓 πœƒ 𝐱 superscript ℝ 𝑀 superscript 𝐷′\mathbf{z}=f_{\theta}(\mathbf{x})\in\mathbb{R}^{M\times D^{\prime}}bold_z = italic_f start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT ( bold_x ) ∈ blackboard_R start_POSTSUPERSCRIPT italic_M Γ— italic_D start_POSTSUPERSCRIPT β€² end_POSTSUPERSCRIPT end_POSTSUPERSCRIPT is a flattened input feature encoded by CNN encoder f ΞΈ subscript 𝑓 πœƒ f_{\theta}italic_f start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT, and k,q π‘˜ π‘ž k,q italic_k , italic_q represents linear projection matrices. Note that softmax operation is normalized in the query (slots) direction, inducing competition among slots. Based on Equation2, the slots are iteratively refined by:

𝐒(n+1)=GRU⁒(𝐒(n),Normalize⁒(𝐀⁒(𝐱,𝐒(n))Tβ‹…v⁒(𝐳))),⁒𝐒(0)βˆΌπ’©β’(ΞΌ,diag⁒(Οƒ)).formulae-sequence superscript 𝐒 𝑛 1 GRU superscript 𝐒 𝑛 Normalize⋅𝐀 superscript 𝐱 superscript 𝐒 𝑛 𝑇 𝑣 𝐳 similar-to superscript 𝐒 0 𝒩 πœ‡ diag 𝜎\displaystyle\mathbf{S}^{(n+1)}=\text{GRU}(\mathbf{S}^{(n)},\text{Normalize}(% \mathbf{A}(\mathbf{x},\mathbf{S}^{(n)})^{T}\cdot v(\mathbf{z}))),{}{}\text{ % }\mathbf{S}^{(0)}\sim\mathcal{N}(\mu,\text{diag}({\sigma})).bold_S start_POSTSUPERSCRIPT ( italic_n + 1 ) end_POSTSUPERSCRIPT = GRU ( bold_S start_POSTSUPERSCRIPT ( italic_n ) end_POSTSUPERSCRIPT , Normalize ( bold_A ( bold_x , bold_S start_POSTSUPERSCRIPT ( italic_n ) end_POSTSUPERSCRIPT ) start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT β‹… italic_v ( bold_z ) ) ) , bold_S start_POSTSUPERSCRIPT ( 0 ) end_POSTSUPERSCRIPT ∼ caligraphic_N ( italic_ΞΌ , diag ( italic_Οƒ ) ) .(3)

Here, 𝐒(n)superscript 𝐒 𝑛\mathbf{S}^{(n)}bold_S start_POSTSUPERSCRIPT ( italic_n ) end_POSTSUPERSCRIPT denotes the slot representation after n 𝑛 n italic_n iterations, ΞΌ,Οƒ πœ‡ 𝜎\mu,\sigma italic_ΞΌ , italic_Οƒ are learnable parameters characterizing the distribution of the initial slots, v 𝑣 v italic_v is a linear projection matrix, and Normalize⁒(β‹…)Normalizeβ‹…\text{Normalize}(\cdot)Normalize ( β‹… ) is a weighted mean operation introduced by Locatello et al. (2020) to improve stability of the attention.

Slot Decoder D Ο• subscript 𝐷 italic-Ο• D_{\phi}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT

While the architectural choice for D Ο• subscript 𝐷 italic-Ο• D_{\phi}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT is not constrained to a specific form in principle, subsequent works(Singh et al., 2022a; Jiang et al., 2023) have empirically found that the choice of the decoder crucially impacts the quality of the object-centric representation. Locatello et al. (2020) proposed a pixel-mixture decoder that renders each slot independently into pixels and combines them with alpha-blending. Although slot-wise decoding provides a strong incentive for the encoder to capture distinct objects in each slot, its limited expressiveness hinders its application to complex scenes. To address this issue, Singh et al. (2022a) employed Transformer decoder that takes the entire slots 𝐒 𝐒\mathbf{S}bold_S as an input and produces an image in an autoregressive manner. By modeling the complex interactions among the slots, it has shown great improvements in slot representation learning even in complex scenes.

Recently, Jiang et al. (2023) employed a diffusion model for the slot decoder. Instead of directly reconstructing an input image 𝐱 𝐱\mathbf{x}bold_x, it optimizes the auto-encoding of Equation1 via denoising objective(Ho et al., 2020) by:

β„’ Diff⁒(ΞΈ,Ο•)=𝔼 Ο΅βˆΌπ’©β’(𝟎,𝐈),t∼U⁒(0,1)⁒[w⁒(t)β‹…β€–D ϕ⁒(𝐱 t,t,𝐒=E θ⁒(𝐱))βˆ’Ο΅β€–2],subscript β„’ Diff πœƒ italic-Ο• subscript 𝔼 formulae-sequence similar-to italic-Ο΅ 𝒩 0 𝐈 similar-to 𝑑 π‘ˆ 0 1 delimited-[]⋅𝑀 𝑑 superscript delimited-βˆ₯βˆ₯subscript 𝐷 italic-Ο• subscript 𝐱 𝑑 𝑑 𝐒 subscript 𝐸 πœƒ 𝐱 italic-Ο΅ 2\begin{gathered}\mathcal{L}{\text{Diff}}(\theta,\phi)=\mathbb{E}{\epsilon% \sim\mathcal{N}(\mathbf{0},\mathbf{I}),t\sim U(0,1)}\left[w(t)\cdot|D_{\phi}(% \mathbf{x}{t},t,\mathbf{S}=E{\theta}(\mathbf{x}))-\epsilon|^{2}\right],\end% {gathered}start_ROW start_CELL caligraphic_L start_POSTSUBSCRIPT Diff end_POSTSUBSCRIPT ( italic_ΞΈ , italic_Ο• ) = blackboard_E start_POSTSUBSCRIPT italic_Ο΅ ∼ caligraphic_N ( bold_0 , bold_I ) , italic_t ∼ italic_U ( 0 , 1 ) end_POSTSUBSCRIPT [ italic_w ( italic_t ) β‹… βˆ₯ italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT ( bold_x start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_t , bold_S = italic_E start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT ( bold_x ) ) - italic_Ο΅ βˆ₯ start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ] , end_CELL end_ROW(4)

where 𝐱 t=Ξ±Β―t⁒𝐱+1βˆ’Ξ±Β―t subscript 𝐱 𝑑 subscript¯𝛼 𝑑 𝐱 1 subscript¯𝛼 𝑑\mathbf{x}{t}=\sqrt{\bar{\alpha}{t}}\mathbf{x}+\sqrt{1-\bar{\alpha}{t}}bold_x start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT = square-root start_ARG overΒ― start_ARG italic_Ξ± end_ARG start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT end_ARG bold_x + square-root start_ARG 1 - overΒ― start_ARG italic_Ξ± end_ARG start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT end_ARG is an corrupted image of an input 𝐱 𝐱\mathbf{x}bold_x by the forward diffusion process at step t 𝑑 t italic_t, Ξ±Β―t=∏i t(1βˆ’Ξ² i)subscript¯𝛼 𝑑 subscript superscript product 𝑑 𝑖 1 subscript 𝛽 𝑖\bar{\alpha}{t}=\prod^{t}{i}(1-\beta{i})overΒ― start_ARG italic_Ξ± end_ARG start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT = ∏ start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ( 1 - italic_Ξ² start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ) is a schedule function, and w⁒(t)𝑀 𝑑 w(t)italic_w ( italic_t ) is the weighting parameter. In practice, the diffusion decoder is implemented based on UNet architecture(Rombach et al., 2022), where each layer consists of a CNN-layer followed by a slot-conditioned Transformer. Once trained, the decoder generates an image 𝐱∼p ϕ⁒(𝐱|𝐒)similar-to 𝐱 subscript 𝑝 italic-Ο• conditional 𝐱 𝐒\mathbf{x}\sim p_{\phi}(\mathbf{x}|\mathbf{S})bold_x ∼ italic_p start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT ( bold_x | bold_S ) using iterative denoising, starting from the random Gaussian noise(Ho et al., 2020; Rombach et al., 2022). Employing a diffusion decoder significantly enhances object-centric representation and generation quality compared to previous arts especially in complex scenes Jiang et al. (2023).

2.1 Limitations

While the slot attention with auto-encoding objectives has shown promise in object-centric learning, its success highly depends on the model architectures, such as number of slots and architectures of the encoder and decoder, where suboptimal configuration often leads to dividing the scenes into tessellations(Singh et al., 2022a; Sajjadi et al., 2022) and objects into the parts(Jiang et al., 2023). However, the optimal model configuration varies depending on the datasets, and discovering them through cross-validation is practically infeasible due to the missing object labels in an unsupervised setting. We argue that such instability is primarily because the auto-encoding objective is inherently misaligned with the one for object-centric learning, since the former guides the encoder only to minimize the information loss on the input, while the latter demands the object-level disentanglement in the representation, potentially sacrificing the reconstruction quality. This motivates us to seek an alternative approach that directly encourages object-level disentanglement in the objective function instead of designing architectural biases.

3 Learning to Compose

Our goal is to improve object-centric learning by modifying its objective function to be more directly aligned with learning compositional slot representation than the auto-encoding loss. Our main intuition is that arbitrary compositions of object representation are likely to yield another valid representation. To realize this intuition, our framework is designed to generate composite images by mixing slot representations from two images and maximize their validity measured by the data prior.

Figure1 illustrates the overall framework of our method. Our framework is built upon the conventional object-centric learning that learns both the slot encoder and decoder by the auto-encoding path on individual images (Section2). To impose compositionality on slot representation, we incorporate an additional composition path that constructs a composite slot representation from two images by the mixing strategy (Section3.1) and assesses the quality of the image generated from the mixed slots by the generative prior (Section3.2). This way, the auto-encoding path ensures that each slot contains the relevant information of an input image, while such slots are constrained to capture composable components of the scenes (e.g., objects) by regularizing the encoder through the composition path.

Image 1: Refer to caption

Figure 1: Overview of our method. Our framework consists of two paths: an auto-encoding path and a composition path. The auto-encoding path ensures slot representations encode relevant information about an image. In contrast, the composition path encourages the compositionality of the representations by constructing the composite representation through the mixture of slots from two separate images (Section3.1), and assessing the quality of the composite image by the generative prior (Section3.2). The encoder is jointly optimized by both paths.

3.1 Mixing Strategy for composing slot representation

Given 𝐒 1,𝐒 2βˆˆβ„ NΓ—D superscript 𝐒 1 superscript 𝐒 2 superscript ℝ 𝑁 𝐷\mathbf{S}^{1},\mathbf{S}^{2}\in\mathbb{R}^{N\times D}bold_S start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT , bold_S start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ∈ blackboard_R start_POSTSUPERSCRIPT italic_N Γ— italic_D end_POSTSUPERSCRIPT extracted from two distinct images 𝐱 1,𝐱 2 superscript 𝐱 1 superscript 𝐱 2\mathbf{x}^{1},\mathbf{x}^{2}bold_x start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT , bold_x start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT, we construct their composite slot representation 𝐒 cβˆˆβ„ NΓ—D superscript 𝐒 𝑐 superscript ℝ 𝑁 𝐷\mathbf{S}^{c}\in\mathbb{R}^{N\times D}bold_S start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT ∈ blackboard_R start_POSTSUPERSCRIPT italic_N Γ— italic_D end_POSTSUPERSCRIPT by

𝐒 c=π⁒(𝐒 1,𝐒 2),superscript 𝐒 𝑐 πœ‹ superscript 𝐒 1 superscript 𝐒 2\displaystyle\mathbf{S}^{c}=\pi(\mathbf{S}^{1},\mathbf{S}^{2}),bold_S start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT = italic_Ο€ ( bold_S start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT , bold_S start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ) ,(5)

where π⁒(β‹…,β‹…)πœ‹β‹…β‹…\pi(\cdot,\cdot)italic_Ο€ ( β‹… , β‹… ) denotes a composition function of two sets. The primary role of the composition function is to simulate potential combinations of slot-wise compositions. Since our goal is to maximize the compositionality of unseen slot combinations, the composition function should be capable of exploring a broad range of compositional possibilities. Below, we introduce simple instantiations of such function.

Random Sampling

In this approach, we randomly sample N 𝑁 N italic_N slots among 2⁒N 2 𝑁 2N 2 italic_N slots i.e., 𝐒 cβ’βˆΌπ‘β’(𝐒 1βˆͺ𝐒 2)superscript 𝐒 𝑐 𝑁 similar-to superscript 𝐒 1 superscript 𝐒 2\mathbf{S}^{c}\overset{N}{\sim}(\mathbf{S}^{1}\cup\mathbf{S}^{2})bold_S start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT overitalic_N start_ARG ∼ end_ARG ( bold_S start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT βˆͺ bold_S start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ). As it explores over all of the possible combinations, this composition function encourages the slot representation itself to be highly composable to generate valid images for any combinations. On the other hand, it may produce invalid combinations of slots on rare occasions, e.g., omitting the background slots or sampling two objects placed in the same location.

Sharing Slot initialization

One way to mitigate such suspicious compositions is to constrain 𝐒 c superscript 𝐒 𝑐\mathbf{S}^{c}bold_S start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT to be valid composition of the scene. However, strictly ensuring this constraint is non-trivial due to the stochastic nature of slot attention i.e., each slot is sampled stochastically from its underlying distribution and the association between the slots and scenes varies depending on the initialization. Instead, we adopt a rather simple approach that employs the identical slot initialization 𝐒(0)superscript 𝐒 0\mathbf{S}^{(0)}bold_S start_POSTSUPERSCRIPT ( 0 ) end_POSTSUPERSCRIPT in Equation3 for two images, and sample the exclusive set of slots. Formally, let I 1 subscript 𝐼 1 I_{1}italic_I start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT and I 2 subscript 𝐼 2 I_{2}italic_I start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT be a random partition of slot indices i.e., I 1βˆͺI 2={1,…,N},I 1∩I 2=βˆ…formulae-sequence subscript 𝐼 1 subscript 𝐼 2 1…𝑁 subscript 𝐼 1 subscript 𝐼 2 I_{1}\cup I_{2}={1,...,N},I_{1}\cap I_{2}=\emptyset italic_I start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT βˆͺ italic_I start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT = { 1 , … , italic_N } , italic_I start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT ∩ italic_I start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT = βˆ…. Then we construct the composite slot by 𝐒 c=𝐒 I 1 1βˆͺ𝐒 I 2 2 superscript 𝐒 𝑐 subscript superscript 𝐒 1 subscript 𝐼 1 subscript superscript 𝐒 2 subscript 𝐼 2\mathbf{S}^{c}=\mathbf{S}^{1}{I{1}}\cup\mathbf{S}^{2}{I{2}}bold_S start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT = bold_S start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_I start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT end_POSTSUBSCRIPT βˆͺ bold_S start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_I start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT end_POSTSUBSCRIPT, where 𝐒 1 superscript 𝐒 1\mathbf{S}^{1}bold_S start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT and 𝐒 2 superscript 𝐒 2\mathbf{S}^{2}bold_S start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT are slots extracted by Equation3 from 𝐱 1 superscript 𝐱 1\mathbf{x}^{1}bold_x start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT and 𝐱 2 superscript 𝐱 2\mathbf{x}^{2}bold_x start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT, respectively, which are initialized with the same 𝐒(0)superscript 𝐒 0\mathbf{S}^{(0)}bold_S start_POSTSUPERSCRIPT ( 0 ) end_POSTSUPERSCRIPT.The underlying intuition is that the slot initialization is reasonably correlated with the objects it captures (Figure7), hence sampling from exclusive slots is likely to be valid scenes than the random sampling.

3.2 Maximizing likelihood of the composite image

Given the composite slot 𝐒 c superscript 𝐒 𝑐\mathbf{S}^{c}bold_S start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT obtained by the previous section, our next step is quantifying its validity i.e., measuring how valid the composition of two image slots is. To this end, we decode it back to an image by 𝐱 c=D ϕ⁒(𝐒 s)superscript 𝐱 𝑐 subscript 𝐷 italic-Ο• superscript 𝐒 𝑠\mathbf{x}^{c}=D_{\phi}(\mathbf{S}^{s})bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT = italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT ( bold_S start_POSTSUPERSCRIPT italic_s end_POSTSUPERSCRIPT ) and measure the likelihood of the image using the generative prior p⁒(𝐱 c)𝑝 superscript 𝐱 𝑐 p(\mathbf{x}^{c})italic_p ( bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT ).

Generative Prior

To model the generative prior p⁒(𝐱 c)𝑝 superscript 𝐱 𝑐 p(\mathbf{x}^{c})italic_p ( bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT ), we opt for a diffusion model(Ho et al., 2020) due to its excellence in generation quality and mode coverage(Xiao et al., 2022). The latter is especially important in our framework since the model evaluates the prior over potentially out-of-distribution samples generated by the composition (Section3.1). Instead of introducing an additional pre-trained diffusion model, we employ the diffusion-based decoder in the auto-encoding path (Section2), and reuse it as a generative prior. This way, our decoder D Ο• subscript 𝐷 italic-Ο• D_{\phi}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT is trained by minimizing the reconstruction loss by denoising objective in Equation4, while serving as a generative prior in the composition path. It greatly improves the parameter-efficiency and memory, and the need for pre-trained generative prior per dataset.

Maximizing p⁒(𝐱 c)𝑝 superscript 𝐱 𝑐 p(\mathbf{x}^{c})italic_p ( bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT )

Given the generative prior, we maximize the likelihood p⁒(𝐱 c)𝑝 superscript 𝐱 𝑐 p(\mathbf{x}^{c})italic_p ( bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT ) with respect to 𝐱 c superscript 𝐱 𝑐\mathbf{x}^{c}bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT in the composition path. Since β„’ Diff subscript β„’ Diff\mathcal{L}{\text{Diff}}caligraphic_L start_POSTSUBSCRIPT Diff end_POSTSUBSCRIPT in Equation 4 is minimizing the upper bound of negative log likelihood of x c superscript π‘₯ 𝑐 x^{c}italic_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT(Ho et al., 2020), minimizing β„’ Diff subscript β„’ Diff\mathcal{L}{\text{Diff}}caligraphic_L start_POSTSUBSCRIPT Diff end_POSTSUBSCRIPT with respect to 𝐱 c superscript 𝐱 𝑐\mathbf{x}^{c}bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT leads to the maximization of the likelihood p⁒(𝐱 c)𝑝 superscript 𝐱 𝑐 p(\mathbf{x}^{c})italic_p ( bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT ). However, computing the gradient of β„’ Diff subscript β„’ Diff\mathcal{L}{\text{Diff}}caligraphic_L start_POSTSUBSCRIPT Diff end_POSTSUBSCRIPT requires expensive computation of Jacobian maxtrix of the decoder and it often degrades the overall training stability. Following (Poole et al., 2022), the gradient of β„’ Diff subscript β„’ Diff\mathcal{L}{\text{Diff}}caligraphic_L start_POSTSUBSCRIPT Diff end_POSTSUBSCRIPT with respect to ΞΈ πœƒ\theta italic_ΞΈ can be approximated as:

βˆ‡ΞΈ β„’ Prior⁒(ΞΈ)=𝔼 t,ϡ⁒[w⁒(t)⁒(D ϕ⁒(𝐱 t c,t,𝐒 c)βˆ’Ο΅)β’βˆ‚π± cβˆ‚ΞΈ].subscriptβˆ‡πœƒ subscript β„’ Prior πœƒ subscript 𝔼 𝑑 italic-Ο΅ delimited-[]𝑀 𝑑 subscript 𝐷 italic-Ο• subscript superscript 𝐱 𝑐 𝑑 𝑑 superscript 𝐒 𝑐 italic-Ο΅ superscript 𝐱 𝑐 πœƒ\displaystyle\nabla_{\theta}\mathcal{L}{\text{Prior}}(\theta)=\mathbb{E}{t,% \epsilon}[w(t)(D_{\phi}(\mathbf{x}^{c}_{t},t,\mathbf{S}^{c})-\epsilon)\frac{% \partial\mathbf{x}^{c}}{\partial\theta}].βˆ‡ start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT caligraphic_L start_POSTSUBSCRIPT Prior end_POSTSUBSCRIPT ( italic_ΞΈ ) = blackboard_E start_POSTSUBSCRIPT italic_t , italic_Ο΅ end_POSTSUBSCRIPT [ italic_w ( italic_t ) ( italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT ( bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT , italic_t , bold_S start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT ) - italic_Ο΅ ) divide start_ARG βˆ‚ bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT end_ARG start_ARG βˆ‚ italic_ΞΈ end_ARG ] .(6)

where Ο΅βˆΌπ’©β’(𝟎,𝐈)similar-to italic-Ο΅ 𝒩 0 𝐈\epsilon\sim\mathcal{N}(\mathbf{0},\mathbf{I})italic_Ο΅ ∼ caligraphic_N ( bold_0 , bold_I ) is a noise, tβˆΌπ’°β’(t min,t max)similar-to 𝑑 𝒰 subscript 𝑑 min subscript 𝑑 max t\sim\mathcal{U}(t_{\text{min}},t_{\text{max}})italic_t ∼ caligraphic_U ( italic_t start_POSTSUBSCRIPT min end_POSTSUBSCRIPT , italic_t start_POSTSUBSCRIPT max end_POSTSUBSCRIPT ) is a timestep, respectively, w⁒(t)𝑀 𝑑 w(t)italic_w ( italic_t ) is a weighting function dependent to t 𝑑 t italic_t, and 𝐱 t c=Ξ±Β―t⁒𝐱 c+Οƒ t⁒ϡ superscript subscript 𝐱 𝑑 𝑐 subscript¯𝛼 𝑑 superscript 𝐱 𝑐 subscript 𝜎 𝑑 italic-Ο΅\mathbf{x}{t}^{c}=\sqrt{\bar{\alpha}{t}}\mathbf{x}^{c}+\sigma_{t}\epsilon bold_x start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT = square-root start_ARG overΒ― start_ARG italic_Ξ± end_ARG start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT end_ARG bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT + italic_Οƒ start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT italic_Ο΅ is a corrupted image of 𝐱 c superscript 𝐱 𝑐\mathbf{x}^{c}bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT from forward diffusion process. By updating the encoder parameters ΞΈ πœƒ\theta italic_ΞΈ with βˆ‡ΞΈ β„’ Prior subscriptβˆ‡πœƒ subscript β„’ Prior\nabla_{\theta}\mathcal{L}_{\text{Prior}}βˆ‡ start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT caligraphic_L start_POSTSUBSCRIPT Prior end_POSTSUBSCRIPT, 𝐱 c superscript 𝐱 𝑐\mathbf{x}^{c}bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT is guided toward high probability density region following the diffusion prior. Note that optimization of the Equation6 is with only respect to the encoder parameter while fixing the decoder. It prevents suspicious collaboration between the encoder and decoders in generating composite images from suboptimal slots.

Surrogate One-Shot Decoder

As discussed earlier, our framework exploits the diffusion model D Ο• subscript 𝐷 italic-Ο• D_{\phi}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT as a decoder and generative prior in the auto-encoding and composition paths, respectively. One drawback is that the diffusion decoder requires an iterative denoising process to generate the composite image 𝐱 c superscript 𝐱 𝑐\mathbf{x}^{c}bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT, which takes significant time and makes the backpropagation through the decoder non-trivial. To address this problem, we employ a one-shot decoder D ψ subscript 𝐷 πœ“ D_{\psi}italic_D start_POSTSUBSCRIPT italic_ψ end_POSTSUBSCRIPT as a surrogate for D Ο• subscript 𝐷 italic-Ο• D_{\phi}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT to support fast and differentiable decoding of 𝐱 c superscript 𝐱 𝑐\mathbf{x}^{c}bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT. 1 1 1 We also consider one-step denoising result of the diffusion decoder using Tweedie’s formula(Stein, 1981; Robbins, 1992) but observe severe degradation in performance due to its inferior quality.

We employ a bidirectional Transformer(Devlin et al., 2019) that takes the composite slot 𝐒 c superscript 𝐒 𝑐\mathbf{S}^{c}bold_S start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT and the learnable mask tokens π¦βˆˆβ„ H⁒WΓ—C 𝐦 superscript ℝ 𝐻 π‘Š 𝐢\mathbf{m}\in\mathbb{R}^{HW\times C}bold_m ∈ blackboard_R start_POSTSUPERSCRIPT italic_H italic_W Γ— italic_C end_POSTSUPERSCRIPT as input, and produces the composite image by a single forward process by 𝐱 c=D ψ⁒(𝐦,𝐒 c)superscript 𝐱 𝑐 subscript 𝐷 πœ“ 𝐦 superscript 𝐒 𝑐\mathbf{x}^{c}=D_{\psi}(\mathbf{m},\mathbf{S}^{c})bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT = italic_D start_POSTSUBSCRIPT italic_ψ end_POSTSUBSCRIPT ( bold_m , bold_S start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT ). The decoder is trained along with the auto-encoding path by:

β„’ Recon⁒(ΞΈ,ψ)=β€–D ψ⁒(𝐦,E θ⁒(𝐱))βˆ’π±β€–2.subscript β„’ Recon πœƒ πœ“ superscript norm subscript 𝐷 πœ“ 𝐦 subscript 𝐸 πœƒ 𝐱 𝐱 2\displaystyle\mathcal{L}{\text{Recon}}(\theta,\psi)=||D{\psi}(\mathbf{m},E_{% \theta}(\mathbf{x}))-\mathbf{x}||^{2}.caligraphic_L start_POSTSUBSCRIPT Recon end_POSTSUBSCRIPT ( italic_ΞΈ , italic_ψ ) = | | italic_D start_POSTSUBSCRIPT italic_ψ end_POSTSUBSCRIPT ( bold_m , italic_E start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT ( bold_x ) ) - bold_x | | start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT .(7)

Note that the generation quality of the one-shot decoder D ψ subscript 𝐷 πœ“ D_{\psi}italic_D start_POSTSUBSCRIPT italic_ψ end_POSTSUBSCRIPT is behind the powerful diffusion decoder D Ο• subscript 𝐷 italic-Ο• D_{\phi}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT, and serves only to compute the 𝐱 c superscript 𝐱 𝑐\mathbf{x}^{c}bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT in Equation6. We observe that such weak decoder is sufficient to compute the meaningful gradient through the Equation6, presumably because the gradients are accumulated over various noise levels t 𝑑 t italic_t.

3.3 Learning Objective

In this section, we summarize the overall framework and objective function. Our framework consists of two paths; auto-encoding path and composition path. In auto-encoding path, encoder E ΞΈ subscript 𝐸 πœƒ E_{\theta}italic_E start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT and two different decoders D Ο•,D ψ subscript 𝐷 italic-Ο• subscript 𝐷 πœ“ D_{\phi},D_{\psi}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT , italic_D start_POSTSUBSCRIPT italic_ψ end_POSTSUBSCRIPT are trained to minimize auto-encoding objective in Equation4 and Equation7. In composition path, we first extract 𝐒 c superscript 𝐒 𝑐\mathbf{S}^{c}bold_S start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT with Equation5 and generate 𝐱 c superscript 𝐱 𝑐\mathbf{x}^{c}bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT with the deterministic decoder D ψ subscript 𝐷 πœ“ D_{\psi}italic_D start_POSTSUBSCRIPT italic_ψ end_POSTSUBSCRIPT, and update the encoder to maximize the Equation6 while fixing decoders D Ο• subscript 𝐷 italic-Ο• D_{\phi}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT and D ψ subscript 𝐷 πœ“ D_{\psi}italic_D start_POSTSUBSCRIPT italic_ψ end_POSTSUBSCRIPT. We find that incorporating an additional regularization term on the slot attention mask is helpful in enhancing object-centric representations:

β„’ Reg⁒(ΞΈ)=𝐀 1β‹…sg⁒(‖𝐱 1βˆ’π± cβ€–2)+𝐀 2β‹…sg⁒(‖𝐱 2βˆ’π± cβ€–2),subscript β„’ Reg πœƒβ‹…superscript 𝐀 1 sg superscript norm superscript 𝐱 1 superscript 𝐱 𝑐 2β‹…superscript 𝐀 2 sg superscript norm superscript 𝐱 2 superscript 𝐱 𝑐 2\mathcal{L}_{\text{Reg}}(\theta)=\mathbf{A}^{1}\cdot\text{sg}(||\mathbf{x}^{1}% -\mathbf{x}^{c}||^{2})+\mathbf{A}^{2}\cdot\text{sg}(||\mathbf{x}^{2}-\mathbf{x% }^{c}||^{2}),caligraphic_L start_POSTSUBSCRIPT Reg end_POSTSUBSCRIPT ( italic_ΞΈ ) = bold_A start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT β‹… sg ( | | bold_x start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT - bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT | | start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ) + bold_A start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT β‹… sg ( | | bold_x start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT - bold_x start_POSTSUPERSCRIPT italic_c end_POSTSUPERSCRIPT | | start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT ) ,(8)

where 𝐀 1=𝐀⁒(𝐱 2,𝐒 1(n)),𝐀 2=𝐀⁒(𝐱 1,𝐒 2(n))formulae-sequence superscript 𝐀 1 𝐀 superscript 𝐱 2 superscript 𝐒 superscript 1 𝑛 superscript 𝐀 2 𝐀 superscript 𝐱 1 superscript 𝐒 superscript 2 𝑛\mathbf{A}^{1}=\mathbf{A}(\mathbf{x}^{2},\mathbf{S}^{1^{(n)}}),\mathbf{A}^{2}=% \mathbf{A}(\mathbf{x}^{1},\mathbf{S}^{2^{(n)}})bold_A start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT = bold_A ( bold_x start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT , bold_S start_POSTSUPERSCRIPT 1 start_POSTSUPERSCRIPT ( italic_n ) end_POSTSUPERSCRIPT end_POSTSUPERSCRIPT ) , bold_A start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT = bold_A ( bold_x start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT , bold_S start_POSTSUPERSCRIPT 2 start_POSTSUPERSCRIPT ( italic_n ) end_POSTSUPERSCRIPT end_POSTSUPERSCRIPT ) are attention masks from the last iteration of slot attention for 𝐱 1,𝐱 2 superscript 𝐱 1 superscript 𝐱 2\mathbf{x}^{1},\mathbf{x}^{2}bold_x start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT , bold_x start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT (Equation2), respectively, and sg⁒(β‹…)sgβ‹…\text{sg}(\cdot)sg ( β‹… ) denotes stop-gradient operator. It encourages the source and the composite images to be consistent over the object area captured by the slots, enhancing the content-preserving composition. The overall objective is then formulated as follow:

β„’ Total⁒(ΞΈ,Ο•,ψ)=Ξ» Prior⁒ℒ Prior⁒(ΞΈ)+Ξ» Diff⁒ℒ Diff⁒(ΞΈ,Ο•)+Ξ» Recon⁒ℒ Recon⁒(ΞΈ,ψ)+Ξ» Reg⁒ℒ Reg⁒(ΞΈ)subscript β„’ Total πœƒ italic-Ο• πœ“ subscript πœ† Prior subscript β„’ Prior πœƒ subscript πœ† Diff subscript β„’ Diff πœƒ italic-Ο• subscript πœ† Recon subscript β„’ Recon πœƒ πœ“ subscript πœ† Reg subscript β„’ Reg πœƒ\begin{gathered}\mathcal{L}{\text{Total}}(\theta,\phi,\psi)=\lambda{\text{% Prior}}\mathcal{L}{\text{Prior}}(\theta)+\lambda{\text{Diff}}\mathcal{L}{% \text{Diff}}(\theta,\phi)+\lambda{\text{Recon}}\mathcal{L}{\text{Recon}}(% \theta,\psi)+\lambda{\text{Reg}}\mathcal{L}_{\text{Reg}}(\theta)\end{gathered}start_ROW start_CELL caligraphic_L start_POSTSUBSCRIPT Total end_POSTSUBSCRIPT ( italic_ΞΈ , italic_Ο• , italic_ψ ) = italic_Ξ» start_POSTSUBSCRIPT Prior end_POSTSUBSCRIPT caligraphic_L start_POSTSUBSCRIPT Prior end_POSTSUBSCRIPT ( italic_ΞΈ ) + italic_Ξ» start_POSTSUBSCRIPT Diff end_POSTSUBSCRIPT caligraphic_L start_POSTSUBSCRIPT Diff end_POSTSUBSCRIPT ( italic_ΞΈ , italic_Ο• ) + italic_Ξ» start_POSTSUBSCRIPT Recon end_POSTSUBSCRIPT caligraphic_L start_POSTSUBSCRIPT Recon end_POSTSUBSCRIPT ( italic_ΞΈ , italic_ψ ) + italic_Ξ» start_POSTSUBSCRIPT Reg end_POSTSUBSCRIPT caligraphic_L start_POSTSUBSCRIPT Reg end_POSTSUBSCRIPT ( italic_ΞΈ ) end_CELL end_ROW(9)

where Ξ» Prior,Ξ» Diff,Ξ» Recon,Ξ» Reg subscript πœ† Prior subscript πœ† Diff subscript πœ† Recon subscript πœ† Reg\lambda_{\text{Prior}},\lambda_{\text{Diff}},\lambda_{\text{Recon}},\lambda_{% \text{Reg}}italic_Ξ» start_POSTSUBSCRIPT Prior end_POSTSUBSCRIPT , italic_Ξ» start_POSTSUBSCRIPT Diff end_POSTSUBSCRIPT , italic_Ξ» start_POSTSUBSCRIPT Recon end_POSTSUBSCRIPT , italic_Ξ» start_POSTSUBSCRIPT Reg end_POSTSUBSCRIPT are hyperparameters for controlling the importance of each term. We empirically find that Ξ» Prior=Ξ» Diff=Ξ» Recon=1.0,Ξ» Reg=0.25 formulae-sequence subscript πœ† Prior subscript πœ† Diff subscript πœ† Recon 1.0 subscript πœ† Reg 0.25\lambda_{\text{Prior}}=\lambda_{\text{Diff}}=\lambda_{\text{Recon}}=1.0,% \lambda_{\text{Reg}}=0.25 italic_Ξ» start_POSTSUBSCRIPT Prior end_POSTSUBSCRIPT = italic_Ξ» start_POSTSUBSCRIPT Diff end_POSTSUBSCRIPT = italic_Ξ» start_POSTSUBSCRIPT Recon end_POSTSUBSCRIPT = 1.0 , italic_Ξ» start_POSTSUBSCRIPT Reg end_POSTSUBSCRIPT = 0.25 generally works well and use it throughout the experiments.

4 Related Work

Object-centric learning

The most dominant paradigm of object-centric learning is employing the auto-encoding objective(Burgess et al., 2019; Greff et al., 2019; Engelcke et al., 2020; 2021; Lin et al., 2020; Jiang et al., 2020; Eslami et al., 2016; Crawford & Pineau, 2019). To guide the model to learn structured representation under reconstruction loss, Locatello et al. (2020) introduces Slot Attention, where each slot is iteratively refined with dot-product attention mechanism normalized in slot direction, inducing competition between the slots. Follow-up studies(Singh et al., 2022a; Seitzer et al., 2022; Sajjadi et al., 2022) demonstrate that Slot Attention with an auto-encoding objective has the potential to attain object-wise disentanglement even in complex scenes. Nonetheless, auto-encoding alone often involves training instability, which leads to attention-leaking problem(Kim et al., 2023), or dividing the scene into Voronoi tessellations(Sajjadi et al., 2022; Jiang et al., 2023). To overcome such challenges, there have been a few attempts on revising the learning objective such as replacing image reconstruction loss with denoising objective(Jiang et al., 2023; Wu et al., 2024) or contrastive loss(HΓ©naff et al., 2022; Wen et al., 2022). Nevertheless, these approaches still do not impose direct learning of object-centric representations.

Generative Prior

There are increasing interests in exploiting the knowledge pre-trained from generative prior to various applications such as solving inverse problems(Chung et al., 2023), guidance in conditional generation(Graikos et al., 2022; Liu et al., 2023), and image manipulations(Ruiz et al., 2023a; Zhang et al., 2023; Ruiz et al., 2023b). One prominent approach in this direction is text-to-3D Generation, where a large-scale pre-trained 2D diffusion model(Rombach et al., 2022; Saharia et al., 2022) is leveraged to generate realistic 3D data without ground-truth(Wang et al., 2023a; Lin et al., 2023; Metzer et al., 2023; Wang et al., 2023b). The seminal work by (Poole et al., 2022) formulates a loss based on a probability density distillation to distill a pre-trained 2D image prior to a 3D model. Back-propagating the loss through a randomly initialized 3D model, e.g., NeRF(Mildenhall et al., 2020), the model gradually updates to generate high-fidelity 3D renderings. Inspired by this line of work, we employ a generative model in our approach to maximize the validity of the given images.

5 Experiment

Implementation Details

We base our implementation on existing frameworks(Singh et al., 2022a; Jiang et al., 2023). We employ the features from the pre-trained auto-encoder 2 2 2 https://huggingface.co/stabilityai/sd-vae-ft-ema-original to represent an image. For the slot encoder, we employ the CNN based on UNet architecture(Singh et al., 2022b; Jiang et al., 2023) to produce a high-resolution attention map. Also, we employ an implicit Slot Attention(Chang et al., 2022) to stabilize the iterative refinement process in slot attention. For the slot mixing strategy, we opt for a sampling with sharing slot initializations for all the experiments unless specified, since it shows slightly better performance than the random sampling strategy. When we compute β„’ Prior subscript β„’ Prior\mathcal{L}{\text{Prior}}caligraphic_L start_POSTSUBSCRIPT Prior end_POSTSUBSCRIPT (Equation6), we use t min=0.02,t max=0.5 formulae-sequence subscript 𝑑 min 0.02 subscript 𝑑 max 0.5 t{\text{min}}=0.02,t_{\text{max}}=0.5 italic_t start_POSTSUBSCRIPT min end_POSTSUBSCRIPT = 0.02 , italic_t start_POSTSUBSCRIPT max end_POSTSUBSCRIPT = 0.5 following a recent report in (Wang et al., 2023b) that employing too high noise level impairs the optimization.

Datasets

We validate our method on four datasets. CLEVRTex(Karazija et al., 2021) consists of various rigid objects with homogeneous textures. MultiShapeNet(Stelzner et al., 2021) includes more complex and realistic furniture objects. PTR(Hong et al., 2021) and Super-CLEVR(Li et al., 2023) contain objects composed of multi-colored parts and textures. All of the datasets are center-cropped and resized to 128x128 resolution images. .

Baselines

We compare our method against two strong baselines in the literature, SLATE(Singh et al., 2022a) and LSD(Jiang et al., 2023), which employ autoregressive Transformer and diffusion-based decoders, respectively. Note that our method without composition path reduces to LSD. For a fair comparison, we employ the same encoder architecture based on slot attention(Locatello et al., 2020) in all compared methods including ours. For LSD and our method, we employ the same pre-trained auto-encoder(Rombach et al., 2022) to represent an input image. Since SLATE runs on discrete features, we employ the features from the pre-trained VQGAN model(Esser et al., 2021) and denote it as SLATE+. All baselines including ours are trained for 200K iterations.

Evaluation Metrics

Following the previous works(Jiang et al., 2023; Singh et al., 2022a; b; Chang et al., 2022), we report the unsupervised segmentation performance with three measures: Adjusted rand index for foreground objects (FG-ARI), mean intersection over union (mIoU), and mean best overlap (mBO). These metrics measure the overlap between the slot attention masks and ground-truth object masks, where FG-ARI focuses more on the coverage of the object area.

Figure 2: Comparison results on unsupervised object segmentation. We evaluate the how well the slot attention masks coincide with the ground-truth objects using FG-ARI, mIoU, and mBO (The higher is better). All results are evaluated on held-out validation set.

(a) CLEVRTex

Model FG-ARI mIoU mBO SLATE+71.29 52.04 52.17 LSD 76.44 72.32 72.44 Ours 93.06 74.82 75.36

(b) MultiShapeNet

Model FG-ARI mIoU mBO SLATE+70.44 15.55 15.64 LSD 67.72 15.39 15.46 Ours 89.8 59.21 59.4

(c) PTR

Model FG-ARI mIoU mBO SLATE+91.25 14.1 14.22 LSD 61.1 10.18 10.33 Ours 90.65 40.89 41.45

(d) Super-CLEVR

Model FG-ARI mIoU mBO SLATE+43.73 29.12 29.49 LSD 54.79 14.12 14.43 Ours 63.08 47.17 48.03

Image 2: Refer to caption

Figure 3: Qualitative results on unsupervised object segmentation. The baselines tend to split an object into different slots (CleverTex) and/or combine different objects and background into a single (MultiShapeNet, PGR, Super-CLEVR). On the other hand, our method produces consistently better object masks, showing improved disentanglement of objects and background in all datasets. More results are presented in the Figure8. Zoom in for better view.

5.1 Unsupervised Object Segmentation

We first present the comparison results of our method with baselines on unsupervised object segmentation. Table0(d) summarizes the quantitative results. Our method significantly improves the FG-ARI scores over the baselines in all datasets (8 to 29% improvement) except PTR, indicating that it captures an object holistically into an individual slot while the baselines tend to split the object into multiple parts and distribute it across multiple slots. In terms of mIoU and mBO, our method improves the baselines over all datasets, especially when the background is monolithic (MultiShapeNet, PTR, and Super-CLEVR). It indicates that the baselines struggle to separate the objects from the background when there exists a strong correlation between them, while our method can still robustly identify the objects. Overall, the results indicate that our method consistently outperforms the baselines by a significant margin. Notably, the consistent and significant improvement over LSD indicates that our regularization on the compositionality is effective in learning object-centric representation.

We also present the qualitative results in Figure3. It shows that SLATE frequently splits the foreground object masks into multiple segments in CLEVRTex and Super-CLEVR datasets, and fails to capture object entities in PTR and MultiShapeNet. Similarly, LSD fails to segment the object in all datasets except CLEVRTex dataset, and tends to rely on positional bias in PTR and Super-CLEVR. In contrast, our method consistently captures objects with tight boundaries.

Image 3: Refer to caption

(a) Number of slots

Image 4: Refer to caption

(b) Encoder architecture

Image 5: Refer to caption

(c) Decoder capacity

Figure 4: Robustness against various architectural biases. We evaluate the robustness of our model various different number of slots, encoder architectures, and decoder capacities. Results based on mIoU and mBO are presented in Figure6.

5.2 Robustness of Compositional Objective

Compared to approaches based on auto-encoding, our method directly incorporates the objective to learn compositional representation, thus is more robust to the choice of architectural biases and hyperparameters. To demonstrate this, we evaluate our method while varying three major factors that are known to be highly sensitive in the previous approaches, such as number of slots, encoder architecture, and decoder capacity. Figure4 summarizes the result on CLEVRTex dataset based on FG-ARI. All methods are trained up to 100K iterations for fair comparison.

Number of slots

Since object-centric learning assumes no prior knowledge on data, the mismatch between the number of objects and slots is inevitable in practice. To evaluate such robustness, we vary the number of slots from 11 to 17. Figure4(a) presents the result. It shows that the performance of the baselines is highly sensitive to the number of slots. Specifically, SLATE tends to deteriorate more as the number of slots increases. Compared to the baseline, our method achieves more robust performance by encoding an object into a slot while leaving excess slots empty.

Encoder architecture

To identify the effect of slot encoder, we consider two popular architectures in the literature; a multi-layer CNN encoder(Singh et al., 2022b) and UNet-based encoder(Ronneberger et al., 2015). Figure4(b) summarizes the result. It shows that employing the weaker encoder generally deteriorates the performance of the baselines significantly, indicating that architectural bias in the encoder is critical in the auto-encoding objective. Interestingly, the performance of our method is hardly affected by such drastic modifications, showing great robustness.

Decoder capacity

It is widely observed that the choice of decoder is also crucial in object-centric learning, since the highly expressive decoder can often bypass the object representation to minimize the reconstruction loss(Singh et al., 2022a). To examine such effect, we gradually increase the feature dimensions of the decoder to 133%percent%%, 166%percent%%, and 200%percent%%. Figure4(c) summarizes the result. It shows that increasing the decoder capacity hampers the performance in SLATE. LSD exhibits the opposite trends showing a large improvement in FG-ARI, although its performance drops significantly in mIoU (Figure6). Compared to the baselines, our method is much less sensitive to the decoder capacity, while the performance tends to improve slightly with increased capacity in all measures.

Overall, the results indicate that the quality of object-centric representation is significantly influenced by various factors in the auto-encoding-based methods. Conversely, our model consistently delivers outstanding performance across all configurations, even with major alterations to the encoder architecture. It demonstrates that our regularization through the composite path can directly encourage the model to learn compositional representation, greatly enhancing robustness to architectural biases.

Table 1: Ablation study on CLEVRTex dataset. All models are trained up to 100K iterations.

β„’ Prior subscript β„’ Prior\mathcal{L}{\text{Prior}}caligraphic_L start_POSTSUBSCRIPT Prior end_POSTSUBSCRIPT β„’ Reg subscript β„’ Reg\mathcal{L}{\text{Reg}}caligraphic_L start_POSTSUBSCRIPT Reg end_POSTSUBSCRIPT Share 𝐒(0)superscript 𝐒 0\mathbf{S}^{(0)}bold_S start_POSTSUPERSCRIPT ( 0 ) end_POSTSUPERSCRIPT FG-ARI mIoU mBO ✘✘✘42.48 52.26 52.41 βœ“βœ˜βœ˜65.76 67.72 67.62 βœ“βœ˜βœ“70.29 69.08 69.28 βœ˜βœ“βœ“65.26 58.81 58.99 βœ“βœ“βœ“88.15 75.30 75.64 Image 6: Refer to caption

Figure 5: Investigating object representation through compositional generation. We investigate the compositionality of learned representations by removing (red arrow) and adding (blue arrow) object slots between two images and generating the composite image. More results are in Figure9.

5.3 Internal Analysis

Component-wise Contributions

To identify the contributions of each component in our framework, we conduct an ablation study and present the result in Table1. The first row corresponds to our model with only the auto-encoding path, while the last row is the complete version of our model. Comparing the first row with the others shows that incorporating the composition path significantly improves overall quality. Adding β„’ Prior subscript β„’ Prior\mathcal{L}{\text{Prior}}caligraphic_L start_POSTSUBSCRIPT Prior end_POSTSUBSCRIPT, we observe a substantial improvement in all three metrics. Considering that FG-ARI measures the correct cluster membership of pixels within the objects, increased FG-ARI indicates that the generative prior encourages the encoder to capture more holistic object representations. This is because the generative prior penalizes the encoder for fragmenting the objects, thereby discouraging the generation of unrealistic partial objects in the composite image. Comparing the second and the third rows, we observe that sharing the slot initialization 𝐒(0)superscript 𝐒 0\mathbf{S}^{(0)}bold_S start_POSTSUPERSCRIPT ( 0 ) end_POSTSUPERSCRIPT slightly enhances mIoU and mBO scores. This improvement is likely attributed to the increased training stability by avoiding invalid slot combinations as shown in Figure7. Incorporating regularization β„’ reg subscript β„’ reg\mathcal{L}{\text{reg}}caligraphic_L start_POSTSUBSCRIPT reg end_POSTSUBSCRIPT alone in the composition path does not improve the performance (fourth row), while combined with generative prior, it leads to significant improvement.

Compositional Generation

We present the compositional generation results to further investigate the impact of our composition path. Figure5 presents the results. Given two images, we construct the composite representation by replacing one object slot from the first image (red arrow) to another from the second image (blue arrow), and producing the image by the decoder. Based on visualization of the learned slots, we observe that the baselines often fail to learn compositional slot representation, by separating objects into multiple slots or encoding background with an object. It leads to failures in object-level manipulation, such as retaining an object after the removal (LSD in MultiShapeNet and PTR), altering the content of the added object (SLATE in MultiShapeNet), or transforming background with the object (SLATE in PTR and LSD in Super-CLEVR). In contrast, our method produces both semantically meaningful and realistic images from composite slot representations, supporting our claim that we can regularize object-centric learning through the proposed compositional path.

6 Conclusion

In this paper, we introduced a method to address the misalignment between object-centric learning and the auto-encoding objective. Our method is based on auto-encoding framework, and incorporates an additional branch to directly assess the compositionality of the representation. This involves constructing composite representations from two separate images and optimizing the encoder jointly with the auto-encoding path to maximize the likelihood of the composite image. Despite the simplicity, our extensive experiments demonstrate that our framework consistently improves the object-centric learning over the auto-encoding frameworks. It also shows that our method greatly enhances the robustness to the choice of architectural biases and hyperparameters, which typically pose sensitivity challenges in auto-encoding-centric approaches.

Acknowledgements

This work was supported in part by Institute of Information & communications Technology Planning & Evaluation (IITP) grant (No.2022-0-00926, 2022-0-00959, 2021-0-02068, and 2019-0-00075) and National Research Foundation of Korea(NRF) grant (2021R1C1C1012540 and 2022R1C1C1009443) funded by the Korea government(MSIT).

References

  • Assouel et al. (2022) Rim Assouel, Pau Rodriguez, Perouz Taslakian, David Vazquez, and Yoshua Bengio. Object-centric compositional imagination for visual abstract reasoning. In ICLR2022 Workshop on the Elements of Reasoning: Objects, Structure and Causality, 2022.
  • Bogin et al. (2021) Ben Bogin, Sanjay Subramanian, Matt Gardner, and Jonathan Berant. Latent compositional representations improve systematic generalization in grounded question answering. Transactions of the Association for Computational Linguistics, 9:195–210, 2021.
  • Burgess et al. (2019) Christopher P Burgess, Loic Matthey, Nicholas Watters, Rishabh Kabra, Irina Higgins, Matt Botvinick, and Alexander Lerchner. Monet: Unsupervised scene decomposition and representation. arXiv preprint arXiv:1901.11390, 2019.
  • Chang et al. (2022) Michael Chang, Tom Griffiths, and Sergey Levine. Object representations as fixed points: Training iterative refinement algorithms with implicit differentiation. In NeurIPS, volume 35, pp. 32694–32708, 2022.
  • Chung et al. (2023) Hyungjin Chung, Dohoon Ryu, Michael T McCann, Marc L Klasky, and Jong Chul Ye. Solving 3d inverse problems using pre-trained 2d diffusion models. In CVPR, 2023.
  • Crawford & Pineau (2019) Eric Crawford and Joelle Pineau. Spatially invariant unsupervised object detection with convolutional neural networks. In AAAI, 2019.
  • D’Amario et al. (2021) Vanessa D’Amario, Tomotake Sasaki, and Xavier Boix. How modular should neural module networks be for systematic generalization? In NeurIPS, 2021.
  • Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In Jill Burstein, Christy Doran, and Thamar Solorio (eds.), Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171–4186, Minneapolis, Minnesota, June 2019. Association for Computational Linguistics. doi: 10.18653/v1/N19-1423. URL https://aclanthology.org/N19-1423.
  • Dittadi et al. (2022) Andrea Dittadi, Samuele Papa, Michele De Vita, Bernhard SchΓΆlkopf, Ole Winther, and Francesco Locatello. Generalization and robustness implications in object-centric learning. In ICML, 2022.
  • Engelcke et al. (2020) Martin Engelcke, Adam R Kosiorek, Oiwi Parker Jones, and Ingmar Posner. Genesis: Generative scene inference and sampling with object-centric latent representations. In ICLR, 2020.
  • Engelcke et al. (2021) Martin Engelcke, Oiwi Parker Jones, and Ingmar Posner. Genesis-v2: Inferring unordered object representations without iterative refinement. In NeurIPS, 2021.
  • Eslami et al. (2016) SM Eslami, Nicolas Heess, Theophane Weber, Yuval Tassa, David Szepesvari, Geoffrey E Hinton, et al. Attend, infer, repeat: Fast scene understanding with generative models. In NeurIPS, 2016.
  • Esser et al. (2021) Patrick Esser, Robin Rombach, and Bjorn Ommer. Taming transformers for high-resolution image synthesis. In CVPR, 2021.
  • Graikos et al. (2022) Alexandros Graikos, Nikolay Malkin, Nebojsa Jojic, and Dimitris Samaras. Diffusion models as plug-and-play priors. In NeurIPS, 2022.
  • Greff et al. (2019) Klaus Greff, RaphaΓ«l Lopez Kaufman, Rishabh Kabra, Nick Watters, Christopher Burgess, Daniel Zoran, Loic Matthey, Matthew Botvinick, and Alexander Lerchner. Multi-object representation learning with iterative variational inference. In ICML, 2019.
  • HΓ©naff et al. (2022) Olivier J HΓ©naff, Skanda Koppula, Evan Shelhamer, Daniel Zoran, Andrew Jaegle, Andrew Zisserman, JoΓ£o Carreira, and Relja ArandjeloviΔ‡. Object discovery and representation networks. In ECCV, 2022.
  • Ho et al. (2020) Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. In NeurIPS, 2020.
  • Hong et al. (2021) Yining Hong, Li Yi, Josh Tenenbaum, Antonio Torralba, and Chuang Gan. Ptr: A benchmark for part-based conceptual, relational, and physical reasoning. In NeurIPS, 2021.
  • Jiang et al. (2020) Jindong Jiang, Sepehr Janghorbani, Gerard De Melo, and Sungjin Ahn. Scalor: Generative world models with scalable object representations. In ICLR, 2020.
  • Jiang et al. (2023) Jindong Jiang, Fei Deng, Gautam Singh, and Sungjin Ahn. Object-centric slot diffusion. In NeurIPS, 2023.
  • Karazija et al. (2021) Laurynas Karazija, Iro Laina, and Christian Rupprecht. ClevrTex: A Texture-Rich Benchmark for Unsupervised Multi-Object Segmentation. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2021.
  • Kim et al. (2023) Jinwoo Kim, Janghyuk Choi, Ho-Jin Choi, and Seon Joo Kim. Shepherding slots to objects: Towards stable and robust object-centric learning. In CVPR, 2023.
  • Kumari et al. (2023) Nupur Kumari, Bingliang Zhang, Richard Zhang, Eli Shechtman, and Jun-Yan Zhu. Multi-concept customization of text-to-image diffusion. In CVPR, 2023.
  • Kuo et al. (2021) Yen-Ling Kuo, Boris Katz, and Andrei Barbu. Compositional networks enable systematic generalization for grounded language understanding. In EMNLP, 2021.
  • Lake et al. (2017) Brenden M Lake, Tomer D Ullman, Joshua B Tenenbaum, and Samuel J Gershman. Building machines that learn and think like people. Behavioral and brain sciences, 40:e253, 2017.
  • Li et al. (2023) Zhuowan Li, Xingrui Wang, Elias Stengel-Eskin, Adam Kortylewski, Wufei Ma, Benjamin Van Durme, and Alan L Yuille. Super-clevr: A virtual benchmark to diagnose domain robustness in visual reasoning. In CVPR, pp. 14963–14973, 2023.
  • Lin et al. (2023) Chen-Hsuan Lin, Jun Gao, Luming Tang, Towaki Takikawa, Xiaohui Zeng, Xun Huang, Karsten Kreis, Sanja Fidler, Ming-Yu Liu, and Tsung-Yi Lin. Magic3d: High-resolution text-to-3d content creation. In CVPR, 2023.
  • Lin et al. (2020) Zhixuan Lin, Yi-Fu Wu, Skand Vishwanath Peri, Weihao Sun, Gautam Singh, Fei Deng, Jindong Jiang, and Sungjin Ahn. Space: Unsupervised object-oriented scene representation via spatial attention and decomposition. In ICLR, 2020.
  • Liu et al. (2023) Xihui Liu, Dong Huk Park, Samaneh Azadi, Gong Zhang, Arman Chopikyan, Yuxiao Hu, Humphrey Shi, Anna Rohrbach, and Trevor Darrell. More control for free! image synthesis with semantic diffusion guidance. In WACV, 2023.
  • Locatello et al. (2020) Francesco Locatello, Dirk Weissenborn, Thomas Unterthiner, Aravindh Mahendran, Georg Heigold, Jakob Uszkoreit, Alexey Dosovitskiy, and Thomas Kipf. Object-centric learning with slot attention. In NeurIPS, 2020.
  • Metzer et al. (2023) Gal Metzer, Elad Richardson, Or Patashnik, Raja Giryes, and Daniel Cohen-Or. Latent-nerf for shape-guided generation of 3d shapes and textures. In CVPR, 2023.
  • Mildenhall et al. (2020) Ben Mildenhall, Pratul P Srinivasan, Matthew Tancik, Jonathan T Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. In ECCV, 2020.
  • Oquab et al. (2023) Maxime Oquab, TimothΓ©e Darcet, ThΓ©o Moutakanni, Huy Vo, Marc Szafraniec, Vasil Khalidov, Pierre Fernandez, Daniel Haziza, Francisco Massa, Alaaeldin El-Nouby, et al. Dinov2: Learning robust visual features without supervision. arXiv preprint arXiv:2304.07193, 2023.
  • Poole et al. (2022) Ben Poole, Ajay Jain, Jonathan T Barron, and Ben Mildenhall. Dreamfusion: Text-to-3d using 2d diffusion. In ICLR, 2022.
  • Rahaman et al. (2021) Nasim Rahaman, Muhammad Waleed Gondal, Shruti Joshi, Peter Gehler, Yoshua Bengio, Francesco Locatello, and Bernhard SchΓΆlkopf. Dynamic inference with neural interpreters. In NeurIPS, 2021.
  • Robbins (1992) Herbert E Robbins. An empirical bayes approach to statistics. In Breakthroughs in Statistics: Foundations and basic theory, pp. 388–394. Springer, 1992.
  • Rombach et al. (2022) Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and BjΓΆrn Ommer. High-resolution image synthesis with latent diffusion models. In CVPR, pp. 10684–10695, 2022.
  • Ronneberger et al. (2015) Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In Nassir Navab, Joachim Hornegger, William M. Wells, and Alejandro F. Frangi (eds.), Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015, pp. 234–241, Cham, 2015. Springer International Publishing. ISBN 978-3-319-24574-4.
  • Ruiz et al. (2023a) Nataniel Ruiz, Yuanzhen Li, Varun Jampani, Yael Pritch, Michael Rubinstein, and Kfir Aberman. Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation. In CVPR, 2023a.
  • Ruiz et al. (2023b) Nataniel Ruiz, Yuanzhen Li, Varun Jampani, Wei Wei, Tingbo Hou, Yael Pritch, Neal Wadhwa, Michael Rubinstein, and Kfir Aberman. Hyperdreambooth: Hypernetworks for fast personalization of text-to-image models. arXiv preprint arXiv:2307.06949, 2023b.
  • Saharia et al. (2022) Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily L Denton, Kamyar Ghasemipour, Raphael Gontijo Lopes, Burcu Karagol Ayan, Tim Salimans, et al. Photorealistic text-to-image diffusion models with deep language understanding. In NeurIPS, 2022.
  • Sajjadi et al. (2022) Mehdi SM Sajjadi, Daniel Duckworth, Aravindh Mahendran, Sjoerd van Steenkiste, Filip Pavetic, Mario Lucic, Leonidas J Guibas, Klaus Greff, and Thomas Kipf. Object scene representation transformer. In NeurIPS, 2022.
  • Seitzer et al. (2022) Maximilian Seitzer, Max Horn, Andrii Zadaianchuk, Dominik Zietlow, Tianjun Xiao, Carl-Johann Simon-Gabriel, Tong He, Zheng Zhang, Bernhard SchΓΆlkopf, Thomas Brox, et al. Bridging the gap to real-world object-centric learning. In ICLR, 2022.
  • Singh et al. (2022a) Gautam Singh, Fei Deng, and Sungjin Ahn. Illiterate dall-e learns to compose. In ICLR, 2022a.
  • Singh et al. (2022b) Gautam Singh, Yi-Fu Wu, and Sungjin Ahn. Simple unsupervised object-centric learning for complex and naturalistic videos. In NeurIPS, 2022b.
  • Spelke (1990) Elizabeth S Spelke. Principles of object perception. Cognitive science, 14(1):29–56, 1990.
  • Stein (1981) Charles M Stein. Estimation of the mean of a multivariate normal distribution. The annals of Statistics, pp. 1135–1151, 1981.
  • Stelzner et al. (2021) Karl Stelzner, Kristian Kersting, and Adam R Kosiorek. Decomposing 3d scenes into objects via unsupervised volume segmentation. arXiv preprint arXiv:2104.01148, 2021.
  • Wang et al. (2023a) Haochen Wang, Xiaodan Du, Jiahao Li, Raymond A Yeh, and Greg Shakhnarovich. Score jacobian chaining: Lifting pretrained 2d diffusion models for 3d generation. In CVPR, 2023a.
  • Wang et al. (2023b) Zhengyi Wang, Cheng Lu, Yikai Wang, Fan Bao, Chongxuan Li, Hang Su, and Jun Zhu. Prolificdreamer: High-fidelity and diverse text-to-3d generation with variational score distillation. In NeurIPS, 2023b.
  • Wen et al. (2022) Xin Wen, Bingchen Zhao, Anlin Zheng, Xiangyu Zhang, and Xiaojuan Qi. Self-supervised visual representation learning with semantic grouping. In NeurIPS, 2022.
  • Wu et al. (2024) Ziyi Wu, Jingyu Hu, Wuyue Lu, Igor Gilitschenski, and Animesh Garg. Slotdiffusion: Object-centric generative modeling with diffusion models. In NeurIPS, 2024.
  • Xiao et al. (2022) Zhisheng Xiao, Karsten Kreis, and Arash Vahdat. Tackling the generative learning trilemma with denoising diffusion gans. In ICLR, 2022.
  • Yu et al. (2020) Fisher Yu, Haofeng Chen, Xin Wang, Wenqi Xian, Yingying Chen, Fangchen Liu, Vashisht Madhavan, and Trevor Darrell. Bdd100k: A diverse driving dataset for heterogeneous multitask learning. In CVPR, 2020.
  • Zhang et al. (2023) Yuxin Zhang, Nisha Huang, Fan Tang, Haibin Huang, Chongyang Ma, Weiming Dong, and Changsheng Xu. Inversion-based style transfer with diffusion models. In CVPR, 2023.

Appendix A Additional Implementation Details

Table2 provides details of hyperparameters used in experiments. For the Slot Attention encoder E ΞΈ subscript 𝐸 πœƒ E_{\theta}italic_E start_POSTSUBSCRIPT italic_ΞΈ end_POSTSUBSCRIPT and a diffusion decoder D Ο• subscript 𝐷 italic-Ο• D_{\phi}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT, we base our implementation on Jiang et al. (2023). Specifically, in the Slot Attention encoder, we employ a CNN-based UNet image encoder. Prior to the UNet encoder, we incorporate a single layer CNN to downsample the original 128Γ—128 128 128 128\times 128 128 Γ— 128 image to a 64Γ—64 64 64 64\times 64 64 Γ— 64 image. Implementing the diffusion decoder D Ο• subscript 𝐷 italic-Ο• D_{\phi}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT, we follow the design of the LSD decoder. The overall structure of D Ο• subscript 𝐷 italic-Ο• D_{\phi}italic_D start_POSTSUBSCRIPT italic_Ο• end_POSTSUBSCRIPT is based on the U-Net architecture, where each layer is composed of CNN layers and a transformer layer. The surrogate decoder D ψ subscript 𝐷 πœ“ D_{\psi}italic_D start_POSTSUBSCRIPT italic_ψ end_POSTSUBSCRIPT is implemented with the Transformer Architecture in Singh et al. (2022a). It takes slots as input through cross-attention layers. In the experimental setting, we augment the Super-CLEVR dataset by randomly altering the background color to another color.

General Batch Size 64 Training Steps 200K Learning Rate 0.0001 CNN Backbone Input Resolution 128 Output Resolution 64 Self Attention Middle Layer Base Channels 128 Channel Multipliers[1,1,2,4]

Heads 8

Res Blocks / Layer 2

Slot Size 192 Slot Attention Input Resolution 64

Iterations 7

Slot Size 192 Auto-Encoder Model KL-8 Input Resolution 128 Output Resolution 16 Output Channels 4 Diffusion Decoder Input Resolution 16 Input Channels 4 Ξ² 𝛽\beta italic_Ξ² scheduler Linear Mid Layer Attention Yes

Res Blocks / Layer 2

Heads 8

Base Channels 192 Attention Resolution[1,2,4,4] Channel Multipliers[1,2,4,4] Surrogate Decoder Layers 8

Heads 8

Hidden Dim 384

Table 2: Hyperparameters used in our experiments.

Appendix B Additional Results

B.1 Additional Results on Robustness Tests

We include results of the robustness test on mIoU, mBO metrics in Figure6. Similar to the results on FG-ARI (Figure4), our model is surprisingly robust to a wide range of hyperparameters. It suggests that directly optimizing the compositionality of the representation significantly reduce a dependency on a choice of hyperparameters.

Image 7: Refer to caption

Image 8: Refer to caption

Image 9: Refer to caption

Image 10: Refer to caption

(a) Number of slots

Image 11: Refer to caption

(b) Encoder architecture

Image 12: Refer to caption

(c) Decoder capacity

Figure 6: Robustness across Various Hyperparameters. We evaluate the robustness of our model across different number of slots, encoder type, and decoder capacity. Among various hyperparameters, our model steadily shows powerful performance against baselines.

B.2 Unsupervised Object Segmentation

We present additional qualitative results for unsupervised segmentation results in Figure8. Our method successfully segmented the object regions across all four datasets. In contrast, baselines easily divide each object into multiple segments or capture a wide area around the objects.

B.3 Effect of Mixing Slot Strategy

As discussed in Section3.1 and Section5.3, sharing 𝐒(0)superscript 𝐒 0\mathbf{S}^{(0)}bold_S start_POSTSUPERSCRIPT ( 0 ) end_POSTSUPERSCRIPT slightly enhances the performance by roughly avoiding suspicious compositions during training. To investigate how sharing slot initialization affects the composition, we obtained the slot representations from multiple scenes with the same slot initialization and grouped those representations by their order, i.e., 𝐬 i subscript 𝐬 𝑖\mathbf{s}_{i}bold_s start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT belongs to i 𝑖 i italic_i-th group. Figure7, we observe that the captured objects from the same initialization are correlated to some degree. The slots in the first row mostly capture the backgrounds of the scenes, while other slots tend to capture foreground objects. Moreover, we observe that the slots in the fourth row tend to capture the objects located in the lower part of the scene. Based on these observations, we conjecture that sharing slot initialization stabilizes our framework by alleviating some suspicious compositions, such as the occlusion of foreground objects or composing multiple backgrounds.

Image 13: Refer to caption

Figure 7: Grouping of slots by sharing the slot initialization. We obtain slot representation from various images while sharing the initial values of 𝐒(0)superscript 𝐒 0\mathbf{S}^{(0)}bold_S start_POSTSUPERSCRIPT ( 0 ) end_POSTSUPERSCRIPT and cluster the representation based on their initial values. Slots initialized as 𝐬 1 subscript 𝐬 1\mathbf{s}_{1}bold_s start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT consistently capture backgrounds.

Image 14: Refer to caption

Figure 8: More Qualitative Results on Unsupervised Object Segmentation.

Image 15: Refer to caption

Figure 9: More Qualitative Results on Compositional Generation between two images.

B.4 Investigation on Compositionality of Slots

In this section, we provide more visual samples of composite images to investigate the compositionality of slot representations in our method. Figure9 illustrates the results of generating composite images by mixing slots from two images, which supplements the Figure5 in the main paper. It shows that the baselines often fail to capture compositional objects into independent slots, while our method successfully learns object-level slots through the composition path. As a result, the composite images generated by the baselines often fail to adhere the object-level manipulation, such as retaining the removed objects or transforming the object identity and background pattern while adding a new object. In contrast, our method preserves these semantics more precisely based on accurate object slots.

B.5 Additional qualitative results on compositional generation

Image 16: Refer to caption

Figure 10: More Qualitative Results on Compositional Generation between two images.

To help a comprehensive understanding of the baselines, we provide more qualitative samples on compositional generation in Figure10. While Figure5 and Figure9 illustrates the common failure cases of the baselines, we additionally present compositional generation results where the baselines also reasonably capture an object into a slot. Despite the reasonable slot attention masks, the composite image produced by the baseline model often distorts the original appearance of the object or creates unrealistic partial objects. In contrast, our model consistently produces faithful composite images, which highlights the importance of the compositional objective.

B.6 Additional Evaluation on Object Property Prediction

To assess the quality of acquired object representations, we employ object property prediction using the learned representation, following the methodology outlined in Jiang et al. (2023); Dittadi et al. (2022). During this process, we train a network to predict the property based on a fixed slot representation. The true label for the slot representation is established through Hungarian matching, comparing the mask of slots with the foreground objects. The remaining slots after matching are considered as backgrounds. For predicting properties, we employ a 4-layer MLPs with a hidden dimension of 196. Accuracy is reported for categorical properties, while mean squared error is reported for continuous properties. We assess the models on datasets that include object properties.

The results for object property prediction are presented in Table3. Our model consistently performs better than the baselines across different properties and datasets. Notably, it excels in predicting shape and position, as observed in the high segmentation performance depicted in Figure3 and Table0(d). Furthermore, our model demonstrates improved performance in predicting materials indicating its ability to capture local and high-frequency information.

On the Super-CLEVR dataset, despite our model’s higher segmentation performance, the mean square error of position remains competitive with other baselines. We attribute this to the challenging nature of the dataset, where scenes often include many small and occluded objects. As a result, both our model and the baselines face increased difficulty in predicting position, leading to a higher error rate compared to other datasets.

Table 3: Results on object property prediction. We evaluate the quality of the learned representation through object property prediction. Our model consistently performs better than the baselines across different properties and datasets.

Dataset CLEVRTex PTR Super-CLEVR Property Position Shape Material Position Shape Position Shape Material (↓)↓(\downarrow)( ↓ )(↑)↑(\uparrow)( ↑ )(↑)↑(\uparrow)( ↑ )(↓)↓(\downarrow)( ↓ )(↑)↑(\uparrow)( ↑ )(↓)↓(\downarrow)( ↓ )(↑)↑(\uparrow)( ↑ )(↑)↑(\uparrow)( ↑ ) SLATE+0.1757 78.72 67.99 0.2218 88.21 0.5397 76.28 68.43 LSD 0.1563 85.07 82.33 0.5999 75.80 0.4372 76.5 69.24 Ours 0.1044 88.86 84.29 0.1424 90.00 0.4262 80.67 71.31

B.7 Additional Results on Real-world dataset

To explore the scalability of our novel objective in a complex real-world dataset, we examine our framework in BDD100k dataset Yu et al. (2020), which consists of diverse driving scenes. Since the images captured on night or rainy days often produce blurry and dark images, we filter the data to collect only sunny and daytime images using metadata, which leaves about 12k, 1.7k images in the training/validation set, respectively. Since it has been widely observed that learning the object-centric representation directly on real-world dataset is challenging, we bootstrap our auto-encoding path with off-the-shelf models following Jiang et al. (2023). Specifically, we employ pretrained DINOv2 Oquab et al. (2023) and Stable Diffusion Rombach et al. (2022) for the image encoder and slot decoder in our auto-encoding path, respectively. Instead of using frozen Stable-Diffusion, we update key and value mapping layers in cross-attention layers to enhance the overall auto-encoding performance following Kumari et al. (2023). For efficient training, we first warm up the auto-encoding path for 200k iterations and then train only the surrogate decoder for 140k iterations on top of frozen slot representations, which significantly boosts up the convergence of the surrogate decoder. Finally, we optimize our compositional path for 100k iterations. For the baseline, we compare our model trained with only auto-encoding objective for 300k iterations, which converges closely to the Stable-LSD Jiang et al. (2023).

Figure11 illustrates qualitative results on unsupervised object segmentation. The slot attention masks of our model successfully capture composable instances such as cars, buildings, trees, font hoods, etc. In contrast, the diffusion model trained without compositional objective often divides the objects into multiple slots or encodes multiple objects into a slot. For example, the car or truck is frequently divided into multiple masks, and multiple cars are often encoded into a single slot.

To further examine the compositionality of the learned slot representations, we qualitatively analyze the visual samples of composite images in Figure12 similar to SectionB.4. We observe that our method successfully generates realistic scenes, modeling complex correlations among objects and environments. It appropriately adapts the appearance of newly added/removed objects, their shadow, reflections in the front glass and hood, and sometimes even global illumination change caused by removing the sun. In contrast, the auto-encoding model often fails to achieve faithful composition. For example, in Row 1 of Figure12, the car still appears in the composite image even after the removal of the corresponding slot. Also, we observe that removing slots containing partial information of the object often leads to undesirable artifacts in composite images such as creating a new car in the first example of Row 2, or leaving unrealistic artifacts in the third example of Row 2. In contrast, our model produces natural object-wise manipulation. Moreover, the baseline model often fails to faithfully generate the inserted object as shown in Row 3, while our model tends to maintain the target object. In Row 4, we identify that our model successfully models complex interaction between slots such as removing sunlight changing the reflection of the bonnet in the first image, or changing a blurry car into a sharp car corresponding to bright weather. In summary, we identify that our novel objective on compositionality can help to learn object-wise disentanglement even in complex scenes and helps to model complex interactions among objects.

Image 17: Refer to caption

Figure 11: Qualitative results on unsupervised segmentation in BDD100k.

Image 18: Refer to caption

Figure 12: Qualitative results on compositional generation in BDD100k.

Xet Storage Details

Size:
101 kB
Β·
Xet hash:
0e0a65900f1800a4c351d31e6ce25979abc0ef595b7030f54e6de80be1583152

Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.