Differencing based self-supervised scene change detection (D-SSCD) with temporal consistency
Abstract:
A computer implemented network for executing a self-supervised scene change detection method in which image pairs (T0, T1) from different time instances are subjected to random photometric transformations to obtain two pairs of augmented images (T0→T0′, T0″; T1→T1′, T1″), which augmented images are passed into an encoder (fθ) and a projection head (gϕ) to provide corresponding feature representations. Absolute feature differencing is applied over the outputs of the projection head (gϕ) to obtain difference representations (d1, d2) of changed features between the pair of images, and a self-supervised objective function (LSSL) is applied on the difference representations d1 and d2 to maximize a cross-correlation of the changed features, wherein d1 and d2 are defined as









d
1

=



"\[LeftBracketingBar]"



g

(

f

(

T
0


)

)

-

g

(

f

(

T
0


)

)




"\[RightBracketingBar]"








d
2

=



"\[LeftBracketingBar]"



g

(

f

(

T
0


)

)

-

g

(

f

(

T
1


)

)




"\[RightBracketingBar]"







(
1
)







Furthermore, an invariant prediction and change consistency loss is applied in the D-SSCD Network to reduce the effects of differences in the lighting conditions or camera viewpoints by enhancing the image alignment between the temporal images in the decision and feature space.
Information query
Patent Agency Ranking
0/0