Circuit Complexity of Visual Search

Abstract

We study computational hardness of feature and conjunction search through the lens of circuit complexity. Let x=(x1,...,xn)x = (x_1, ... , x_n) (resp., y=(y1,...,yn)y = (y_1, ... , y_n)) be Boolean variables each of which takes the value one if and only if a neuron at place ii detects a feature (resp., another feature). We then simply formulate the feature and conjunction search as Boolean functions FTRn(x)=⋁i=1nxi{\rm FTR}_n(x) = \bigvee_{i=1}^n x_i and CONJn(x,y)=⋁i=1nxi∧yi{\rm CONJ}_n(x, y) = \bigvee_{i=1}^n x_i \wedge y_i, respectively. We employ a threshold circuit or a discretized circuit (such as a sigmoid circuit or a ReLU circuit with discretization) as our models of neural networks, and consider the following four computational resources: [i] the number of neurons (size), [ii] the number of levels (depth), [iii] the number of active neurons outputting non-zero values (energy), and [iv] synaptic weight resolution (weight). We first prove that any threshold circuit CC of size ss, depth dd, energy ee and weight ww satisfies log⁑rk(MC)≀ed(log⁑s+log⁑w+log⁑n)\log rk(M_C) \le ed (\log s + \log w + \log n), where rk(MC)rk(M_C) is the rank of the communication matrix MCM_C of a 2n2n-variable Boolean function that CC computes. Since CONJn{\rm CONJ}_n has rank 2n2^n, we have n≀ed(log⁑s+log⁑w+log⁑n)n \le ed (\log s + \log w + \log n). Thus, an exponential lower bound on the size of even sublinear-depth threshold circuits exists if the energy and weight are sufficiently small. Since FTRn{\rm FTR}_n is computable independently of nn, our result suggests that computational capacity for the feature and conjunction search are different. We also show that the inequality is tight up to a constant factor if ed=o(n/log⁑n)ed = o(n/ \log n). We next show that a similar inequality holds for any discretized circuit. Thus, if we regard the number of gates outputting non-zero values as a measure for sparse activity, our results suggest that larger depth helps neural networks to acquire sparse activity

    Similar works

    Full text

    thumbnail-image

    Available Versions