I've been researching neural networks and boolean formulae. From my efforts, it doesn't seem that neural networks can generally learn boolean formulae using back propagation. This makes sense intuitively, since boolean formulae outputs can exhibit dramatic changes based on input values, and so there will be many discontinuities and consequently local optima.
On the other hand, I also understand that any boolean formula can be represented with a neural network, as per the Universal Approximation Theorem, so there is no inherent reason why a neural network cannot represent, and consequently possibly learn an arbitrary boolean formula. The problem seems to be the learning algorithm, and all general machine learning algorithms seem like they'll get stuck in local optima, regardless of whether I use gradient descent, evolutionary algorithms, expectation maximization, and so on, since they all are based on the premise that local incremental improvements are the route to a global optimum solution.
That being said, I also know that there are other sorts of algorithms like Quine-McCluskey and Espresso that can derive minimal boolean formulae from truth tables. It would be possible to use these algorithms to subsequently generate a neural network that embeds the minimal boolean formula the algorithm derived from a truth table. Or, even simpler, just translate the truth table into a neural network. However, these are very specific algorithms that target boolean formulae in particular, and are not used in a more general machine learning context, as far as I am aware.
So, this brings me to my question. Are there any proofs demonstrating that neural networks are capable or incapable of learning arbitrary boolean formulae using backpropagation and gradient descent, or any other general purpose machine learning algorithms?
I have checked Cross Validated, and Googled the question, but have been unable to find anything definitive.