Overview
Rapid development in artificial intelligence (AI) research has led to increasingly larger artificial neural networks (ANNs). Some of the largest ANNs now have parameter sizes that rival the neuron and synaptic counts of intelligent biological organisms. However, these models have yet to demonstrate the capacity to reason in areas outside of their training domain, leaving a gap in AI research efforts towards artificial general intelligence (AGI).
The combined inability of ANNs to replicate the complex graph structure and temporal statefulness of information travel in biological neural circuits we believe results in ANN hypothesis classes that are too narrow for general reasoning. To address both of these issues, we propose ANT, a stateful, toplogoically nonlinear network formulation.
Unlike other RL formulations, ANT trains with a combined gradient descent and genetic algorithm to properly search the enlarged hyperparameter space brought upon by lifted restrictions on the neural graph.
Results and Current Direction
We found that ANT demonstrates superior performance and task-generalizable capabilities in reinforcement learning (RL) settings compared to conventional artificial neural networks. Specifically, we were able to demonstrate that, with equal parameters, ANT converged quicker, more regularly, and to a better solution than an analogous ANN running on a similar algorithm.
ANT has unique computational benefits and challenges associated with it. Unlike neural networks which have linear complexity with respect to depth and sublinear complexity elsewhere, ANT has linear complexity with respect to neuron size but is entirely parallelizable, as individual neuron computations can be performed independently.
Collaborators
Jonah Schwam (M.S. CS, Brown University), Pavani Nerella (M.S. CS, Brown University), Ilija Nikolov (Ph.D. Physics, Brown University), Taishi Nishizawa (M.S. CS, Brown University)