A neural-network learning theory and a polynomial time RBF algorithm

Asim Roy, Sandeep Govil, Raymond Miranda

Research output: Contribution to journalArticlepeer-review

68 Scopus citations


This paper presents a new learning theory (a set of principles for brain-like learning) and a corresponding algorithm for the neural-network field. The learning theory defines computational characteristics that are much more brain-like than that of classical connectionist learning. Robust and reliable learning algorithms would result if these learning principles are followed rigorously when developing neural-network algorithms. This paper also presents a new algorithm for generating radial basis function (RBF) nets for function approximation. The design of the algorithm is based on the proposed set of learning principles. The net generated by this algorithm is not a typical RBF net, but a combination of "truncated" RBF and other types of hidden units. The algorithm uses random clustering and linear programming (LP) to design and train this "mixed"RBF net. Polynomial time complexity of the algorithm is proven and computational results are provided for the well-known Mackey-Glass chaotic time series problem, the logistic map prediction problem, various neuro-control problems, and several time series forecasting problems. The algorithm can also be implemented as an on-line adaptive algorithm.

Original languageEnglish (US)
Pages (from-to)1301-1313
Number of pages13
JournalIEEE Transactions on Neural Networks
Issue number6
StatePublished - 1997


  • Designing neural networks
  • Feedforward nets
  • Learning complexity
  • Learning theory
  • Linear programming
  • Polynomial time complexity
  • Radial basis function networks

ASJC Scopus subject areas

  • Software
  • Computer Science Applications
  • Computer Networks and Communications
  • Artificial Intelligence


Dive into the research topics of 'A neural-network learning theory and a polynomial time RBF algorithm'. Together they form a unique fingerprint.

Cite this