CHARMM c42b2 gnn.doc



File: gnn -=- Node: Top
Up: (commands.doc) -=- Next: Syntax


                         Genetic Neural Network

A genetic neural network (GNN) method is provided for efficient determination
of quantitative structure property relationships. See the references given
below for a description of the GNN and its application. Some details specific
to the CHARMM implementation follow.

The GNN keyword must be included in pref.dat for the code to be compiled.

The input and output vectors of the data set are internally scaled to take 
values between 0.1 and 0.9. The format of the data file is described in the 
examples section.

Steepest descent back-propagation neural network is used to evaluate model
predictive quality. Jackknife cross-validation residual rms errors are
reported if no test data are specified. Only one hidden layer is employed.

Exhaustive enumeration and two genetic algorithm variants, genetic function
approximation (GFA) and evolutionary programming (EP), are available for
selecting models (sets of descriptors). The stochastic reminder method and
elitism are included for GFA reproduction.

* Menu:

* Syntax::           Syntax required to invoke GNN
* Description::      Description of GNN specific keywords
* Examples::         Examples
* References::       References



File: gnn -=- Node: Syntax
Up: Top -=- Next: Description -=- Previous: Top


                     Syntax required to invoke GNN

GNN [ data-spec ] [ nn-spec ] [ ga-spec ]

data-spec ::= [ NDATa 1 ] [ NPROd 0 ] [ NPARa 1 ] [ UNIT -1 ] [ SEED 123 ]

nn-spec ::= [ NDES 1 ] [ NHIDden 2 ] [ NTARg 1 ] [ NSWEep 100 ] [ MU 0.5 ] [ ETA 0.5 ]

ga-spec ::= [ EXHAust ] [ GFA ] [ EP ] [ NPOPu 500 ] [ NGEN 200 ] [ FITNess 5.0 ]



File: gnn -=- Node: Description
Up: Top -=- Next: Examples -=- Previous: Syntax


                     Description of GNN specific keywords

NDATa    Number of data points in the training set.

NPROd    Number of data points in the test set.

NPARa    Number of candidate descriptors.

UNIT     Unit number from which data are imported.

SEED     Seed for random number generator.

NDES     Number of descriptors for the neural network.

NHIDden  Number of nodes in the hidden layer.

NTARg    Number of target parameters to predict.

NSWEep   Number of sweeps training.

MU       Momentum constant.

ETA      Learning rate.

EXHAust  Exhuastive enumeration.

GFA      Genetic function approximation.

EP       Evolutionary programming.

NPOPu    Number of individual models in reproduction pool.

NGEN     Number of generations to reproduce.

FITNess  Average fitness of models in the reproduction pool before terminating
         genetic algorithms. Fitness is defined as the reciprocal of the
         residual rms error.



File: gnn -=- Node: Examples
Up: Top -=- Previous: Description -=- Next: References


                     Examples

[ Data File (represented here symbolically) ]

P1(1) P2(1) P3(1) P4(1) P5(1)
P1(2) P2(2) P3(2) P4(2) P5(2)
P1(3) P2(3) P3(3) P4(3) P5(3)
P1(4) P2(4) P3(4) P4(4) P5(4)
P1(5) P2(5) P3(5) P4(5) P5(5)

Note: Descriptors P1, P2 and P3, target parameters P4 and P5, training set
with data points (1), (2), and (3), and test set with data points (4) and (5).

[ Input ]

For example, file gnn.dat has 53 lines and 6 columns.

open read card unit 18 name gnn.dat

1. Exhaustive enumeration + Cross-validation + 1-Descriptor network

gnn ndata 53 nprod 0 npara 5 unit 18 seed 123 -
    ndes 1 nhidden 2 ntarg 1 nsweep 100 mu 0.5 eta 0.5 -
    exhaust

2. GFA + Test set residual rms error evaluated + 2-Descriptor network

gnn ndata 30 nprod 23 npara 4 unit 18 seed 123 -
    ndes 2 nhidden 3 ntarg 2 nsweep 100 mu 0.5 eta 0.5 -
    gfa npopu 2 ngen 10 fitness 5.0

Note: ndata + nprod = 53 (number of lines),
      npara + ntarg = 6 (number of columns).



File: gnn -=- Node: References
Up: Top -=- Previous: Examples -=- Next: Top


			     References

The GNN method was originally introduced in:

Sung-Sau So and Martin Karplus, Evolutionary optimization in quantitative
structure-activity relationship: An application of genetic neural networks,
J. Med. Chem., 39:1521-1530 (1996).

Sung-Sau So and Martin Karplus, Genetic Neural Networks for Quantitative
Structure-Activity Relationships:  Improvements and application of
benzodiazepine affinity for benzodiazepine/GABA_A receptors, J. Med. Chem., 
39:5246-5256 (1996).

Jie Hu and Aaron Dinner implemented the version in CHARMM.  It differs from
the HIPPO program of So and Karplus primarily in that steepest descents rather 
than scaled conjugate gradients optimization is used to train the neural 
networks.  Performance is comparable for the data in:

Jie Hu, Ao Ma, and Aaron R. Dinner, A two-step nucleotide flipping mechanism
enable kinetic discrimination of DNA lesions by AGT, Proc. Natl. Acad. Sci.
USA, in press (2008).

In addition to the above studies, papers using the CHARMM GNN method should
cite the introduction of the use of the GNN (and, more generally, informatic
methods) for determination of reaction coordinates:

Ao Ma and Aaron R. Dinner, Automatic method for identifying reaction
coordinates in complex systems, J. Phys. Chem. B, 109:6769-6779 (2005).

For a review of the GNN in other contexts, see:

Aaron R. Dinner, Sung-Sau So, and Martin Karplus, Statistical analysis of
protein folding kinetics, Adv. Chem. Phys., 120:1-34 (2002).

For a more detailed discussion of the neural network component used in the
CHARMM implementation, see the book:

Jure Zupan and Johann Gasteiger, Neural Networks for Chemists:  An
Introduction, VCH, New York (1993).


NIH Helix/Biowulf Systems
charmm.org Homepage

CHARMM Documentation / Rick_Venable@nih.gov