ROM Generation Tips

  • Typically, for physical problems, two to three layers are enough.
  • If the number of suggested epochs is about or greater than 100, overfitting might be an issue. If overfitting occurs, try using a smaller neural net. Though, if doing so reduces the accuracy, try instead to increase the regularization coefficient.
    Note: Poor generalization properties will more likely occur with a larger net compared to a smaller net.
  • A regularization coefficient can help to increase generalization properties (in particular, when deploying the tanh activation function).
  • Use the relu activation function on the first attempt to train/generate a romAI model for nonlinear problems.
  • With linear problems use the linear model type.
  • At each training run, neural net weights and biases are randomly initialized. Hence, when the proper neural net structure and activation function have been selected, consider re-training the model multiple times and save it in a different location. The optimization problem solved by romAI is a non-convex one (many local minima). Different initializations of the neural net weights might lead to different local minima.
  • Proper setting of the Early Stopping parameters can greatly reduce the training time, though this requires some familiarity with the model being trained and the model's loss curve.
  • Increasing the learning rate parameter can potentially speed-up the learning process.
  • If after many tries, the neural net appears to be inaccurate, the selection of inputs, state variables, or outputs might be wrong. When selecting inputs and state variables, consider that they should have a physically meaningful relation with the outputs.
  • When dealing with highly non-linear systems, the romAI approach can fail when facing problems with more than four states, however with linear problems, the number of states can be much higher (more than 20) even though the learning might require more epochs when increasing the number of states.
  • If signals have been filtered in the Pre-Processor tab before using them in the training, or if signals are too noisy, the R2 coefficients computed in the Viewer tab might not return reliable information.