To whom it may concern,
I am writing regarding the problem with giving my topology network noise
after simulation with the signal. The network is oscillating with noise
generator which is created as single layer which provides noise for other
nodes (which consist of layers). I stimulate my network with stronger
signal(another different layer) and after the simulation the Nodes stop
reacting to the noise input and become stuck on refresh Voltage and its
for longer periods than refractory period.
The model is derivation of Hill-Tononi example network.
The Nest is running on Ubuntu 20.04 , Conda/Spyder The nest is version
To visualize the problem here are the images which illustrate the problem
Here is the snippet of the code with the connections and layer setups
Bc. Filip Blaštík