brainpy.neurons.ExpIF#
- class brainpy.neurons.ExpIF(*args, input_var=True, noise=None, spike_fun=None, **kwargs)[source]#
Exponential integrate-and-fire neuron model.
Model Descriptions
In the exponential integrate-and-fire model [1], the differential equation for the membrane potential is given by
\[\begin{split}\tau\frac{d V}{d t}= - (V-V_{rest}) + \Delta_T e^{\frac{V-V_T}{\Delta_T}} + RI(t), \\ \text{after} \, V(t) \gt V_{th}, V(t) = V_{reset} \, \text{last} \, \tau_{ref} \, \text{ms}\end{split}\]This equation has an exponential nonlinearity with “sharpness” parameter \(\Delta_{T}\) and “threshold” \(\vartheta_{rh}\).
The moment when the membrane potential reaches the numerical threshold \(V_{th}\) defines the firing time \(t^{(f)}\). After firing, the membrane potential is reset to \(V_{rest}\) and integration restarts at time \(t^{(f)}+\tau_{\rm ref}\), where \(\tau_{\rm ref}\) is an absolute refractory time. If the numerical threshold is chosen sufficiently high, \(V_{th}\gg v+\Delta_T\), its exact value does not play any role. The reason is that the upswing of the action potential for \(v\gg v +\Delta_{T}\) is so rapid, that it goes to infinity in an incredibly short time. The threshold \(V_{th}\) is introduced mainly for numerical convenience. For a formal mathematical analysis of the model, the threshold can be pushed to infinity.
The model was first introduced by Nicolas Fourcaud-Trocmé, David Hansel, Carl van Vreeswijk and Nicolas Brunel [1]. The exponential nonlinearity was later confirmed by Badel et al. [3]. It is one of the prominent examples of a precise theoretical prediction in computational neuroscience that was later confirmed by experimental neuroscience.
Two important remarks:
(i) The right-hand side of the above equation contains a nonlinearity that can be directly extracted from experimental data [3]. In this sense the exponential nonlinearity is not an arbitrary choice but directly supported by experimental evidence.
(ii) Even though it is a nonlinear model, it is simple enough to calculate the firing rate for constant input, and the linear response to fluctuations, even in the presence of input noise [4].
Model Examples
>>> import brainpy as bp >>> group = bp.neurons.ExpIF(1) >>> runner = bp.DSRunner(group, monitors=['V'], inputs=('input', 10.)) >>> runner.run(300., ) >>> bp.visualize.line_plot(runner.mon.ts, runner.mon.V, ylabel='V', show=True)
Model Parameters
Parameter
Init Value
Unit
Explanation
V_rest
-65
mV
Resting potential.
V_reset
-68
mV
Reset potential after spike.
V_th
-30
mV
Threshold potential of spike.
V_T
-59.9
mV
Threshold potential of generating action potential.
delta_T
3.48
Spike slope factor.
R
1
Membrane resistance.
tau
10
Membrane time constant. Compute by R * C.
tau_ref
1.7
Refractory period length.
Model Variables
Variables name
Initial Value
Explanation
V
0
Membrane potential.
input
0
External and synaptic input current.
spike
False
Flag to mark whether the neuron is spiking.
refractory
False
Flag to mark whether the neuron is in refractory period.
t_last_spike
-1e7
Last spike time stamp.
References
Methods
__init__
(*args[, input_var, noise, spike_fun])add_aft_update
(key, fun)Add the after update into this node
add_bef_update
(key, fun)Add the before update into this node
add_inp_fun
(key, fun[, label, category])Add an input function.
clear_input
()Empty function of clearing inputs.
cpu
()Move all variable into the CPU device.
cuda
()Move all variables into the GPU device.
derivative
(V, t, I)get_aft_update
(key)Get the after update of this node by the given
key
.get_batch_shape
([batch_size])get_bef_update
(key)Get the before update of this node by the given
key
.get_delay_data
(identifier, delay_pos, *indices)Get delay data according to the provided delay steps.
get_delay_var
(name)get_inp_fun
(key)Get the input function.
get_local_delay
(var_name, delay_name)Get the delay at the given identifier (name).
has_aft_update
(key)Whether this node has the after update of the given
key
.has_bef_update
(key)Whether this node has the before update of the given
key
.init_param
(param[, shape, sharding])Initialize parameters.
init_variable
(var_data, batch_or_mode[, ...])Initialize variables.
inv_scaling
(x[, scale])jit_step_run
(i, *args, **kwargs)The jitted step function for running.
load_state
(state_dict, **kwargs)Load states from a dictionary.
load_state_dict
(state_dict[, warn, compatible])Copy parameters and buffers from
state_dict
into this module and its descendants.nodes
([method, level, include_self])Collect all children nodes.
offset_scaling
(x[, bias, scale])register_delay
(identifier, delay_step, ...)Register delay variable.
register_implicit_nodes
(*nodes[, node_cls])register_implicit_vars
(*variables[, var_cls])register_local_delay
(var_name, delay_name[, ...])Register local relay at the given delay time.
reset
(*args, **kwargs)Reset function which reset the whole variables in the model (including its children models).
reset_local_delays
([nodes])Reset local delay variables.
reset_state
([batch_size])return_info
()save_state
(**kwargs)Save states as a dictionary.
setattr
(key, value)- rtype:
state_dict
(**kwargs)Returns a dictionary containing a whole state of the module.
std_scaling
(x[, scale])step_run
(i, *args, **kwargs)The step run function.
sum_current_inputs
(*args[, init, label])Summarize all current inputs by the defined input functions
.current_inputs
.sum_delta_inputs
(*args[, init, label])Summarize all delta inputs by the defined input functions
.delta_inputs
.sum_inputs
(*args, **kwargs)to
(device)Moves all variables into the given device.
tpu
()Move all variables into the TPU device.
tracing_variable
(name, init, shape[, ...])Initialize the variable which can be traced during computations and transformations.
train_vars
([method, level, include_self])The shortcut for retrieving all trainable variables.
tree_flatten
()Flattens the object as a PyTree.
tree_unflatten
(aux, dynamic_values)Unflatten the data to construct an object of this class.
unique_name
([name, type_])Get the unique name for this object.
update
([x])The function to specify the updating rule.
update_local_delays
([nodes])Update local delay variables.
vars
([method, level, include_self, ...])Collect all variables in this node and the children nodes.
Attributes
after_updates
before_updates
cur_inputs
current_inputs
delta_inputs
implicit_nodes
implicit_vars
mode
Mode of the model, which is useful to control the multiple behaviors of the model.
name
Name of the model.
spk_dtype
supported_modes
Supported computing modes.
varshape
The shape of variables in the neuron group.