# brainpy.dyn.synapses.AMPA#

class brainpy.dyn.synapses.AMPA(pre, post, conn, output=COBA, stp=None, comp_method='dense', g_max=0.42, delay_step=None, alpha=0.98, beta=0.18, T=0.5, T_duration=0.5, method='exp_auto', name=None, mode=NormalMode, stop_spike_gradient=False)[source]#

AMPA synapse model.

Model Descriptions

AMPA receptor is an ionotropic receptor, which is an ion channel. When it is bound by neurotransmitters, it will immediately open the ion channel, causing the change of membrane potential of postsynaptic neurons.

A classical model is to use the Markov process to model ion channel switch. Here $$g$$ represents the probability of channel opening, $$1-g$$ represents the probability of ion channel closing, and $$\alpha$$ and $$\beta$$ are the transition probability. Because neurotransmitters can open ion channels, the transfer probability from $$1-g$$ to $$g$$ is affected by the concentration of neurotransmitters. We denote the concentration of neurotransmitters as $$[T]$$ and get the following Markov process.

We obtained the following formula when describing the process by a differential equation.

$\frac{ds}{dt} =\alpha[T](1-g)-\beta g$

where $$\alpha [T]$$ denotes the transition probability from state $$(1-g)$$ to state $$(g)$$; and $$\beta$$ represents the transition probability of the other direction. $$\alpha$$ is the binding constant. $$\beta$$ is the unbinding constant. $$[T]$$ is the neurotransmitter concentration, and has the duration of 0.5 ms.

Moreover, the post-synaptic current on the post-synaptic neuron is formulated as

$I_{syn} = g_{max} g (V-E)$

where $$g_{max}$$ is the maximum conductance, and E is the reverse potential.

Model Examples

>>> import brainpy as bp
>>> from brainpy.dyn import neurons, synapses
>>> import matplotlib.pyplot as plt
>>>
>>> neu1 = neurons.HH(1)
>>> neu2 = neurons.HH(1)
>>> syn1 = synapses.AMPA(neu1, neu2, bp.connect.All2All())
>>> net = bp.dyn.Network(pre=neu1, syn=syn1, post=neu2)
>>>
>>> runner = bp.dyn.DSRunner(net, inputs=[('pre.input', 5.)], monitors=['pre.V', 'post.V', 'syn.g'])
>>> runner.run(150.)
>>>
>>> fig, gs = bp.visualize.get_figure(2, 1, 3, 8)
>>> plt.plot(runner.mon.ts, runner.mon['pre.V'], label='pre-V')
>>> plt.plot(runner.mon.ts, runner.mon['post.V'], label='post-V')
>>> plt.legend()
>>>
>>> plt.plot(runner.mon.ts, runner.mon['syn.g'], label='g')
>>> plt.legend()
>>> plt.show()
Parameters
• pre (NeuGroup) – The pre-synaptic neuron group.

• post (NeuGroup) – The post-synaptic neuron group.

• conn (optional, ndarray, JaxArray, dict of (str, ndarray), TwoEndConnector) – The synaptic connections.

• comp_method (str) – The connection type used for model speed optimization. It can be sparse and dense. The default is dense.

• delay_step (int, ndarray, JaxArray, Initializer, Callable) – The delay length. It should be the value of $$\mathrm{delay\_time / dt}$$.

• E (float, JaxArray, ndarray) –

The reversal potential for the synaptic current. [mV]

Deprecated since version 2.1.13: E is deprecated in AMPA model. Please define E with brainpy.dyn.synouts.COBA. This parameter will be removed since 2.2.0

• g_max (float, ndarray, JaxArray, Initializer, Callable) – The synaptic strength (the maximum conductance). Default is 1.

• alpha (float, JaxArray, ndarray) – Binding constant.

• beta (float, JaxArray, ndarray) – Unbinding constant.

• T (float, JaxArray, ndarray) – Transmitter concentration when synapse is triggered by a pre-synaptic spike.. Default 1 [mM].

• T_duration (float, JaxArray, ndarray) – Transmitter concentration duration time after being triggered. Default 1 [ms]

• name (str) – The name of this synaptic projection.

• method (str) – The numerical integration methods.

References

1

Vijayan S, Kopell N J. Thalamic model of awake alpha oscillations and implications for stimulus processing[J]. Proceedings of the National Academy of Sciences, 2012, 109(45): 18553-18558.

__init__(pre, post, conn, output=COBA, stp=None, comp_method='dense', g_max=0.42, delay_step=None, alpha=0.98, beta=0.18, T=0.5, T_duration=0.5, method='exp_auto', name=None, mode=NormalMode, stop_spike_gradient=False)[source]#

Methods

 __init__(pre, post, conn[, output, stp, ...]) check_post_attrs(*attrs) Check whether post group satisfies the requirement. check_pre_attrs(*attrs) Check whether pre group satisfies the requirement. clear_input() dg(g, t, TT) get_delay_data(identifier, delay_step, *indices) Get delay data according to the provided delay steps. init_weights(weight, comp_method[, sparse_data]) rtype Union[float, TypeVar(Array, JaxArray, Variable, TrainVar, Array, ndarray)] load_states(filename[, verbose]) Load the model states. nodes([method, level, include_self]) Collect all children nodes. offline_fit(target, fit_record) offline_init() online_fit(target, fit_record) online_init() register_delay(identifier, delay_step, ...) Register delay variable. register_implicit_nodes(*nodes, **named_nodes) register_implicit_vars(*variables, ...) reset([batch_size]) Reset function which reset the whole variables in the model. reset_local_delays([nodes]) Reset local delay variables. reset_state([batch_size]) Reset function which reset the states in the model. save_states(filename[, variables]) Save the model states. syn2post_with_all2all(syn_value, syn_weight) syn2post_with_dense(syn_value, syn_weight, ...) syn2post_with_one2one(syn_value, syn_weight) train_vars([method, level, include_self]) The shortcut for retrieving all trainable variables. unique_name([name, type_]) Get the unique name for this object. update(tdi[, pre_spike]) The function to specify the updating rule. update_local_delays([nodes]) Update local delay variables. vars([method, level, include_self]) Collect all variables in this node and the children nodes.

Attributes

 global_delay_data mode Mode of the model, which is useful to control the multiple behaviors of the model. name Name of the model.