Home
last modified time | relevance | path

Searched refs:softplus (Results 1 – 25 of 62) sorted by relevance

123

/dports/misc/ncnn/ncnn-20211208/tools/pnnx/tests/
H A Dtest_F_softplus.py24 x = F.softplus(x)
25 y = F.softplus(y, 2, 1.2)
26 z = F.softplus(z, -0.7, 15)
27 w = F.softplus(w, 0.1, 0.3)
H A Dtest_F_mish.py20 return x * F.softplus(x).tanh()
23 return x.mul(torch.tanh(F.softplus(x)))
/dports/misc/ncnn/ncnn-20211208/tools/pnnx/src/pass_level1/
H A Dnn_Softplus.cpp36 const torch::jit::Node* softplus = find_node_by_kind(graph, "aten::softplus"); in write() local
38 op->params["beta"] = softplus->namedInput("beta"); in write()
39 op->params["threshold"] = softplus->namedInput("threshold"); in write()
/dports/science/py-chainer/chainer-7.8.0/chainer/functions/loss/
H A Dvae.py3 from chainer.functions.activation import softplus
114 loss = softplus.softplus(y) - x * y
/dports/science/py-chainer-chemistry/chainer-chemistry-0.7.1/chainer_chemistry/links/update/
H A Dcgcnn_update.py36 feat_core = functions.softplus(feat_core)
39 out = functions.softplus(site_feat + feat_sum)
/dports/misc/ncnn/ncnn-20211208/tools/pnnx/tests/ncnn/
H A Dtest_F_mish.py20 return x * F.softplus(x).tanh()
23 return x.mul(torch.tanh(F.softplus(x)))
/dports/misc/mxnet/incubator-mxnet-1.9.0/python/mxnet/contrib/onnx/onnx2mx/
H A D_import_helper.py24 from ._op_translations import softplus, shape, gather, lp_pooling, size
140 'Softplus' : softplus,
/dports/misc/py-mxnet/incubator-mxnet-1.9.0/python/mxnet/contrib/onnx/onnx2mx/
H A D_import_helper.py24 from ._op_translations import softplus, shape, gather, lp_pooling, size
140 'Softplus' : softplus,
/dports/math/py-keras/Keras-2.4.3/tests/keras/
H A Dactivations_test.py110 def softplus(x): function
114 f = K.function([x], [activations.softplus(x)])
118 expected = softplus(test_values)
/dports/math/py-theano/Theano-1.0.5/theano/tensor/nnet/
H A Dsigm.py372 softplus = elemwise.Elemwise(scalar_softplus, name='softplus') variable
374 pprint.assign(softplus, printing.FunctionPrinter('softplus'))
385 (tensor.neg, (softplus, (tensor.neg, 'x'))),
411 (tensor.neg, (softplus, 'x')),
420 (softplus, 'x'),
427 (tensor.neg, (softplus, 'x')),
H A D__init__.py25 from .sigm import (softplus, sigmoid, sigmoid_inplace,
/dports/math/py-jax/jax-0.2.9/jax/nn/
H A D__init__.py41 softplus,
/dports/misc/py-onnx-tf/onnx-tf-1.6.0/onnx_tf/handlers/backend/
H A Dsoftplus.py9 @tf_func(tf.nn.softplus)
/dports/science/py-chainer-chemistry/chainer-chemistry-0.7.1/chainer_chemistry/functions/activation/
H A Dshifted_softplus.py20 functions.softplus(x, beta=beta))
/dports/math/py-flax/flax-0.3.3/flax/core/nn/
H A D__init__.py24 softplus, swish, silu, tanh)
/dports/science/py-chainer-chemistry/chainer-chemistry-0.7.1/chainer_chemistry/links/readout/
H A Dcgcnn_readout.py22 h = functions.softplus(h)
/dports/math/py-flax/flax-0.3.3/flax/nn/
H A D__init__.py21 softplus, swish, silu, tanh)
H A Dactivation.py32 from jax.nn import softplus
/dports/math/py-keras/Keras-2.4.3/keras/
H A Dactivations.py6 from tensorflow.keras.activations import softplus
/dports/math/py-flax/flax-0.3.3/flax/linen/
H A D__init__.py22 softplus, swish, silu, tanh)
H A Dactivation.py32 from jax.nn import softplus
/dports/math/py-jax/jax-0.2.9/jax/_src/nn/
H A Dfunctions.py45 def softplus(x: Array) -> Array: function
95 return -softplus(-x)
/dports/science/py-chainer-chemistry/chainer-chemistry-0.7.1/chainer_chemistry/links/scaler/
H A Dflow_scaler.py66 return chainer.functions.softplus(self.W1_)
70 return chainer.functions.softplus(self.W2_)
/dports/math/py-pymc3/pymc-3.11.4/pymc3/distributions/
H A Dtransforms.py234 return tt.nnet.softplus(x)
248 return -tt.nnet.softplus(-x)
297 s = tt.nnet.softplus(-x)
/dports/math/R-cran-recipes/recipes/man/
H A Dstep_relu.Rd40 \item{smooth}{A logical indicating if the softplus function, a smooth
65 will apply the rectified linear or softplus transformations to numeric

123