/dports/misc/ncnn/ncnn-20211208/tools/pnnx/tests/ |
H A D | test_F_softplus.py | 24 x = F.softplus(x) 25 y = F.softplus(y, 2, 1.2) 26 z = F.softplus(z, -0.7, 15) 27 w = F.softplus(w, 0.1, 0.3)
|
H A D | test_F_mish.py | 20 return x * F.softplus(x).tanh() 23 return x.mul(torch.tanh(F.softplus(x)))
|
/dports/misc/ncnn/ncnn-20211208/tools/pnnx/src/pass_level1/ |
H A D | nn_Softplus.cpp | 36 const torch::jit::Node* softplus = find_node_by_kind(graph, "aten::softplus"); in write() local 38 op->params["beta"] = softplus->namedInput("beta"); in write() 39 op->params["threshold"] = softplus->namedInput("threshold"); in write()
|
/dports/science/py-chainer/chainer-7.8.0/chainer/functions/loss/ |
H A D | vae.py | 3 from chainer.functions.activation import softplus 114 loss = softplus.softplus(y) - x * y
|
/dports/science/py-chainer-chemistry/chainer-chemistry-0.7.1/chainer_chemistry/links/update/ |
H A D | cgcnn_update.py | 36 feat_core = functions.softplus(feat_core) 39 out = functions.softplus(site_feat + feat_sum)
|
/dports/misc/ncnn/ncnn-20211208/tools/pnnx/tests/ncnn/ |
H A D | test_F_mish.py | 20 return x * F.softplus(x).tanh() 23 return x.mul(torch.tanh(F.softplus(x)))
|
/dports/misc/mxnet/incubator-mxnet-1.9.0/python/mxnet/contrib/onnx/onnx2mx/ |
H A D | _import_helper.py | 24 from ._op_translations import softplus, shape, gather, lp_pooling, size 140 'Softplus' : softplus,
|
/dports/misc/py-mxnet/incubator-mxnet-1.9.0/python/mxnet/contrib/onnx/onnx2mx/ |
H A D | _import_helper.py | 24 from ._op_translations import softplus, shape, gather, lp_pooling, size 140 'Softplus' : softplus,
|
/dports/math/py-keras/Keras-2.4.3/tests/keras/ |
H A D | activations_test.py | 110 def softplus(x): function 114 f = K.function([x], [activations.softplus(x)]) 118 expected = softplus(test_values)
|
/dports/math/py-theano/Theano-1.0.5/theano/tensor/nnet/ |
H A D | sigm.py | 372 softplus = elemwise.Elemwise(scalar_softplus, name='softplus') variable 374 pprint.assign(softplus, printing.FunctionPrinter('softplus')) 385 (tensor.neg, (softplus, (tensor.neg, 'x'))), 411 (tensor.neg, (softplus, 'x')), 420 (softplus, 'x'), 427 (tensor.neg, (softplus, 'x')),
|
H A D | __init__.py | 25 from .sigm import (softplus, sigmoid, sigmoid_inplace,
|
/dports/math/py-jax/jax-0.2.9/jax/nn/ |
H A D | __init__.py | 41 softplus,
|
/dports/misc/py-onnx-tf/onnx-tf-1.6.0/onnx_tf/handlers/backend/ |
H A D | softplus.py | 9 @tf_func(tf.nn.softplus)
|
/dports/science/py-chainer-chemistry/chainer-chemistry-0.7.1/chainer_chemistry/functions/activation/ |
H A D | shifted_softplus.py | 20 functions.softplus(x, beta=beta))
|
/dports/math/py-flax/flax-0.3.3/flax/core/nn/ |
H A D | __init__.py | 24 softplus, swish, silu, tanh)
|
/dports/science/py-chainer-chemistry/chainer-chemistry-0.7.1/chainer_chemistry/links/readout/ |
H A D | cgcnn_readout.py | 22 h = functions.softplus(h)
|
/dports/math/py-flax/flax-0.3.3/flax/nn/ |
H A D | __init__.py | 21 softplus, swish, silu, tanh)
|
H A D | activation.py | 32 from jax.nn import softplus
|
/dports/math/py-keras/Keras-2.4.3/keras/ |
H A D | activations.py | 6 from tensorflow.keras.activations import softplus
|
/dports/math/py-flax/flax-0.3.3/flax/linen/ |
H A D | __init__.py | 22 softplus, swish, silu, tanh)
|
H A D | activation.py | 32 from jax.nn import softplus
|
/dports/math/py-jax/jax-0.2.9/jax/_src/nn/ |
H A D | functions.py | 45 def softplus(x: Array) -> Array: function 95 return -softplus(-x)
|
/dports/science/py-chainer-chemistry/chainer-chemistry-0.7.1/chainer_chemistry/links/scaler/ |
H A D | flow_scaler.py | 66 return chainer.functions.softplus(self.W1_) 70 return chainer.functions.softplus(self.W2_)
|
/dports/math/py-pymc3/pymc-3.11.4/pymc3/distributions/ |
H A D | transforms.py | 234 return tt.nnet.softplus(x) 248 return -tt.nnet.softplus(-x) 297 s = tt.nnet.softplus(-x)
|
/dports/math/R-cran-recipes/recipes/man/ |
H A D | step_relu.Rd | 40 \item{smooth}{A logical indicating if the softplus function, a smooth 65 will apply the rectified linear or softplus transformations to numeric
|