Method List
-
#<=> Torch::Tensor
-
#[] Torch::Utils::Data::TensorDataset
-
#[] Torch::NN::ParameterList
-
#[] Torch::Tensor
-
#[] Torch::Utils::Data::Subset
-
#[] Torch::NN::ModuleList
-
#[]= Torch::Tensor
-
#_activation_fn Torch::NN::Utils
-
#_apply Torch::NN::Module
-
#_apply Torch::NN::RNNBase
-
_calculate_fan_in_and_fan_out Torch::NN::Init
-
#_check_input_dim Torch::NN::InstanceNorm2d
-
#_check_input_dim Torch::NN::BatchNorm3d
-
#_check_input_dim Torch::NN::InstanceNorm3d
-
#_check_input_dim Torch::NN::InstanceNorm1d
-
#_check_input_dim Torch::NN::BatchNorm1d
-
#_check_input_dim Torch::NN::BatchNorm2d
-
#_clones Torch::NN::Utils
-
_dtype_to_numo Torch
-
_from_blob_ref Torch
-
_make_tensor_class Torch
-
#_ntuple Torch::NN::Utils
-
#_pair Torch::NN::Utils
-
#_quadrupal Torch::NN::Utils
-
#_single Torch::NN::Utils
-
#_triple Torch::NN::Utils
-
adaptive_avg_pool1d Torch::NN::Functional
-
adaptive_avg_pool2d Torch::NN::Functional
-
adaptive_avg_pool3d Torch::NN::Functional
-
adaptive_max_pool1d Torch::NN::Functional
-
adaptive_max_pool2d Torch::NN::Functional
-
adaptive_max_pool3d Torch::NN::Functional
-
#add_module Torch::NN::Module
-
#add_param_group Torch::Optim::Optimizer
-
alpha_dropout Torch::NN::Functional
-
#append Torch::NN::ModuleList
-
#apply Torch::NN::Module
-
avg_pool1d Torch::NN::Functional
-
avg_pool2d Torch::NN::Functional
-
avg_pool3d Torch::NN::Functional
-
#batch_first? Torch::NN::MultiheadAttention
-
#batch_first? Torch::NN::Transformer
-
batch_norm Torch::NN::Functional
-
bilinear Torch::NN::Functional
-
binary_cross_entropy Torch::NN::Functional
-
binary_cross_entropy_with_logits Torch::NN::Functional
-
#buffers Torch::NN::Module
-
#bytes Torch::ByteStorage
-
calculate_gain Torch::NN::Init
-
#call Torch::NN::Module
-
#check_forward_args Torch::NN::LSTM
-
#children Torch::NN::Module
-
#close Torch::Utils::Data::DataPipes::Iter::StreamWrapper
-
close_streams Torch::Utils::Data::DataPipes::Iter::StreamWrapper
-
#coerce Torch::Tensor
-
#concat Torch::NN::ModuleList
-
#concat Torch::NN::ParameterList
-
conv1d Torch::NN::Functional
-
conv2d Torch::NN::Functional
-
conv3d Torch::NN::Functional
-
cosine_embedding_loss Torch::NN::Functional
-
cosine_similarity Torch::NN::Functional
-
#cpu Torch::Tensor
-
#cpu Torch::NN::Module
-
cross_entropy Torch::NN::Functional
-
ctc_loss Torch::NN::Functional
-
#cuda Torch::Tensor
-
#cuda Torch::NN::Module
-
#d_model Torch::NN::Transformer
-
#dataset Torch::Utils::Data::DataLoader
-
#decoder Torch::NN::Transformer
-
#deep_dup Torch::NN::Module
-
device Torch
-
#dilation Torch::NN::ConvNd
-
#double Torch::NN::Module
-
download_url_to_file Torch::Hub
-
dropout Torch::NN::Functional
-
dropout2d Torch::NN::Functional
-
dropout3d Torch::NN::Functional
-
#dtype Torch::Tensor
-
#dup Torch::NN::Parameter
-
#dup Torch::Tensor
-
#each Torch::Utils::Data::DataPipes::Iter::IterableWrapper
-
#each Torch::NN::ParameterList
-
#each Torch::Utils::Data::DataPipes::Iter::FileLister
-
#each Torch::Tensor
-
#each Torch::Utils::Data::DataPipes::FilterIterDataPipe
-
#each Torch::Utils::Data::DataPipes::Iter::FileOpener
-
#each Torch::Utils::Data::DataPipes::IterDataPipe
-
#each Torch::Utils::Data::DataLoader
-
#each Torch::NN::ModuleList
-
#each Torch::Utils::Data::IterableDataset
-
elu Torch::NN::Functional
-
embedding Torch::NN::Functional
-
embedding_bag Torch::NN::Functional
-
enable_grad Torch
-
#encoder Torch::NN::Transformer
-
#eval Torch::NN::Module
-
#extra_inspect Torch::NN::PReLU
-
#extra_inspect Torch::NN::ReLU
-
#extra_inspect Torch::NN::GroupNorm
-
#extra_inspect Torch::NN::ReflectionPadNd
-
#extra_inspect Torch::NN::LayerNorm
-
#extra_inspect Torch::NN::Hardshrink
-
#extra_inspect Torch::NN::MaxPoolNd
-
#extra_inspect Torch::NN::Conv2d
-
#extra_inspect Torch::NN::Softmax
-
#extra_inspect Torch::NN::LeakyReLU
-
#extra_inspect Torch::NN::LPPoolNd
-
#extra_inspect Torch::NN::GELU
-
#extra_inspect Torch::NN::Softshrink
-
#extra_inspect Torch::NN::ReplicationPadNd
-
#extra_inspect Torch::NN::ConstantPadNd
-
#extra_inspect Torch::NN::RNNBase
-
#extra_inspect Torch::NN::Bilinear
-
#extra_inspect Torch::NN::DropoutNd
-
#extra_inspect Torch::NN::ELU
-
#extra_inspect Torch::NN::Upsample
-
#extra_inspect Torch::NN::MaxUnpoolNd
-
#extra_inspect Torch::NN::AdaptiveAvgPoolNd
-
#extra_inspect Torch::NN::AvgPoolNd
-
#extra_inspect Torch::NN::Softplus
-
#extra_inspect Torch::NN::Linear
-
#extra_inspect Torch::NN::BatchNorm
-
#extra_inspect Torch::NN::AdaptiveMaxPoolNd
-
#extra_inspect Torch::NN::LocalResponseNorm
-
feature_alpha_dropout Torch::NN::Functional
-
#flatten_parameters Torch::NN::RNNBase
-
#float Torch::NN::Module
-
fold Torch::NN::Functional
-
#format Torch::Inspector::Formatter
-
#forward Torch::NN::Conv3d
-
#forward Torch::NN::Conv1d
-
#forward Torch::NN::LayerNorm
-
#forward Torch::NN::Conv2d
-
#forward Torch::NN::FeatureAlphaDropout
-
#forward Torch::NN::ReplicationPadNd
-
#forward Torch::NN::PReLU
-
#forward Torch::NN::ReLU
-
#forward Torch::NN::LSTM
-
#forward Torch::NN::MaxPool3d
-
#forward Torch::NN::Hardshrink
-
#forward Torch::NN::Tanh
-
#forward Torch::NN::EmbeddingBag
-
#forward Torch::NN::AdaptiveMaxPool1d
-
#forward Torch::NN::HingeEmbeddingLoss
-
#forward Torch::NN::PoissonNLLLoss
-
#forward Torch::NN::TransformerEncoderLayer
-
#forward Torch::NN::TransformerEncoder
-
#forward Torch::NN::SoftMarginLoss
-
#forward Torch::NN::AdaptiveAvgPool3d
-
#forward Torch::NN::AdaptiveMaxPool2d
-
#forward Torch::NN::MultiheadAttention
-
#forward Torch::NN::GroupNorm
-
#forward Torch::NN::Transformer
-
#forward Torch::NN::MaxUnpool3d
-
#forward Torch::NN::Module
-
#forward Torch::NN::AlphaDropout
-
#forward Torch::NN::Linear
-
#forward Torch::NN::TransformerDecoder
-
#forward Torch::NN::MaxUnpool2d
-
#forward Torch::NN::AdaptiveAvgPool1d
-
#forward Torch::NN::BCELoss
-
#forward Torch::NN::BatchNorm
-
#forward Torch::NN::Dropout
-
#forward Torch::NN::L1Loss
-
#forward Torch::NN::ConstantPadNd
-
#forward Torch::NN::MultiMarginLoss
-
#forward Torch::NN::MarginRankingLoss
-
#forward Torch::NN::Unfold
-
#forward Torch::NN::Sigmoid
-
#forward Torch::NN::ReflectionPadNd
-
#forward Torch::NN::Softmax
-
#forward Torch::NN::Softmin
-
#forward Torch::NN::AdaptiveMaxPool3d
-
#forward Torch::NN::AvgPool2d
-
#forward Torch::NN::CosineEmbeddingLoss
-
#forward Torch::NN::GELU
-
#forward Torch::NN::Fold
-
#forward Torch::NN::Softshrink
-
#forward Torch::NN::LeakyReLU
-
#forward Torch::NN::MaxPool1d
-
#forward Torch::NN::MaxPool2d
-
#forward Torch::NN::MultiLabelMarginLoss
-
#forward Torch::NN::CrossEntropyLoss
-
#forward Torch::NN::Sequential
-
#forward Torch::NN::PairwiseDistance
-
#forward Torch::NN::GRU
-
#forward Torch::NN::Tanhshrink
-
#forward Torch::NN::TransformerDecoderLayer
-
#forward Torch::NN::AvgPool3d
-
#forward Torch::NN::ELU
-
#forward Torch::NN::LogSigmoid
-
#forward Torch::NN::LogSoftmax
-
#forward Torch::NN::KLDivLoss
-
#forward Torch::NN::LocalResponseNorm
-
#forward Torch::NN::InstanceNorm
-
#forward Torch::NN::AvgPool1d
-
#forward Torch::NN::Softmax2d
-
#forward Torch::NN::Softplus
-
#forward Torch::NN::Softsign
-
#forward Torch::NN::RNNBase
-
#forward Torch::NN::BCEWithLogitsLoss
-
#forward Torch::NN::NLLLoss
-
#forward Torch::NN::MSELoss
-
#forward Torch::NN::Identity
-
#forward Torch::NN::CTCLoss
-
#forward Torch::NN::AdaptiveAvgPool2d
-
#forward Torch::NN::Dropout3d
-
#forward Torch::NN::Dropout2d
-
#forward Torch::NN::LPPool2d
-
#forward Torch::NN::SmoothL1Loss
-
#forward Torch::NN::Embedding
-
#forward Torch::NN::LPPool1d
-
#forward Torch::NN::TripletMarginLoss
-
#forward Torch::NN::MultiLabelSoftMarginLoss
-
#forward Torch::NN::CosineSimilarity
-
#forward Torch::NN::Bilinear
-
#forward Torch::NN::Upsample
-
#forward Torch::NN::MaxUnpool1d
-
#forward_impl Torch::NN::LSTM
-
#forward_impl Torch::NN::GRU
-
#forward_tensor Torch::NN::LSTM
-
#forward_tensor Torch::NN::GRU
-
from_buffer Torch::ByteStorage
-
from_numo Torch
-
functional_datapipe Torch::Utils::Data::DataPipes::IterDataPipe
-
functions Torch::Utils::Data::DataPipes::IterDataPipe
-
gelu Torch::NN::Functional
-
#generate_square_subsequent_mask Torch::NN::Transformer
-
#get_lr Torch::Optim::LRScheduler::ExponentialLR
-
#get_lr Torch::Optim::LRScheduler::CosineAnnealingLR
-
#get_lr Torch::Optim::LRScheduler::LambdaLR
-
#get_lr Torch::Optim::LRScheduler::MultiplicativeLR
-
#get_lr Torch::Optim::LRScheduler::MultiStepLR
-
#get_lr Torch::Optim::LRScheduler::StepLR
-
#gets Torch::Utils::Data::DataPipes::Iter::StreamWrapper
-
grad_enabled Torch
-
group_norm Torch::NN::Functional
-
#groups Torch::NN::ConvNd
-
#half Torch::NN::Module
-
hardshrink Torch::NN::Functional
-
hinge_embedding_loss Torch::NN::Functional
-
#imag Torch::Tensor
-
#in_channels Torch::NN::ConvNd
-
#in_features Torch::NN::Linear
-
in_projection Torch::NN::Functional
-
in_projection_packed Torch::NN::Functional
-
#initialize Torch::Optim::ASGD
-
#initialize Torch::Optim::Adam
-
#initialize Torch::Optim::LRScheduler::LRScheduler
-
#initialize Torch::NN::Softmin
-
#initialize Torch::NN::AvgPool2d
-
#initialize Torch::NN::AvgPool3d
-
#initialize Torch::NN::AvgPool1d
-
#initialize Torch::Optim::Adamax
-
#initialize Torch::NN::BatchNorm
-
#initialize Torch::NN::TransformerEncoder
-
#initialize Torch::NN::LocalResponseNorm
-
#initialize Torch::NN::MultiLabelSoftMarginLoss
-
#initialize Torch::Optim::Adadelta
-
#initialize Torch::NN::Embedding
-
#initialize Torch::NN::DropoutNd
-
#initialize Torch::Optim::Rprop
-
#initialize Torch::Optim::AdamW
-
#initialize Torch::NN::LPPoolNd
-
#initialize Torch::NN::Softplus
-
#initialize Torch::NN::Upsample
-
#initialize Torch::Optim::LRScheduler::CosineAnnealingLR
-
#initialize Torch::NN::RNNBase
-
#initialize Torch::NN::NLLLoss
-
#initialize Torch::NN::Identity
-
#initialize Torch::NN::Bilinear
-
#initialize Torch::NN::CTCLoss
-
#initialize Torch::NN::MSELoss
-
#initialize Torch::Utils::Data::DataPipes::Iter::IterableWrapper
-
#initialize Torch::NN::ConstantPadNd
-
#initialize Torch::NN::ConstantPad2d
-
#initialize Torch::Optim::LRScheduler::StepLR
-
#initialize Torch::NN::WeightedLoss
-
#initialize Torch::NN::TransformerDecoderLayer
-
#initialize Torch::NN::ParameterList
-
#initialize Torch::NN::MaxUnpool3d
-
#initialize Torch::NN::SmoothL1Loss
-
#initialize Torch::NN::TransformerEncoderLayer
-
#initialize Torch::NN::TripletMarginLoss
-
#initialize Torch::NN::Softmax
-
#initialize Torch::Optim::LRScheduler::MultiStepLR
-
#initialize Torch::NN::L1Loss
-
#initialize Torch::Optim::SGD
-
#initialize Torch::NN::Unfold
-
#initialize Torch::NN::BCELoss
-
#initialize Torch::Optim::LRScheduler::MultiplicativeLR
-
#initialize Torch::NN::MarginRankingLoss
-
#initialize Torch::NN::ReplicationPad3d
-
#initialize Torch::NN::SoftMarginLoss
-
#initialize Torch::Utils::Data::DataPipes::Iter::FileLister
-
#initialize Torch::Utils::Data::DataPipes::Iter::StreamWrapper
-
#initialize Torch::NN::ReflectionPad2d
-
#initialize Torch::NN::ReflectionPad1d
-
#initialize Torch::NN::MultiheadAttention
-
#initialize Torch::NN::Linear
-
#initialize Torch::NN::CrossEntropyLoss
-
#initialize Torch::NN::ConvNd
-
#initialize Torch::NN::Module
-
#initialize Torch::Optim::LRScheduler::ExponentialLR
-
#initialize Torch::NN::TransformerDecoder
-
#initialize Torch::NN::AdaptiveAvgPoolNd
-
#initialize Torch::NN::Conv3d
-
#initialize Torch::Utils::Data::DataPipes::FilterIterDataPipe
-
#initialize Torch::NN::Conv1d
-
#initialize Torch::NN::Conv2d
-
#initialize Torch::Inspector::Formatter
-
#initialize Torch::NN::EmbeddingBag
-
#initialize Torch::NN::PReLU
-
#initialize Torch::NN::ReLU
-
#initialize Torch::NN::LSTM
-
#initialize Torch::NN::Loss
-
#initialize Torch::Optim::LRScheduler::LambdaLR
-
#initialize Torch::NN::CosineEmbeddingLoss
-
#initialize Torch::NN::Fold
-
#initialize Torch::NN::GELU
-
#initialize Torch::NN::ConstantPad1d
-
#initialize Torch::Utils::Data::DataLoader
-
#initialize Torch::NN::HingeEmbeddingLoss
-
#initialize Torch::NN::ReplicationPad1d
-
#initialize Torch::Optim::Optimizer
-
#initialize Torch::Utils::Data::DataPipes::Iter::FileOpener
-
#initialize Torch::Utils::Data::TensorDataset
-
#initialize Torch::NN::RNN
-
#initialize Torch::NN::ConstantPad3d
-
#initialize Torch::NN::ELU
-
#initialize Torch::NN::GRU
-
#initialize Torch::NN::CosineSimilarity
-
#initialize Torch::NN::PairwiseDistance
-
#initialize Torch::NN::MultiMarginLoss
-
#initialize Torch::ByteStorage
-
#initialize Torch::NN::PoissonNLLLoss
-
#initialize Torch::NN::InstanceNorm
-
#initialize Torch::NN::MaxUnpool2d
-
#initialize Torch::NN::MaxPoolNd
-
#initialize Torch::NN::LeakyReLU
-
#initialize Torch::NN::LayerNorm
-
#initialize Torch::NN::Hardshrink
-
#initialize Torch::NN::MultiLabelMarginLoss
-
#initialize Torch::NN::KLDivLoss
-
#initialize Torch::Optim::RMSprop
-
#initialize Torch::Optim::Adagrad
-
#initialize Torch::NN::ZeroPad2d
-
#initialize Torch::NN::LogSoftmax
-
#initialize Torch::NN::Transformer
-
#initialize Torch::NN::ModuleList
-
#initialize Torch::NN::Softshrink
-
#initialize Torch::NN::GroupNorm
-
#initialize Torch::NN::BCEWithLogitsLoss
-
#initialize Torch::NN::Sequential
-
#initialize Torch::NN::AdaptiveMaxPoolNd
-
#initialize Torch::NN::ReplicationPad2d
-
#initialize Torch::Utils::Data::Subset
-
#initialize Torch::NN::MaxUnpool1d
-
#inspect Torch::NN::Parameter
-
#inspect Torch::NN::Embedding
-
#inspect Torch::NN::Module
-
#inspect Torch::Inspector
-
instance_norm Torch::NN::Functional
-
interpolate Torch::NN::Functional
-
#item Torch::Tensor
-
kaiming_normal! Torch::NN::Init
-
kaiming_uniform! Torch::NN::Init
-
#kernel_size Torch::NN::ConvNd
-
kl_div Torch::NN::Functional
-
l1_loss Torch::NN::Functional
-
layer_norm Torch::NN::Functional
-
#layout Torch::Tensor
-
leaky_relu Torch::NN::Functional
-
#length Torch::NN::ParameterList
-
#length Torch::Utils::Data::DataPipes::Iter::IterableWrapper
-
#length Torch::Tensor
-
#length Torch::Utils::Data::Subset
-
#length Torch::NN::ModuleList
-
linear Torch::NN::Functional
-
list Torch::Hub
-
load Torch
-
#load_state_dict Torch::Optim::Optimizer
-
#load_state_dict Torch::NN::Module
-
load_state_dict_from_url Torch::Hub
-
local_response_norm Torch::NN::Functional
-
log_sigmoid Torch::NN::Functional
-
log_softmax Torch::NN::Functional
-
margin_ranking_loss Torch::NN::Functional
-
max_pool1d Torch::NN::Functional
-
max_pool2d Torch::NN::Functional
-
max_pool3d Torch::NN::Functional
-
max_unpool1d Torch::NN::Functional
-
max_unpool2d Torch::NN::Functional
-
max_unpool3d Torch::NN::Functional
-
#message Torch::NotImplementedYet
-
#method_missing Torch::NN::Module
-
#modules Torch::NN::Module
-
mse_loss Torch::NN::Functional
-
multi_head_attention_forward Torch::NN::Functional
-
multi_margin_loss Torch::NN::Functional
-
multilabel_margin_loss Torch::NN::Functional
-
multilabel_soft_margin_loss Torch::NN::Functional
-
#named_buffers Torch::NN::Module
-
#named_children Torch::NN::Module
-
#named_modules Torch::NN::Module
-
#named_parameters Torch::NN::Module
-
new Torch::NN::Parameter
-
#new Torch::Tensor
-
new Torch::Tensor
-
new Torch::Autograd::Variable
-
#nhead Torch::NN::Transformer
-
nll_loss Torch::NN::Functional
-
no_grad Torch
-
#non_empty? Torch::Utils::Data::DataPipes::FilterIterDataPipe
-
normal! Torch::NN::Init
-
normalize Torch::NN::Functional
-
#numo Torch::Tensor
-
orthogonal! Torch::NN::Init
-
#out_channels Torch::NN::ConvNd
-
#out_features Torch::NN::Linear
-
#output_paddding Torch::NN::ConvNd
-
pad Torch::NN::Functional
-
#padding Torch::NN::ConvNd
-
#padding_mode Torch::NN::ConvNd
-
pairwise_distance Torch::NN::Functional
-
#param_groups Torch::Optim::Optimizer
-
#parameters Torch::NN::Module
-
#permute_hidden Torch::NN::RNNBase
-
#permute_hidden Torch::NN::LSTM
-
poisson_nll_loss Torch::NN::Functional
-
prelu Torch::NN::Functional
-
#random! Torch::Tensor
-
random_split Torch::Utils::Data
-
#real Torch::Tensor
-
#register_buffer Torch::NN::Module
-
register_datapipe_as_function Torch::Utils::Data::DataPipes::IterDataPipe
-
#register_parameter Torch::NN::Module
-
relu Torch::NN::Functional
-
#requires_grad! Torch::NN::Module
-
#requires_grad= Torch::Tensor
-
#reset Torch::Utils::Data::DataPipes::IterDataPipe
-
#reset_parameters Torch::NN::BatchNorm
-
#reset_parameters Torch::NN::Embedding
-
#reset_parameters Torch::NN::RNNBase
-
#reset_parameters Torch::NN::Bilinear
-
#reset_parameters Torch::NN::Linear
-
#reset_parameters Torch::NN::ConvNd
-
#reset_parameters Torch::NN::MultiheadAttention
-
#reset_parameters Torch::NN::EmbeddingBag
-
#reset_parameters Torch::NN::Transformer
-
#reset_parameters Torch::NN::LayerNorm
-
#reset_parameters Torch::NN::GroupNorm
-
#reset_running_stats Torch::NN::BatchNorm
-
#respond_to? Torch::NN::Module
-
#return_if_true Torch::Utils::Data::DataPipes::FilterIterDataPipe
-
#run_impl Torch::NN::GRU
-
save Torch
-
scaled_dot_product_attention Torch::NN::Functional
-
#share_memory Torch::NN::Module
-
#share_memory Torch::Optim::Adagrad
-
#size Torch::Utils::Data::TensorDataset
-
#size Torch::Utils::Data::DataLoader
-
#size Torch::Tensor
-
smooth_l1_loss Torch::NN::Functional
-
soft_margin_loss Torch::NN::Functional
-
softmax Torch::NN::Functional
-
softmin Torch::NN::Functional
-
softplus Torch::NN::Functional
-
softshrink Torch::NN::Functional
-
softsign Torch::NN::Functional
-
sparse! Torch::NN::Init
-
#state_dict Torch::Optim::Optimizer
-
#state_dict Torch::NN::Module
-
#step Torch::Optim::Adadelta
-
#step Torch::Optim::RMSprop
-
#step Torch::Optim::Adagrad
-
#step Torch::Optim::Adamax
-
#step Torch::Optim::AdamW
-
#step Torch::Optim::Rprop
-
#step Torch::Optim::Adam
-
#step Torch::Optim::ASGD
-
#step Torch::Optim::SGD
-
#step Torch::Optim::LRScheduler::LRScheduler
-
#stride Torch::NN::ConvNd
-
#stride Torch::Tensor
-
tanhshrink Torch::NN::Functional
-
tensor Torch
-
tensor? Torch
-
#tensor_totype Torch::Inspector::Formatter
-
#to Torch::NN::Module
-
#to Torch::Tensor
-
#to_a Torch::Utils::Data::Subset
-
#to_a Torch::Tensor
-
#to_f Torch::Tensor
-
#to_i Torch::Tensor
-
#to_s Torch::Tensor
-
#train Torch::NN::Module
-
#training Torch::NN::Module
-
#transposed Torch::NN::ConvNd
-
triplet_margin_loss Torch::NN::Functional
-
#type Torch::NN::Module
-
#type Torch::Tensor
-
unfold Torch::NN::Functional
-
uniform! Torch::NN::Init
-
#width Torch::Inspector::Formatter
-
xavier_normal! Torch::NN::Init
-
xavier_uniform! Torch::NN::Init
-
#zero_grad Torch::Optim::Optimizer
-
#zero_grad Torch::NN::Module