Method List
-
#% Torch::Tensor
-
#* Torch::Tensor
-
#** Torch::Tensor
-
#+ Torch::Tensor
-
#- Torch::Tensor
-
#-@ Torch::Tensor
-
#/ Torch::Tensor
-
#<=> Torch::Tensor
-
#[] Torch::Tensor
-
#[] Torch::Utils::Data::TensorDataset
-
#[]= Torch::Tensor
-
#_apply Torch::NN::RNNBase
-
#_apply Torch::NN::Module
-
_calculate_fan_in_and_fan_out Torch::NN::Init
-
#_check_input_dim Torch::NN::BatchNorm1d
-
#_check_input_dim Torch::NN::InstanceNorm1d
-
#_check_input_dim Torch::NN::InstanceNorm2d
-
#_check_input_dim Torch::NN::BatchNorm3d
-
#_check_input_dim Torch::NN::BatchNorm2d
-
#_check_input_dim Torch::NN::InstanceNorm3d
-
_dtype_to_numo Torch
-
_make_tensor_class Torch
-
#_ntuple Torch::NN::Utils
-
#_pair Torch::NN::Utils
-
#_quadrupal Torch::NN::Utils
-
#_single Torch::NN::Utils
-
#_triple Torch::NN::Utils
-
adaptive_avg_pool1d Torch::NN::Functional
-
adaptive_avg_pool2d Torch::NN::Functional
-
adaptive_avg_pool3d Torch::NN::Functional
-
adaptive_max_pool1d Torch::NN::Functional
-
adaptive_max_pool2d Torch::NN::Functional
-
adaptive_max_pool3d Torch::NN::Functional
-
#add! Torch::Tensor
-
#add_module Torch::NN::Module
-
#add_param_group Torch::Optim::Optimizer
-
alpha_dropout Torch::NN::Functional
-
#apply Torch::NN::Module
-
arange Torch
-
#args Torch::Native::Function
-
avg_pool1d Torch::NN::Functional
-
avg_pool2d Torch::NN::Functional
-
avg_pool3d Torch::NN::Functional
-
#backward Torch::Tensor
-
#base_name Torch::Native::Function
-
batch_norm Torch::NN::Functional
-
bilinear Torch::NN::Functional
-
binary_cross_entropy Torch::NN::Functional
-
binary_cross_entropy_with_logits Torch::NN::Functional
-
bind Torch::Native::Dispatcher
-
bind_functions Torch::Native::Dispatcher
-
#buffers Torch::NN::Module
-
calculate_gain Torch::NN::Init
-
#call Torch::NN::Module
-
#check_forward_args Torch::NN::LSTM
-
#children Torch::NN::Module
-
constant! Torch::NN::Init
-
conv1d Torch::NN::Functional
-
conv2d Torch::NN::Functional
-
conv3d Torch::NN::Functional
-
cosine_embedding_loss Torch::NN::Functional
-
cosine_similarity Torch::NN::Functional
-
#cpp_name Torch::Native::Function
-
#cpu Torch::Tensor
-
#cpu Torch::NN::Module
-
cross_entropy Torch::NN::Functional
-
ctc_loss Torch::NN::Functional
-
#cuda Torch::Tensor
-
#cuda Torch::NN::Module
-
#dataset Torch::Utils::Data::DataLoader
-
device Torch
-
dirac! Torch::NN::Init
-
#double Torch::NN::Module
-
dropout Torch::NN::Functional
-
dropout2d Torch::NN::Functional
-
dropout3d Torch::NN::Functional
-
#dtype Torch::Tensor
-
#each Torch::Utils::Data::DataLoader
-
embedding Torch::NN::Functional
-
embedding_bag Torch::NN::Functional
-
empty Torch
-
empty_like Torch
-
#eval Torch::NN::Module
-
#extra_inspect Torch::NN::ReflectionPadNd
-
#extra_inspect Torch::NN::RNNBase
-
#extra_inspect Torch::NN::Softplus
-
#extra_inspect Torch::NN::Softmax
-
#extra_inspect Torch::NN::AdaptiveMaxPoolNd
-
#extra_inspect Torch::NN::LayerNorm
-
#extra_inspect Torch::NN::ReplicationPadNd
-
#extra_inspect Torch::NN::LeakyReLU
-
#extra_inspect Torch::NN::AdaptiveAvgPoolNd
-
#extra_inspect Torch::NN::LocalResponseNorm
-
#extra_inspect Torch::NN::Bilinear
-
#extra_inspect Torch::NN::AvgPoolNd
-
#extra_inspect Torch::NN::Hardshrink
-
#extra_inspect Torch::NN::GroupNorm
-
#extra_inspect Torch::NN::LPPoolNd
-
#extra_inspect Torch::NN::MaxUnpoolNd
-
#extra_inspect Torch::NN::DropoutNd
-
#extra_inspect Torch::NN::ReLU
-
#extra_inspect Torch::NN::Linear
-
#extra_inspect Torch::NN::MaxPoolNd
-
#extra_inspect Torch::NN::Conv2d
-
#extra_inspect Torch::NN::ConstantPadNd
-
#extra_inspect Torch::NN::Softshrink
-
#extra_inspect Torch::NN::PReLU
-
eye Torch
-
eye! Torch::NN::Init
-
feature_alpha_dropout Torch::NN::Functional
-
#flatten_parameters Torch::NN::RNNBase
-
#float Torch::NN::Module
-
fold Torch::NN::Functional
-
#forward Torch::NN::Module
-
#forward Torch::NN::HingeEmbeddingLoss
-
#forward Torch::NN::AdaptiveMaxPool3d
-
#forward Torch::NN::AdaptiveAvgPool2d
-
#forward Torch::NN::Linear
-
#forward Torch::NN::Conv3d
-
#forward Torch::NN::Conv2d
-
#forward Torch::NN::Conv1d
-
#forward Torch::NN::PReLU
-
#forward Torch::NN::ReLU
-
#forward Torch::NN::Tanh
-
#forward Torch::NN::MaxPool3d
-
#forward Torch::NN::CrossEntropyLoss
-
#forward Torch::NN::Tanhshrink
-
#forward Torch::NN::AdaptiveMaxPool2d
-
#forward Torch::NN::LocalResponseNorm
-
#forward Torch::NN::ReflectionPadNd
-
#forward Torch::NN::MaxPool1d
-
#forward Torch::NN::LSTM
-
#forward Torch::NN::CosineSimilarity
-
#forward Torch::NN::LogSigmoid
-
#forward Torch::NN::KLDivLoss
-
#forward Torch::NN::Fold
-
#forward Torch::NN::PairwiseDistance
-
#forward Torch::NN::Sequential
-
#forward Torch::NN::AdaptiveAvgPool1d
-
#forward Torch::NN::MultiMarginLoss
-
#forward Torch::NN::TripletMarginLoss
-
#forward Torch::NN::FeatureAlphaDropout
-
#forward Torch::NN::MultiLabelMarginLoss
-
#forward Torch::NN::LogSoftmax
-
#forward Torch::NN::AdaptiveMaxPool1d
-
#forward Torch::NN::PoissonNLLLoss
-
#forward Torch::NN::MaxUnpool2d
-
#forward Torch::NN::Bilinear
-
#forward Torch::NN::BCELoss
-
#forward Torch::NN::AlphaDropout
-
#forward Torch::NN::Softmin
-
#forward Torch::NN::Softmax
-
#forward Torch::NN::LeakyReLU
-
#forward Torch::NN::L1Loss
-
#forward Torch::NN::Dropout
-
#forward Torch::NN::SmoothL1Loss
-
#forward Torch::NN::Sigmoid
-
#forward Torch::NN::Unfold
-
#forward Torch::NN::ConstantPadNd
-
#forward Torch::NN::MultiLabelSoftMarginLoss
-
#forward Torch::NN::CTCLoss
-
#forward Torch::NN::CosineEmbeddingLoss
-
#forward Torch::NN::Softplus
-
#forward Torch::NN::RNNBase
-
#forward Torch::NN::MarginRankingLoss
-
#forward Torch::NN::NLLLoss
-
#forward Torch::NN::EmbeddingBag
-
#forward Torch::NN::MSELoss
-
#forward Torch::NN::Identity
-
#forward Torch::NN::BCEWithLogitsLoss
-
#forward Torch::NN::ReplicationPadNd
-
#forward Torch::NN::Dropout2d
-
#forward Torch::NN::Dropout3d
-
#forward Torch::NN::Softsign
-
#forward Torch::NN::InstanceNorm
-
#forward Torch::NN::LayerNorm
-
#forward Torch::NN::GRU
-
#forward Torch::NN::MaxUnpool3d
-
#forward Torch::NN::Softshrink
-
#forward Torch::NN::SoftMarginLoss
-
#forward Torch::NN::AdaptiveAvgPool3d
-
#forward Torch::NN::GroupNorm
-
#forward Torch::NN::BatchNorm
-
#forward Torch::NN::AvgPool2d
-
#forward Torch::NN::AvgPool3d
-
#forward Torch::NN::Softmax2d
-
#forward Torch::NN::LPPool2d
-
#forward Torch::NN::LPPool1d
-
#forward Torch::NN::Embedding
-
#forward Torch::NN::AvgPool1d
-
#forward Torch::NN::MaxPool2d
-
#forward Torch::NN::MaxUnpool1d
-
#forward Torch::NN::Hardshrink
-
#forward_impl Torch::NN::LSTM
-
#forward_impl Torch::NN::GRU
-
#forward_tensor Torch::NN::LSTM
-
#forward_tensor Torch::NN::GRU
-
from_numo Torch
-
full Torch
-
full_like Torch
-
#func Torch::Native::Function
-
#function Torch::Native::Function
-
generate_cpp_functions Torch::Native::Generator
-
#get_lr Torch::Optim::LRScheduler::LambdaLR
-
#get_lr Torch::Optim::LRScheduler::MultiStepLR
-
#get_lr Torch::Optim::LRScheduler::ExponentialLR
-
#get_lr Torch::Optim::LRScheduler::StepLR
-
#get_lr Torch::Optim::LRScheduler::CosineAnnealingLR
-
#get_lr Torch::Optim::LRScheduler::MultiplicativeLR
-
group_norm Torch::NN::Functional
-
grouped_functions Torch::Native::Generator
-
#half Torch::NN::Module
-
hardshrink Torch::NN::Functional
-
hinge_embedding_loss Torch::NN::Functional
-
initial_seed Torch::Random
-
#initialize Torch::Optim::Rprop
-
#initialize Torch::Optim::AdamW
-
#initialize Torch::NN::DropoutNd
-
#initialize Torch::NN::Softplus
-
#initialize Torch::NN::RNNBase
-
#initialize Torch::NN::Identity
-
#initialize Torch::NN::MSELoss
-
#initialize Torch::NN::NLLLoss
-
#initialize Torch::NN::Bilinear
-
#initialize Torch::Optim::ASGD
-
#initialize Torch::NN::BCELoss
-
#initialize Torch::NN::CTCLoss
-
#initialize Torch::NN::Softmin
-
#initialize Torch::Optim::Adam
-
#initialize Torch::Optim::LRScheduler::StepLR
-
#initialize Torch::NN::SmoothL1Loss
-
#initialize Torch::NN::L1Loss
-
#initialize Torch::Optim::SGD
-
#initialize Torch::NN::Unfold
-
#initialize Torch::NN::Embedding
-
#initialize Torch::Optim::LRScheduler::CosineAnnealingLR
-
#initialize Torch::NN::ConstantPad2d
-
#initialize Torch::NN::AdaptiveAvgPoolNd
-
#initialize Torch::NN::CrossEntropyLoss
-
#initialize Torch::Optim::LRScheduler::MultiplicativeLR
-
#initialize Torch::NN::Softmax
-
#initialize Torch::NN::ReplicationPad1d
-
#initialize Torch::Optim::Optimizer
-
#initialize Torch::NN::LocalResponseNorm
-
#initialize Torch::NN::Module
-
#initialize Torch::NN::Linear
-
#initialize Torch::NN::Conv3d
-
#initialize Torch::NN::ConvNd
-
#initialize Torch::NN::Conv1d
-
#initialize Torch::NN::Conv2d
-
#initialize Torch::NN::PReLU
-
#initialize Torch::NN::CosineEmbeddingLoss
-
#initialize Torch::NN::ConstantPad1d
-
#initialize Torch::NN::ReplicationPad3d
-
#initialize Torch::NN::MultiLabelMarginLoss
-
#initialize Torch::NN::MaxPoolNd
-
#initialize Torch::NN::MultiLabelSoftMarginLoss
-
#initialize Torch::NN::HingeEmbeddingLoss
-
#initialize Torch::Optim::LRScheduler::LambdaLR
-
#initialize Torch::NN::MaxUnpool1d
-
#initialize Torch::NN::PairwiseDistance
-
#initialize Torch::NN::ConstantPadNd
-
#initialize Torch::NN::ReplicationPad2d
-
#initialize Torch::NN::InstanceNorm
-
#initialize Torch::NN::BatchNorm
-
#initialize Torch::NN::AvgPool2d
-
#initialize Torch::NN::AvgPool3d
-
#initialize Torch::Optim::LRScheduler::LRScheduler
-
#initialize Torch::Native::Parser
-
#initialize Torch::Optim::Adamax
-
#initialize Torch::NN::AvgPool1d
-
#initialize Torch::NN::TripletMarginLoss
-
#initialize Torch::NN::LPPoolNd
-
#initialize Torch::NN::AdaptiveMaxPoolNd
-
#initialize Torch::NN::GroupNorm
-
#initialize Torch::Optim::LRScheduler::MultiStepLR
-
#initialize Torch::NN::ReflectionPad2d
-
#initialize Torch::NN::WeightedLoss
-
#initialize Torch::NN::ConstantPad3d
-
#initialize Torch::NN::ReflectionPad1d
-
#initialize Torch::NN::LayerNorm
-
#initialize Torch::NN::Hardshrink
-
#initialize Torch::NN::LeakyReLU
-
#initialize Torch::NN::PoissonNLLLoss
-
#initialize Torch::NN::MaxUnpool2d
-
#initialize Torch::NN::EmbeddingBag
-
#initialize Torch::NN::MarginRankingLoss
-
#initialize Torch::NN::Loss
-
#initialize Torch::NN::LSTM
-
#initialize Torch::NN::ReLU
-
#initialize Torch::Utils::Data::TensorDataset
-
#initialize Torch::NN::Fold
-
#initialize Torch::NN::Softshrink
-
#initialize Torch::NN::MaxUnpool3d
-
#initialize Torch::NN::CosineSimilarity
-
#initialize Torch::NN::SoftMarginLoss
-
#initialize Torch::NN::BCEWithLogitsLoss
-
#initialize Torch::Optim::LRScheduler::ExponentialLR
-
#initialize Torch::NN::GRU
-
#initialize Torch::NN::RNN
-
#initialize Torch::Utils::Data::DataLoader
-
#initialize Torch::NN::Sequential
-
#initialize Torch::NN::KLDivLoss
-
#initialize Torch::Optim::RMSprop
-
#initialize Torch::NN::ZeroPad2d
-
#initialize Torch::Optim::Adagrad
-
#initialize Torch::NN::MultiMarginLoss
-
#initialize Torch::NN::LogSoftmax
-
#initialize Torch::Optim::Adadelta
-
#initialize Torch::Native::Function
-
#inspect Torch::NN::Parameter
-
#inspect Torch::NN::Embedding
-
#inspect Torch::NN::Module
-
#inspect Torch::Inspector
-
instance_norm Torch::NN::Functional
-
#item Torch::Tensor
-
kaiming_normal! Torch::NN::Init
-
kaiming_uniform! Torch::NN::Init
-
kl_div Torch::NN::Functional
-
l1_loss Torch::NN::Functional
-
layer_norm Torch::NN::Functional
-
#layout Torch::Tensor
-
leaky_relu Torch::NN::Functional
-
#length Torch::Tensor
-
linear Torch::NN::Functional
-
linspace Torch
-
list Torch::Hub
-
load Torch
-
#load_state_dict Torch::Optim::Optimizer
-
#load_state_dict Torch::NN::Module
-
local_response_norm Torch::NN::Functional
-
log_sigmoid Torch::NN::Functional
-
log_softmax Torch::NN::Functional
-
logspace Torch
-
margin_ranking_loss Torch::NN::Functional
-
max_pool1d Torch::NN::Functional
-
max_pool2d Torch::NN::Functional
-
max_pool3d Torch::NN::Functional
-
max_unpool1d Torch::NN::Functional
-
max_unpool2d Torch::NN::Functional
-
max_unpool3d Torch::NN::Functional
-
#message Torch::NotImplementedYet
-
#method_missing Torch::NN::Module
-
#modules Torch::NN::Module
-
mse_loss Torch::NN::Functional
-
multi_margin_loss Torch::NN::Functional
-
multilabel_margin_loss Torch::NN::Functional
-
multilabel_soft_margin_loss Torch::NN::Functional
-
#name Torch::Native::Function
-
#named_buffers Torch::NN::Module
-
#named_children Torch::NN::Module
-
#named_modules Torch::NN::Module
-
#named_parameters Torch::NN::Module
-
new Torch::NN::Parameter
-
#new Torch::Tensor
-
new Torch::Tensor
-
#new_ones Torch::Tensor
-
nll_loss Torch::NN::Functional
-
no_grad Torch
-
normal! Torch::NN::Init
-
#numo Torch::Tensor
-
ones Torch
-
ones! Torch::NN::Init
-
ones_like Torch
-
orthogonal! Torch::NN::Init
-
#out? Torch::Native::Function
-
#out_size Torch::Native::Function
-
pad Torch::NN::Functional
-
pairwise_distance Torch::NN::Functional
-
#param_groups Torch::Optim::Optimizer
-
#parameters Torch::NN::Module
-
#parse Torch::Native::Parser
-
#permute_hidden Torch::NN::RNNBase
-
#permute_hidden Torch::NN::LSTM
-
poisson_nll_loss Torch::NN::Functional
-
prelu Torch::NN::Functional
-
#python_module Torch::Native::Function
-
rand Torch
-
rand_like Torch
-
randint Torch
-
randint_like Torch
-
randn Torch
-
randn_like Torch
-
#random! Torch::Tensor
-
randperm Torch
-
#register_buffer Torch::NN::Module
-
#register_parameter Torch::NN::Module
-
relu Torch::NN::Functional
-
#requires_grad! Torch::NN::Module
-
#requires_grad! Torch::Tensor
-
#reset_parameters Torch::NN::LayerNorm
-
#reset_parameters Torch::NN::GroupNorm
-
#reset_parameters Torch::NN::BatchNorm
-
#reset_parameters Torch::NN::RNNBase
-
#reset_parameters Torch::NN::Bilinear
-
#reset_parameters Torch::NN::ConvNd
-
#reset_parameters Torch::NN::Linear
-
#reset_parameters Torch::NN::Embedding
-
#reset_parameters Torch::NN::EmbeddingBag
-
#reset_running_stats Torch::NN::BatchNorm
-
#reshape Torch::Tensor
-
#respond_to? Torch::NN::Module
-
#ret_size Torch::Native::Function
-
#ruby_name Torch::Native::Function
-
#run_impl Torch::NN::GRU
-
save Torch
-
#shape Torch::Tensor
-
#share_memory Torch::NN::Module
-
#share_memory Torch::Optim::Adagrad
-
#size Torch::Utils::Data::DataLoader
-
#size Torch::Tensor
-
#size Torch::Utils::Data::TensorDataset
-
smooth_l1_loss Torch::NN::Functional
-
soft_margin_loss Torch::NN::Functional
-
softmax Torch::NN::Functional
-
softmin Torch::NN::Functional
-
softplus Torch::NN::Functional
-
softshrink Torch::NN::Functional
-
softsign Torch::NN::Functional
-
sparse! Torch::NN::Init
-
#state_dict Torch::NN::Module
-
#state_dict Torch::Optim::Optimizer
-
#step Torch::Optim::Adamax
-
#step Torch::Optim::LRScheduler::LRScheduler
-
#step Torch::Optim::Rprop
-
#step Torch::Optim::AdamW
-
#step Torch::Optim::Adam
-
#step Torch::Optim::ASGD
-
#step Torch::Optim::SGD
-
#step Torch::Optim::Adadelta
-
#step Torch::Optim::Adagrad
-
#step Torch::Optim::RMSprop
-
tanhshrink Torch::NN::Functional
-
tensor Torch
-
tensor? Torch
-
#to Torch::NN::Module
-
#to Torch::Tensor
-
#to_a Torch::Tensor
-
#to_s Torch::Tensor
-
#train Torch::NN::Module
-
triplet_margin_loss Torch::NN::Functional
-
#type Torch::NN::Module
-
#type Torch::Tensor
-
unfold Torch::NN::Functional
-
uniform! Torch::NN::Init
-
#variants Torch::Native::Function
-
#view Torch::Tensor
-
xavier_normal! Torch::NN::Init
-
xavier_uniform! Torch::NN::Init
-
#zero_grad Torch::NN::Module
-
#zero_grad Torch::Optim::Optimizer
-
zeros Torch
-
zeros! Torch::NN::Init
-
zeros_like Torch