Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
  • Learn
    Get Started

    Run PyTorch locally or get started quickly with one of the supported cloud platforms

    Tutorials

    Whats new in PyTorch tutorials

    Learn the Basics

    Familiarize yourself with PyTorch concepts and modules

    PyTorch Recipes

    Bite-size, ready-to-deploy PyTorch code examples

    Intro to PyTorch - YouTube Series

    Master PyTorch basics with our engaging YouTube tutorial series

  • Ecosystem
    Tools

    Learn about the tools and frameworks in the PyTorch Ecosystem

    Community

    Join the PyTorch developer community to contribute, learn, and get your questions answered

    Forums

    A place to discuss PyTorch code, issues, install, research

    Developer Resources

    Find resources and get questions answered

    Contributor Awards - 2024

    Award winners announced at this year's PyTorch Conference

  • Edge
    About PyTorch Edge

    Build innovative and privacy-aware AI experiences for edge devices

    ExecuTorch

    End-to-end solution for enabling on-device inference capabilities across mobile and edge devices

    ExecuTorch Docs
  • Docs
    PyTorch

    Explore the documentation for comprehensive guidance on how to use PyTorch

    PyTorch Domains

    Read the PyTorch Domains documentation to learn more about domain-specific libraries

  • Blogs & News
    PyTorch Blog

    Catch up on the latest technical news and happenings

    Community Blog

    Stories from the PyTorch ecosystem

    Videos

    Learn about the latest PyTorch tutorials, new, and more

    Community Stories

    Learn how our community solves real, everyday machine learning problems with PyTorch

    Events

    Find events, webinars, and podcasts

    Newsletter

    Stay up-to-date with the latest updates

  • About
    PyTorch Foundation

    Learn more about the PyTorch Foundation

    Governing Board Cloud Credit Program Technical Advisory Council Staff Contact Us
  • Become a Member
  • Learn
    • Get Started
    • Tutorials
    • Learn the Basics
    • PyTorch Recipes
    • Introduction to PyTorch - YouTube Series
  • Ecosystem
    • Tools
    • Community
    • Forums
    • Developer Resources
    • Contributor Awards - 2024
  • Edge
    • About PyTorch Edge
    • ExecuTorch
    • ExecuTorch Documentation
  • Docs
    • PyTorch
    • PyTorch Domains
  • Blog & News
    • PyTorch Blog
    • Community Blog
    • Videos
    • Community Stories
    • Events
    • Newsletter
  • About
    • PyTorch Foundation
    • Governing Board
    • Cloud Credit Program
    • Technical Advisory Council
    • Staff
    • Contact Us
Skip to main content
Ctrl+K

Home

  • Python API
  • Developer Notes
  • Tutorials
Ctrl+K
  • X
  • GitHub
  • PyTorch Forum
  • PyPi
  • Python API
  • Developer Notes
  • Tutorials
Ctrl+K
  • X
  • GitHub
  • PyTorch Forum
  • PyPi

Section Navigation

  • torch
    • torch.is_tensor
    • torch.is_storage
    • torch.is_complex
    • torch.is_conj
    • torch.is_floating_point
    • torch.is_nonzero
    • torch.set_default_dtype
    • torch.get_default_dtype
    • torch.set_default_device
    • torch.get_default_device
    • torch.set_default_tensor_type
    • torch.numel
    • torch.set_printoptions
    • torch.set_flush_denormal
    • torch.tensor
    • torch.sparse_coo_tensor
    • torch.sparse_csr_tensor
    • torch.sparse_csc_tensor
    • torch.sparse_bsr_tensor
    • torch.sparse_bsc_tensor
    • torch.asarray
    • torch.as_tensor
    • torch.as_strided
    • torch.from_file
    • torch.from_numpy
    • torch.from_dlpack
    • torch.frombuffer
    • torch.zeros
    • torch.zeros_like
    • torch.ones
    • torch.ones_like
    • torch.arange
    • torch.range
    • torch.linspace
    • torch.logspace
    • torch.eye
    • torch.empty
    • torch.empty_like
    • torch.empty_strided
    • torch.full
    • torch.full_like
    • torch.quantize_per_tensor
    • torch.quantize_per_channel
    • torch.dequantize
    • torch.complex
    • torch.polar
    • torch.heaviside
    • torch.adjoint
    • torch.argwhere
    • torch.cat
    • torch.concat
    • torch.concatenate
    • torch.conj
    • torch.chunk
    • torch.dsplit
    • torch.column_stack
    • torch.dstack
    • torch.gather
    • torch.hsplit
    • torch.hstack
    • torch.index_add
    • torch.index_copy
    • torch.index_reduce
    • torch.index_select
    • torch.masked_select
    • torch.movedim
    • torch.moveaxis
    • torch.narrow
    • torch.narrow_copy
    • torch.nonzero
    • torch.permute
    • torch.reshape
    • torch.row_stack
    • torch.select
    • torch.scatter
    • torch.diagonal_scatter
    • torch.select_scatter
    • torch.slice_scatter
    • torch.scatter_add
    • torch.scatter_reduce
    • torch.split
    • torch.squeeze
    • torch.stack
    • torch.swapaxes
    • torch.swapdims
    • torch.t
    • torch.take
    • torch.take_along_dim
    • torch.tensor_split
    • torch.tile
    • torch.transpose
    • torch.unbind
    • torch.unravel_index
    • torch.unsqueeze
    • torch.vsplit
    • torch.vstack
    • torch.where
    • Stream
    • Event
    • Generator
    • torch.seed
    • torch.manual_seed
    • torch.initial_seed
    • torch.get_rng_state
    • torch.set_rng_state
    • torch.bernoulli
    • torch.multinomial
    • torch.normal
    • torch.poisson
    • torch.rand
    • torch.rand_like
    • torch.randint
    • torch.randint_like
    • torch.randn
    • torch.randn_like
    • torch.randperm
    • SobolEngine
    • torch.save
    • torch.load
    • torch.get_num_threads
    • torch.set_num_threads
    • torch.get_num_interop_threads
    • torch.set_num_interop_threads
    • no_grad
    • enable_grad
    • set_grad_enabled
    • torch.is_grad_enabled
    • inference_mode
    • torch.is_inference_mode_enabled
    • torch.abs
    • torch.absolute
    • torch.acos
    • torch.arccos
    • torch.acosh
    • torch.arccosh
    • torch.add
    • torch.addcdiv
    • torch.addcmul
    • torch.angle
    • torch.asin
    • torch.arcsin
    • torch.asinh
    • torch.arcsinh
    • torch.atan
    • torch.arctan
    • torch.atanh
    • torch.arctanh
    • torch.atan2
    • torch.arctan2
    • torch.bitwise_not
    • torch.bitwise_and
    • torch.bitwise_or
    • torch.bitwise_xor
    • torch.bitwise_left_shift
    • torch.bitwise_right_shift
    • torch.ceil
    • torch.clamp
    • torch.clip
    • torch.conj_physical
    • torch.copysign
    • torch.cos
    • torch.cosh
    • torch.deg2rad
    • torch.div
    • torch.divide
    • torch.digamma
    • torch.erf
    • torch.erfc
    • torch.erfinv
    • torch.exp
    • torch.exp2
    • torch.expm1
    • torch.fake_quantize_per_channel_affine
    • torch.fake_quantize_per_tensor_affine
    • torch.fix
    • torch.float_power
    • torch.floor
    • torch.floor_divide
    • torch.fmod
    • torch.frac
    • torch.frexp
    • torch.gradient
    • torch.imag
    • torch.ldexp
    • torch.lerp
    • torch.lgamma
    • torch.log
    • torch.log10
    • torch.log1p
    • torch.log2
    • torch.logaddexp
    • torch.logaddexp2
    • torch.logical_and
    • torch.logical_not
    • torch.logical_or
    • torch.logical_xor
    • torch.logit
    • torch.hypot
    • torch.i0
    • torch.igamma
    • torch.igammac
    • torch.mul
    • torch.multiply
    • torch.mvlgamma
    • torch.nan_to_num
    • torch.neg
    • torch.negative
    • torch.nextafter
    • torch.polygamma
    • torch.positive
    • torch.pow
    • torch.quantized_batch_norm
    • torch.quantized_max_pool1d
    • torch.quantized_max_pool2d
    • torch.rad2deg
    • torch.real
    • torch.reciprocal
    • torch.remainder
    • torch.round
    • torch.rsqrt
    • torch.sigmoid
    • torch.sign
    • torch.sgn
    • torch.signbit
    • torch.sin
    • torch.sinc
    • torch.sinh
    • torch.softmax
    • torch.sqrt
    • torch.square
    • torch.sub
    • torch.subtract
    • torch.tan
    • torch.tanh
    • torch.true_divide
    • torch.trunc
    • torch.xlogy
    • torch.argmax
    • torch.argmin
    • torch.amax
    • torch.amin
    • torch.aminmax
    • torch.all
    • torch.any
    • torch.max
    • torch.min
    • torch.dist
    • torch.logsumexp
    • torch.mean
    • torch.nanmean
    • torch.median
    • torch.nanmedian
    • torch.mode
    • torch.norm
    • torch.nansum
    • torch.prod
    • torch.quantile
    • torch.nanquantile
    • torch.std
    • torch.std_mean
    • torch.sum
    • torch.unique
    • torch.unique_consecutive
    • torch.var
    • torch.var_mean
    • torch.count_nonzero
    • torch.allclose
    • torch.argsort
    • torch.eq
    • torch.equal
    • torch.ge
    • torch.greater_equal
    • torch.gt
    • torch.greater
    • torch.isclose
    • torch.isfinite
    • torch.isin
    • torch.isinf
    • torch.isposinf
    • torch.isneginf
    • torch.isnan
    • torch.isreal
    • torch.kthvalue
    • torch.le
    • torch.less_equal
    • torch.lt
    • torch.less
    • torch.maximum
    • torch.minimum
    • torch.fmax
    • torch.fmin
    • torch.ne
    • torch.not_equal
    • torch.sort
    • torch.topk
    • torch.msort
    • torch.stft
    • torch.istft
    • torch.bartlett_window
    • torch.blackman_window
    • torch.hamming_window
    • torch.hann_window
    • torch.kaiser_window
    • torch.atleast_1d
    • torch.atleast_2d
    • torch.atleast_3d
    • torch.bincount
    • torch.block_diag
    • torch.broadcast_tensors
    • torch.broadcast_to
    • torch.broadcast_shapes
    • torch.bucketize
    • torch.cartesian_prod
    • torch.cdist
    • torch.clone
    • torch.combinations
    • torch.corrcoef
    • torch.cov
    • torch.cross
    • torch.cummax
    • torch.cummin
    • torch.cumprod
    • torch.cumsum
    • torch.diag
    • torch.diag_embed
    • torch.diagflat
    • torch.diagonal
    • torch.diff
    • torch.einsum
    • torch.flatten
    • torch.flip
    • torch.fliplr
    • torch.flipud
    • torch.kron
    • torch.rot90
    • torch.gcd
    • torch.histc
    • torch.histogram
    • torch.histogramdd
    • torch.meshgrid
    • torch.lcm
    • torch.logcumsumexp
    • torch.ravel
    • torch.renorm
    • torch.repeat_interleave
    • torch.roll
    • torch.searchsorted
    • torch.tensordot
    • torch.trace
    • torch.tril
    • torch.tril_indices
    • torch.triu
    • torch.triu_indices
    • torch.unflatten
    • torch.vander
    • torch.view_as_real
    • torch.view_as_complex
    • torch.resolve_conj
    • torch.resolve_neg
    • torch.addbmm
    • torch.addmm
    • torch.addmv
    • torch.addr
    • torch.baddbmm
    • torch.bmm
    • torch.chain_matmul
    • torch.cholesky
    • torch.cholesky_inverse
    • torch.cholesky_solve
    • torch.dot
    • torch.geqrf
    • torch.ger
    • torch.inner
    • torch.inverse
    • torch.det
    • torch.logdet
    • torch.slogdet
    • torch.lu
    • torch.lu_solve
    • torch.lu_unpack
    • torch.matmul
    • torch.matrix_power
    • torch.matrix_exp
    • torch.mm
    • torch.mv
    • torch.orgqr
    • torch.ormqr
    • torch.outer
    • torch.pinverse
    • torch.qr
    • torch.svd
    • torch.svd_lowrank
    • torch.pca_lowrank
    • torch.lobpcg
    • torch.trapz
    • torch.trapezoid
    • torch.cumulative_trapezoid
    • torch.triangular_solve
    • torch.vdot
    • torch._foreach_abs
    • torch._foreach_abs_
    • torch._foreach_acos
    • torch._foreach_acos_
    • torch._foreach_asin
    • torch._foreach_asin_
    • torch._foreach_atan
    • torch._foreach_atan_
    • torch._foreach_ceil
    • torch._foreach_ceil_
    • torch._foreach_cos
    • torch._foreach_cos_
    • torch._foreach_cosh
    • torch._foreach_cosh_
    • torch._foreach_erf
    • torch._foreach_erf_
    • torch._foreach_erfc
    • torch._foreach_erfc_
    • torch._foreach_exp
    • torch._foreach_exp_
    • torch._foreach_expm1
    • torch._foreach_expm1_
    • torch._foreach_floor
    • torch._foreach_floor_
    • torch._foreach_log
    • torch._foreach_log_
    • torch._foreach_log10
    • torch._foreach_log10_
    • torch._foreach_log1p
    • torch._foreach_log1p_
    • torch._foreach_log2
    • torch._foreach_log2_
    • torch._foreach_neg
    • torch._foreach_neg_
    • torch._foreach_tan
    • torch._foreach_tan_
    • torch._foreach_sin
    • torch._foreach_sin_
    • torch._foreach_sinh
    • torch._foreach_sinh_
    • torch._foreach_round
    • torch._foreach_round_
    • torch._foreach_sqrt
    • torch._foreach_sqrt_
    • torch._foreach_lgamma
    • torch._foreach_lgamma_
    • torch._foreach_frac
    • torch._foreach_frac_
    • torch._foreach_reciprocal
    • torch._foreach_reciprocal_
    • torch._foreach_sigmoid
    • torch._foreach_sigmoid_
    • torch._foreach_trunc
    • torch._foreach_trunc_
    • torch._foreach_zero_
    • torch.compiled_with_cxx11_abi
    • torch.result_type
    • torch.can_cast
    • torch.promote_types
    • torch.use_deterministic_algorithms
    • torch.are_deterministic_algorithms_enabled
    • torch.is_deterministic_algorithms_warn_only_enabled
    • torch.set_deterministic_debug_mode
    • torch.get_deterministic_debug_mode
    • torch.set_float32_matmul_precision
    • torch.get_float32_matmul_precision
    • torch.set_warn_always
    • torch.get_device_module
    • torch.is_warn_always_enabled
    • torch.vmap
    • torch._assert
    • torch.sym_float
    • torch.sym_fresh_size
    • torch.sym_int
    • torch.sym_max
    • torch.sym_min
    • torch.sym_not
    • torch.sym_ite
    • torch.sym_sum
    • torch.cond
    • torch.compile
  • torch.nn
    • Buffer
    • Parameter
    • UninitializedParameter
    • UninitializedBuffer
    • Module
    • Sequential
    • ModuleList
    • ModuleDict
    • ParameterList
    • ParameterDict
    • torch.nn.modules.module.register_module_forward_pre_hook
    • torch.nn.modules.module.register_module_forward_hook
    • torch.nn.modules.module.register_module_backward_hook
    • torch.nn.modules.module.register_module_full_backward_pre_hook
    • torch.nn.modules.module.register_module_full_backward_hook
    • torch.nn.modules.module.register_module_buffer_registration_hook
    • torch.nn.modules.module.register_module_module_registration_hook
    • torch.nn.modules.module.register_module_parameter_registration_hook
    • Conv1d
    • Conv2d
    • Conv3d
    • ConvTranspose1d
    • ConvTranspose2d
    • ConvTranspose3d
    • LazyConv1d
    • LazyConv2d
    • LazyConv3d
    • LazyConvTranspose1d
    • LazyConvTranspose2d
    • LazyConvTranspose3d
    • Unfold
    • Fold
    • MaxPool1d
    • MaxPool2d
    • MaxPool3d
    • MaxUnpool1d
    • MaxUnpool2d
    • MaxUnpool3d
    • AvgPool1d
    • AvgPool2d
    • AvgPool3d
    • FractionalMaxPool2d
    • FractionalMaxPool3d
    • LPPool1d
    • LPPool2d
    • LPPool3d
    • AdaptiveMaxPool1d
    • AdaptiveMaxPool2d
    • AdaptiveMaxPool3d
    • AdaptiveAvgPool1d
    • AdaptiveAvgPool2d
    • AdaptiveAvgPool3d
    • ReflectionPad1d
    • ReflectionPad2d
    • ReflectionPad3d
    • ReplicationPad1d
    • ReplicationPad2d
    • ReplicationPad3d
    • ZeroPad1d
    • ZeroPad2d
    • ZeroPad3d
    • ConstantPad1d
    • ConstantPad2d
    • ConstantPad3d
    • CircularPad1d
    • CircularPad2d
    • CircularPad3d
    • ELU
    • Hardshrink
    • Hardsigmoid
    • Hardtanh
    • Hardswish
    • LeakyReLU
    • LogSigmoid
    • MultiheadAttention
    • PReLU
    • ReLU
    • ReLU6
    • RReLU
    • SELU
    • CELU
    • GELU
    • Sigmoid
    • SiLU
    • Mish
    • Softplus
    • Softshrink
    • Softsign
    • Tanh
    • Tanhshrink
    • Threshold
    • GLU
    • Softmin
    • Softmax
    • Softmax2d
    • LogSoftmax
    • AdaptiveLogSoftmaxWithLoss
    • BatchNorm1d
    • BatchNorm2d
    • BatchNorm3d
    • LazyBatchNorm1d
    • LazyBatchNorm2d
    • LazyBatchNorm3d
    • GroupNorm
    • SyncBatchNorm
    • InstanceNorm1d
    • InstanceNorm2d
    • InstanceNorm3d
    • LazyInstanceNorm1d
    • LazyInstanceNorm2d
    • LazyInstanceNorm3d
    • LayerNorm
    • LocalResponseNorm
    • RMSNorm
    • RNNBase
    • RNN
    • LSTM
    • GRU
    • RNNCell
    • LSTMCell
    • GRUCell
    • Transformer
    • TransformerEncoder
    • TransformerDecoder
    • TransformerEncoderLayer
    • TransformerDecoderLayer
    • Identity
    • Linear
    • Bilinear
    • LazyLinear
    • Dropout
    • Dropout1d
    • Dropout2d
    • Dropout3d
    • AlphaDropout
    • FeatureAlphaDropout
    • Embedding
    • EmbeddingBag
    • CosineSimilarity
    • PairwiseDistance
    • L1Loss
    • MSELoss
    • CrossEntropyLoss
    • CTCLoss
    • NLLLoss
    • PoissonNLLLoss
    • GaussianNLLLoss
    • KLDivLoss
    • BCELoss
    • BCEWithLogitsLoss
    • MarginRankingLoss
    • HingeEmbeddingLoss
    • MultiLabelMarginLoss
    • HuberLoss
    • SmoothL1Loss
    • SoftMarginLoss
    • MultiLabelSoftMarginLoss
    • CosineEmbeddingLoss
    • MultiMarginLoss
    • TripletMarginLoss
    • TripletMarginWithDistanceLoss
    • PixelShuffle
    • PixelUnshuffle
    • Upsample
    • UpsamplingNearest2d
    • UpsamplingBilinear2d
    • ChannelShuffle
    • DataParallel
    • DistributedDataParallel
    • torch.nn.utils.clip_grad_norm_
    • torch.nn.utils.clip_grad_norm
    • torch.nn.utils.clip_grad_value_
    • torch.nn.utils.get_total_norm
    • torch.nn.utils.clip_grads_with_norm_
    • torch.nn.utils.parameters_to_vector
    • torch.nn.utils.vector_to_parameters
    • torch.nn.utils.fuse_conv_bn_eval
    • torch.nn.utils.fuse_conv_bn_weights
    • torch.nn.utils.fuse_linear_bn_eval
    • torch.nn.utils.fuse_linear_bn_weights
    • torch.nn.utils.convert_conv2d_weight_memory_format
    • torch.nn.utils.convert_conv3d_weight_memory_format
    • torch.nn.utils.weight_norm
    • torch.nn.utils.remove_weight_norm
    • torch.nn.utils.spectral_norm
    • torch.nn.utils.remove_spectral_norm
    • torch.nn.utils.skip_init
    • BasePruningMethod
    • PruningContainer
    • Identity
    • RandomUnstructured
    • L1Unstructured
    • RandomStructured
    • LnStructured
    • CustomFromMask
    • torch.nn.utils.prune.identity
    • torch.nn.utils.prune.random_unstructured
    • torch.nn.utils.prune.l1_unstructured
    • torch.nn.utils.prune.random_structured
    • torch.nn.utils.prune.ln_structured
    • torch.nn.utils.prune.global_unstructured
    • torch.nn.utils.prune.custom_from_mask
    • torch.nn.utils.prune.remove
    • torch.nn.utils.prune.is_pruned
    • torch.nn.utils.parametrizations.orthogonal
    • torch.nn.utils.parametrizations.weight_norm
    • torch.nn.utils.parametrizations.spectral_norm
    • torch.nn.utils.parametrize.register_parametrization
    • torch.nn.utils.parametrize.remove_parametrizations
    • torch.nn.utils.parametrize.cached
    • torch.nn.utils.parametrize.is_parametrized
    • ParametrizationList
    • torch.nn.utils.stateless.functional_call
    • PackedSequence
    • torch.nn.utils.rnn.pack_padded_sequence
    • torch.nn.utils.rnn.pad_packed_sequence
    • torch.nn.utils.rnn.pad_sequence
    • torch.nn.utils.rnn.pack_sequence
    • torch.nn.utils.rnn.unpack_sequence
    • torch.nn.utils.rnn.unpad_sequence
    • Flatten
    • Unflatten
    • LazyModuleMixin
    • RMSNorm
  • torch.nn.functional
    • torch.nn.functional.conv1d
    • torch.nn.functional.conv2d
    • torch.nn.functional.conv3d
    • torch.nn.functional.conv_transpose1d
    • torch.nn.functional.conv_transpose2d
    • torch.nn.functional.conv_transpose3d
    • torch.nn.functional.unfold
    • torch.nn.functional.fold
    • torch.nn.functional.avg_pool1d
    • torch.nn.functional.avg_pool2d
    • torch.nn.functional.avg_pool3d
    • torch.nn.functional.max_pool1d
    • torch.nn.functional.max_pool2d
    • torch.nn.functional.max_pool3d
    • torch.nn.functional.max_unpool1d
    • torch.nn.functional.max_unpool2d
    • torch.nn.functional.max_unpool3d
    • torch.nn.functional.lp_pool1d
    • torch.nn.functional.lp_pool2d
    • torch.nn.functional.lp_pool3d
    • torch.nn.functional.adaptive_max_pool1d
    • torch.nn.functional.adaptive_max_pool2d
    • torch.nn.functional.adaptive_max_pool3d
    • torch.nn.functional.adaptive_avg_pool1d
    • torch.nn.functional.adaptive_avg_pool2d
    • torch.nn.functional.adaptive_avg_pool3d
    • torch.nn.functional.fractional_max_pool2d
    • torch.nn.functional.fractional_max_pool3d
    • torch.nn.functional.scaled_dot_product_attention
    • torch.nn.functional.threshold
    • torch.nn.functional.threshold_
    • torch.nn.functional.relu
    • torch.nn.functional.relu_
    • torch.nn.functional.hardtanh
    • torch.nn.functional.hardtanh_
    • torch.nn.functional.hardswish
    • torch.nn.functional.relu6
    • torch.nn.functional.elu
    • torch.nn.functional.elu_
    • torch.nn.functional.selu
    • torch.nn.functional.celu
    • torch.nn.functional.leaky_relu
    • torch.nn.functional.leaky_relu_
    • torch.nn.functional.prelu
    • torch.nn.functional.rrelu
    • torch.nn.functional.rrelu_
    • torch.nn.functional.glu
    • torch.nn.functional.gelu
    • torch.nn.functional.logsigmoid
    • torch.nn.functional.hardshrink
    • torch.nn.functional.tanhshrink
    • torch.nn.functional.softsign
    • torch.nn.functional.softplus
    • torch.nn.functional.softmin
    • torch.nn.functional.softmax
    • torch.nn.functional.softshrink
    • torch.nn.functional.gumbel_softmax
    • torch.nn.functional.log_softmax
    • torch.nn.functional.tanh
    • torch.nn.functional.sigmoid
    • torch.nn.functional.hardsigmoid
    • torch.nn.functional.silu
    • torch.nn.functional.mish
    • torch.nn.functional.batch_norm
    • torch.nn.functional.group_norm
    • torch.nn.functional.instance_norm
    • torch.nn.functional.layer_norm
    • torch.nn.functional.local_response_norm
    • torch.nn.functional.rms_norm
    • torch.nn.functional.normalize
    • torch.nn.functional.linear
    • torch.nn.functional.bilinear
    • torch.nn.functional.dropout
    • torch.nn.functional.alpha_dropout
    • torch.nn.functional.feature_alpha_dropout
    • torch.nn.functional.dropout1d
    • torch.nn.functional.dropout2d
    • torch.nn.functional.dropout3d
    • torch.nn.functional.embedding
    • torch.nn.functional.embedding_bag
    • torch.nn.functional.one_hot
    • torch.nn.functional.pairwise_distance
    • torch.nn.functional.cosine_similarity
    • torch.nn.functional.pdist
    • torch.nn.functional.binary_cross_entropy
    • torch.nn.functional.binary_cross_entropy_with_logits
    • torch.nn.functional.poisson_nll_loss
    • torch.nn.functional.cosine_embedding_loss
    • torch.nn.functional.cross_entropy
    • torch.nn.functional.ctc_loss
    • torch.nn.functional.gaussian_nll_loss
    • torch.nn.functional.hinge_embedding_loss
    • torch.nn.functional.kl_div
    • torch.nn.functional.l1_loss
    • torch.nn.functional.mse_loss
    • torch.nn.functional.margin_ranking_loss
    • torch.nn.functional.multilabel_margin_loss
    • torch.nn.functional.multilabel_soft_margin_loss
    • torch.nn.functional.multi_margin_loss
    • torch.nn.functional.nll_loss
    • torch.nn.functional.huber_loss
    • torch.nn.functional.smooth_l1_loss
    • torch.nn.functional.soft_margin_loss
    • torch.nn.functional.triplet_margin_loss
    • torch.nn.functional.triplet_margin_with_distance_loss
    • torch.nn.functional.pixel_shuffle
    • torch.nn.functional.pixel_unshuffle
    • torch.nn.functional.pad
    • torch.nn.functional.interpolate
    • torch.nn.functional.upsample
    • torch.nn.functional.upsample_nearest
    • torch.nn.functional.upsample_bilinear
    • torch.nn.functional.grid_sample
    • torch.nn.functional.affine_grid
    • torch.nn.functional.torch.nn.parallel.data_parallel
  • torch.Tensor
    • torch.Tensor.new_tensor
    • torch.Tensor.new_full
    • torch.Tensor.new_empty
    • torch.Tensor.new_ones
    • torch.Tensor.new_zeros
    • torch.Tensor.is_cuda
    • torch.Tensor.is_quantized
    • torch.Tensor.is_meta
    • torch.Tensor.device
    • torch.Tensor.grad
    • torch.Tensor.ndim
    • torch.Tensor.real
    • torch.Tensor.imag
    • torch.Tensor.nbytes
    • torch.Tensor.itemsize
    • torch.Tensor.abs
    • torch.Tensor.abs_
    • torch.Tensor.absolute
    • torch.Tensor.absolute_
    • torch.Tensor.acos
    • torch.Tensor.acos_
    • torch.Tensor.arccos
    • torch.Tensor.arccos_
    • torch.Tensor.add
    • torch.Tensor.add_
    • torch.Tensor.addbmm
    • torch.Tensor.addbmm_
    • torch.Tensor.addcdiv
    • torch.Tensor.addcdiv_
    • torch.Tensor.addcmul
    • torch.Tensor.addcmul_
    • torch.Tensor.addmm
    • torch.Tensor.addmm_
    • torch.Tensor.sspaddmm
    • torch.Tensor.addmv
    • torch.Tensor.addmv_
    • torch.Tensor.addr
    • torch.Tensor.addr_
    • torch.Tensor.adjoint
    • torch.Tensor.allclose
    • torch.Tensor.amax
    • torch.Tensor.amin
    • torch.Tensor.aminmax
    • torch.Tensor.angle
    • torch.Tensor.apply_
    • torch.Tensor.argmax
    • torch.Tensor.argmin
    • torch.Tensor.argsort
    • torch.Tensor.argwhere
    • torch.Tensor.asin
    • torch.Tensor.asin_
    • torch.Tensor.arcsin
    • torch.Tensor.arcsin_
    • torch.Tensor.as_strided
    • torch.Tensor.atan
    • torch.Tensor.atan_
    • torch.Tensor.arctan
    • torch.Tensor.arctan_
    • torch.Tensor.atan2
    • torch.Tensor.atan2_
    • torch.Tensor.arctan2
    • torch.Tensor.arctan2_
    • torch.Tensor.all
    • torch.Tensor.any
    • torch.Tensor.backward
    • torch.Tensor.baddbmm
    • torch.Tensor.baddbmm_
    • torch.Tensor.bernoulli
    • torch.Tensor.bernoulli_
    • torch.Tensor.bfloat16
    • torch.Tensor.bincount
    • torch.Tensor.bitwise_not
    • torch.Tensor.bitwise_not_
    • torch.Tensor.bitwise_and
    • torch.Tensor.bitwise_and_
    • torch.Tensor.bitwise_or
    • torch.Tensor.bitwise_or_
    • torch.Tensor.bitwise_xor
    • torch.Tensor.bitwise_xor_
    • torch.Tensor.bitwise_left_shift
    • torch.Tensor.bitwise_left_shift_
    • torch.Tensor.bitwise_right_shift
    • torch.Tensor.bitwise_right_shift_
    • torch.Tensor.bmm
    • torch.Tensor.bool
    • torch.Tensor.byte
    • torch.Tensor.broadcast_to
    • torch.Tensor.cauchy_
    • torch.Tensor.ceil
    • torch.Tensor.ceil_
    • torch.Tensor.char
    • torch.Tensor.cholesky
    • torch.Tensor.cholesky_inverse
    • torch.Tensor.cholesky_solve
    • torch.Tensor.chunk
    • torch.Tensor.clamp
    • torch.Tensor.clamp_
    • torch.Tensor.clip
    • torch.Tensor.clip_
    • torch.Tensor.clone
    • torch.Tensor.contiguous
    • torch.Tensor.copy_
    • torch.Tensor.conj
    • torch.Tensor.conj_physical
    • torch.Tensor.conj_physical_
    • torch.Tensor.resolve_conj
    • torch.Tensor.resolve_neg
    • torch.Tensor.copysign
    • torch.Tensor.copysign_
    • torch.Tensor.cos
    • torch.Tensor.cos_
    • torch.Tensor.cosh
    • torch.Tensor.cosh_
    • torch.Tensor.corrcoef
    • torch.Tensor.count_nonzero
    • torch.Tensor.cov
    • torch.Tensor.acosh
    • torch.Tensor.acosh_
    • torch.Tensor.arccosh
    • torch.Tensor.arccosh_
    • torch.Tensor.cpu
    • torch.Tensor.cross
    • torch.Tensor.cuda
    • torch.Tensor.logcumsumexp
    • torch.Tensor.cummax
    • torch.Tensor.cummin
    • torch.Tensor.cumprod
    • torch.Tensor.cumprod_
    • torch.Tensor.cumsum
    • torch.Tensor.cumsum_
    • torch.Tensor.chalf
    • torch.Tensor.cfloat
    • torch.Tensor.cdouble
    • torch.Tensor.data_ptr
    • torch.Tensor.deg2rad
    • torch.Tensor.dequantize
    • torch.Tensor.det
    • torch.Tensor.dense_dim
    • torch.Tensor.detach
    • torch.Tensor.detach_
    • torch.Tensor.diag
    • torch.Tensor.diag_embed
    • torch.Tensor.diagflat
    • torch.Tensor.diagonal
    • torch.Tensor.diagonal_scatter
    • torch.Tensor.fill_diagonal_
    • torch.Tensor.fmax
    • torch.Tensor.fmin
    • torch.Tensor.diff
    • torch.Tensor.digamma
    • torch.Tensor.digamma_
    • torch.Tensor.dim
    • torch.Tensor.dim_order
    • torch.Tensor.dist
    • torch.Tensor.div
    • torch.Tensor.div_
    • torch.Tensor.divide
    • torch.Tensor.divide_
    • torch.Tensor.dot
    • torch.Tensor.double
    • torch.Tensor.dsplit
    • torch.Tensor.element_size
    • torch.Tensor.eq
    • torch.Tensor.eq_
    • torch.Tensor.equal
    • torch.Tensor.erf
    • torch.Tensor.erf_
    • torch.Tensor.erfc
    • torch.Tensor.erfc_
    • torch.Tensor.erfinv
    • torch.Tensor.erfinv_
    • torch.Tensor.exp
    • torch.Tensor.exp_
    • torch.Tensor.expm1
    • torch.Tensor.expm1_
    • torch.Tensor.expand
    • torch.Tensor.expand_as
    • torch.Tensor.exponential_
    • torch.Tensor.fix
    • torch.Tensor.fix_
    • torch.Tensor.fill_
    • torch.Tensor.flatten
    • torch.Tensor.flip
    • torch.Tensor.fliplr
    • torch.Tensor.flipud
    • torch.Tensor.float
    • torch.Tensor.float_power
    • torch.Tensor.float_power_
    • torch.Tensor.floor
    • torch.Tensor.floor_
    • torch.Tensor.floor_divide
    • torch.Tensor.floor_divide_
    • torch.Tensor.fmod
    • torch.Tensor.fmod_
    • torch.Tensor.frac
    • torch.Tensor.frac_
    • torch.Tensor.frexp
    • torch.Tensor.gather
    • torch.Tensor.gcd
    • torch.Tensor.gcd_
    • torch.Tensor.ge
    • torch.Tensor.ge_
    • torch.Tensor.greater_equal
    • torch.Tensor.greater_equal_
    • torch.Tensor.geometric_
    • torch.Tensor.geqrf
    • torch.Tensor.ger
    • torch.Tensor.get_device
    • torch.Tensor.gt
    • torch.Tensor.gt_
    • torch.Tensor.greater
    • torch.Tensor.greater_
    • torch.Tensor.half
    • torch.Tensor.hardshrink
    • torch.Tensor.heaviside
    • torch.Tensor.histc
    • torch.Tensor.histogram
    • torch.Tensor.hsplit
    • torch.Tensor.hypot
    • torch.Tensor.hypot_
    • torch.Tensor.i0
    • torch.Tensor.i0_
    • torch.Tensor.igamma
    • torch.Tensor.igamma_
    • torch.Tensor.igammac
    • torch.Tensor.igammac_
    • torch.Tensor.index_add_
    • torch.Tensor.index_add
    • torch.Tensor.index_copy_
    • torch.Tensor.index_copy
    • torch.Tensor.index_fill_
    • torch.Tensor.index_fill
    • torch.Tensor.index_put_
    • torch.Tensor.index_put
    • torch.Tensor.index_reduce_
    • torch.Tensor.index_reduce
    • torch.Tensor.index_select
    • torch.Tensor.indices
    • torch.Tensor.inner
    • torch.Tensor.int
    • torch.Tensor.int_repr
    • torch.Tensor.inverse
    • torch.Tensor.isclose
    • torch.Tensor.isfinite
    • torch.Tensor.isinf
    • torch.Tensor.isposinf
    • torch.Tensor.isneginf
    • torch.Tensor.isnan
    • torch.Tensor.is_contiguous
    • torch.Tensor.is_complex
    • torch.Tensor.is_conj
    • torch.Tensor.is_floating_point
    • torch.Tensor.is_inference
    • torch.Tensor.is_leaf
    • torch.Tensor.is_pinned
    • torch.Tensor.is_set_to
    • torch.Tensor.is_shared
    • torch.Tensor.is_signed
    • torch.Tensor.is_sparse
    • torch.Tensor.istft
    • torch.Tensor.isreal
    • torch.Tensor.item
    • torch.Tensor.kthvalue
    • torch.Tensor.lcm
    • torch.Tensor.lcm_
    • torch.Tensor.ldexp
    • torch.Tensor.ldexp_
    • torch.Tensor.le
    • torch.Tensor.le_
    • torch.Tensor.less_equal
    • torch.Tensor.less_equal_
    • torch.Tensor.lerp
    • torch.Tensor.lerp_
    • torch.Tensor.lgamma
    • torch.Tensor.lgamma_
    • torch.Tensor.log
    • torch.Tensor.log_
    • torch.Tensor.logdet
    • torch.Tensor.log10
    • torch.Tensor.log10_
    • torch.Tensor.log1p
    • torch.Tensor.log1p_
    • torch.Tensor.log2
    • torch.Tensor.log2_
    • torch.Tensor.log_normal_
    • torch.Tensor.logaddexp
    • torch.Tensor.logaddexp2
    • torch.Tensor.logsumexp
    • torch.Tensor.logical_and
    • torch.Tensor.logical_and_
    • torch.Tensor.logical_not
    • torch.Tensor.logical_not_
    • torch.Tensor.logical_or
    • torch.Tensor.logical_or_
    • torch.Tensor.logical_xor
    • torch.Tensor.logical_xor_
    • torch.Tensor.logit
    • torch.Tensor.logit_
    • torch.Tensor.long
    • torch.Tensor.lt
    • torch.Tensor.lt_
    • torch.Tensor.less
    • torch.Tensor.less_
    • torch.Tensor.lu
    • torch.Tensor.lu_solve
    • torch.Tensor.as_subclass
    • torch.Tensor.map_
    • torch.Tensor.masked_scatter_
    • torch.Tensor.masked_scatter
    • torch.Tensor.masked_fill_
    • torch.Tensor.masked_fill
    • torch.Tensor.masked_select
    • torch.Tensor.matmul
    • torch.Tensor.matrix_power
    • torch.Tensor.matrix_exp
    • torch.Tensor.max
    • torch.Tensor.maximum
    • torch.Tensor.mean
    • torch.Tensor.module_load
    • torch.Tensor.nanmean
    • torch.Tensor.median
    • torch.Tensor.nanmedian
    • torch.Tensor.min
    • torch.Tensor.minimum
    • torch.Tensor.mm
    • torch.Tensor.smm
    • torch.Tensor.mode
    • torch.Tensor.movedim
    • torch.Tensor.moveaxis
    • torch.Tensor.msort
    • torch.Tensor.mul
    • torch.Tensor.mul_
    • torch.Tensor.multiply
    • torch.Tensor.multiply_
    • torch.Tensor.multinomial
    • torch.Tensor.mv
    • torch.Tensor.mvlgamma
    • torch.Tensor.mvlgamma_
    • torch.Tensor.nansum
    • torch.Tensor.narrow
    • torch.Tensor.narrow_copy
    • torch.Tensor.ndimension
    • torch.Tensor.nan_to_num
    • torch.Tensor.nan_to_num_
    • torch.Tensor.ne
    • torch.Tensor.ne_
    • torch.Tensor.not_equal
    • torch.Tensor.not_equal_
    • torch.Tensor.neg
    • torch.Tensor.neg_
    • torch.Tensor.negative
    • torch.Tensor.negative_
    • torch.Tensor.nelement
    • torch.Tensor.nextafter
    • torch.Tensor.nextafter_
    • torch.Tensor.nonzero
    • torch.Tensor.norm
    • torch.Tensor.normal_
    • torch.Tensor.numel
    • torch.Tensor.numpy
    • torch.Tensor.orgqr
    • torch.Tensor.ormqr
    • torch.Tensor.outer
    • torch.Tensor.permute
    • torch.Tensor.pin_memory
    • torch.Tensor.pinverse
    • torch.Tensor.polygamma
    • torch.Tensor.polygamma_
    • torch.Tensor.positive
    • torch.Tensor.pow
    • torch.Tensor.pow_
    • torch.Tensor.prod
    • torch.Tensor.put_
    • torch.Tensor.qr
    • torch.Tensor.qscheme
    • torch.Tensor.quantile
    • torch.Tensor.nanquantile
    • torch.Tensor.q_scale
    • torch.Tensor.q_zero_point
    • torch.Tensor.q_per_channel_scales
    • torch.Tensor.q_per_channel_zero_points
    • torch.Tensor.q_per_channel_axis
    • torch.Tensor.rad2deg
    • torch.Tensor.random_
    • torch.Tensor.ravel
    • torch.Tensor.reciprocal
    • torch.Tensor.reciprocal_
    • torch.Tensor.record_stream
    • torch.Tensor.register_hook
    • torch.Tensor.register_post_accumulate_grad_hook
    • torch.Tensor.remainder
    • torch.Tensor.remainder_
    • torch.Tensor.renorm
    • torch.Tensor.renorm_
    • torch.Tensor.repeat
    • torch.Tensor.repeat_interleave
    • torch.Tensor.requires_grad
    • torch.Tensor.requires_grad_
    • torch.Tensor.reshape
    • torch.Tensor.reshape_as
    • torch.Tensor.resize_
    • torch.Tensor.resize_as_
    • torch.Tensor.retain_grad
    • torch.Tensor.retains_grad
    • torch.Tensor.roll
    • torch.Tensor.rot90
    • torch.Tensor.round
    • torch.Tensor.round_
    • torch.Tensor.rsqrt
    • torch.Tensor.rsqrt_
    • torch.Tensor.scatter
    • torch.Tensor.scatter_
    • torch.Tensor.scatter_add_
    • torch.Tensor.scatter_add
    • torch.Tensor.scatter_reduce_
    • torch.Tensor.scatter_reduce
    • torch.Tensor.select
    • torch.Tensor.select_scatter
    • torch.Tensor.set_
    • torch.Tensor.share_memory_
    • torch.Tensor.short
    • torch.Tensor.sigmoid
    • torch.Tensor.sigmoid_
    • torch.Tensor.sign
    • torch.Tensor.sign_
    • torch.Tensor.signbit
    • torch.Tensor.sgn
    • torch.Tensor.sgn_
    • torch.Tensor.sin
    • torch.Tensor.sin_
    • torch.Tensor.sinc
    • torch.Tensor.sinc_
    • torch.Tensor.sinh
    • torch.Tensor.sinh_
    • torch.Tensor.asinh
    • torch.Tensor.asinh_
    • torch.Tensor.arcsinh
    • torch.Tensor.arcsinh_
    • torch.Tensor.shape
    • torch.Tensor.size
    • torch.Tensor.slogdet
    • torch.Tensor.slice_scatter
    • torch.Tensor.softmax
    • torch.Tensor.sort
    • torch.Tensor.split
    • torch.Tensor.sparse_mask
    • torch.Tensor.sparse_dim
    • torch.Tensor.sqrt
    • torch.Tensor.sqrt_
    • torch.Tensor.square
    • torch.Tensor.square_
    • torch.Tensor.squeeze
    • torch.Tensor.squeeze_
    • torch.Tensor.std
    • torch.Tensor.stft
    • torch.Tensor.storage
    • torch.Tensor.untyped_storage
    • torch.Tensor.storage_offset
    • torch.Tensor.storage_type
    • torch.Tensor.stride
    • torch.Tensor.sub
    • torch.Tensor.sub_
    • torch.Tensor.subtract
    • torch.Tensor.subtract_
    • torch.Tensor.sum
    • torch.Tensor.sum_to_size
    • torch.Tensor.svd
    • torch.Tensor.swapaxes
    • torch.Tensor.swapdims
    • torch.Tensor.t
    • torch.Tensor.t_
    • torch.Tensor.tensor_split
    • torch.Tensor.tile
    • torch.Tensor.to
    • torch.Tensor.to_mkldnn
    • torch.Tensor.take
    • torch.Tensor.take_along_dim
    • torch.Tensor.tan
    • torch.Tensor.tan_
    • torch.Tensor.tanh
    • torch.Tensor.tanh_
    • torch.Tensor.atanh
    • torch.Tensor.atanh_
    • torch.Tensor.arctanh
    • torch.Tensor.arctanh_
    • torch.Tensor.tolist
    • torch.Tensor.topk
    • torch.Tensor.to_dense
    • torch.Tensor.to_sparse
    • torch.Tensor.to_sparse_csr
    • torch.Tensor.to_sparse_csc
    • torch.Tensor.to_sparse_bsr
    • torch.Tensor.to_sparse_bsc
    • torch.Tensor.trace
    • torch.Tensor.transpose
    • torch.Tensor.transpose_
    • torch.Tensor.triangular_solve
    • torch.Tensor.tril
    • torch.Tensor.tril_
    • torch.Tensor.triu
    • torch.Tensor.triu_
    • torch.Tensor.true_divide
    • torch.Tensor.true_divide_
    • torch.Tensor.trunc
    • torch.Tensor.trunc_
    • torch.Tensor.type
    • torch.Tensor.type_as
    • torch.Tensor.unbind
    • torch.Tensor.unflatten
    • torch.Tensor.unfold
    • torch.Tensor.uniform_
    • torch.Tensor.unique
    • torch.Tensor.unique_consecutive
    • torch.Tensor.unsqueeze
    • torch.Tensor.unsqueeze_
    • torch.Tensor.values
    • torch.Tensor.var
    • torch.Tensor.vdot
    • torch.Tensor.view
    • torch.Tensor.view_as
    • torch.Tensor.vsplit
    • torch.Tensor.where
    • torch.Tensor.xlogy
    • torch.Tensor.xlogy_
    • torch.Tensor.xpu
    • torch.Tensor.zero_
  • Tensor Attributes
  • Tensor Views
  • torch.amp
  • torch.autograd
    • torch.autograd.backward
    • torch.autograd.grad
    • dual_level
    • torch.autograd.forward_ad.make_dual
    • torch.autograd.forward_ad.unpack_dual
    • torch.autograd.forward_ad.enter_dual_level
    • torch.autograd.forward_ad.exit_dual_level
    • UnpackedDualTensor
    • torch.autograd.functional.jacobian
    • torch.autograd.functional.hessian
    • torch.autograd.functional.vjp
    • torch.autograd.functional.jvp
    • torch.autograd.functional.vhp
    • torch.autograd.functional.hvp
    • torch.autograd.Function.forward
    • torch.autograd.Function.backward
    • torch.autograd.Function.jvp
    • torch.autograd.Function.vmap
    • torch.autograd.function.FunctionCtx.mark_dirty
    • torch.autograd.function.FunctionCtx.mark_non_differentiable
    • torch.autograd.function.FunctionCtx.save_for_backward
    • torch.autograd.function.FunctionCtx.set_materialize_grads
    • torch.autograd.function.once_differentiable
    • BackwardCFunction
    • InplaceFunction
    • NestedIOFunction
    • torch.autograd.gradcheck.gradcheck
    • torch.autograd.gradcheck.gradgradcheck
    • torch.autograd.gradcheck.GradcheckError
    • torch.autograd.profiler.profile.export_chrome_trace
    • torch.autograd.profiler.profile.key_averages
    • torch.autograd.profiler.profile.self_cpu_time_total
    • torch.autograd.profiler.profile.total_average
    • torch.autograd.profiler.parse_nvprof_trace
    • EnforceUnique
    • KinetoStepTracker
    • record_function
    • Interval
    • Kernel
    • MemRecordsAcc
    • StringTable
    • torch.autograd.profiler.load_nvprof
    • set_multithreading_enabled
    • torch.autograd.graph.Node.name
    • torch.autograd.graph.Node.metadata
    • torch.autograd.graph.Node.next_functions
    • torch.autograd.graph.Node.register_hook
    • torch.autograd.graph.Node.register_prehook
    • torch.autograd.graph.increment_version
  • torch.library
  • torch.accelerator
    • torch.accelerator.device_count
    • torch.accelerator.is_available
    • torch.accelerator.current_accelerator
    • torch.accelerator.set_device_index
    • torch.accelerator.set_device_idx
    • torch.accelerator.current_device_index
    • torch.accelerator.current_device_idx
    • torch.accelerator.set_stream
    • torch.accelerator.current_stream
    • torch.accelerator.synchronize
    • device_index
  • torch.cpu
    • torch.cpu.current_device
    • torch.cpu.current_stream
    • torch.cpu.is_available
    • torch.cpu.synchronize
    • torch.cpu.stream
    • torch.cpu.set_device
    • torch.cpu.device_count
    • StreamContext
    • Stream
  • torch.cuda
    • StreamContext
    • torch.cuda.can_device_access_peer
    • torch.cuda.current_blas_handle
    • torch.cuda.current_device
    • torch.cuda.current_stream
    • torch.cuda.cudart
    • torch.cuda.default_stream
    • device
    • torch.cuda.device_count
    • torch.cuda.device_memory_used
    • device_of
    • torch.cuda.get_arch_list
    • torch.cuda.get_device_capability
    • torch.cuda.get_device_name
    • torch.cuda.get_device_properties
    • torch.cuda.get_gencode_flags
    • torch.cuda.get_stream_from_external
    • torch.cuda.get_sync_debug_mode
    • torch.cuda.init
    • torch.cuda.ipc_collect
    • torch.cuda.is_available
    • torch.cuda.is_initialized
    • torch.cuda.is_tf32_supported
    • torch.cuda.memory_usage
    • torch.cuda.set_device
    • torch.cuda.set_stream
    • torch.cuda.set_sync_debug_mode
    • torch.cuda.stream
    • torch.cuda.synchronize
    • torch.cuda.utilization
    • torch.cuda.temperature
    • torch.cuda.power_draw
    • torch.cuda.clock_rate
    • torch.cuda.OutOfMemoryError
    • torch.cuda.get_rng_state
    • torch.cuda.get_rng_state_all
    • torch.cuda.set_rng_state
    • torch.cuda.set_rng_state_all
    • torch.cuda.manual_seed
    • torch.cuda.manual_seed_all
    • torch.cuda.seed
    • torch.cuda.seed_all
    • torch.cuda.initial_seed
    • torch.cuda.comm.broadcast
    • torch.cuda.comm.broadcast_coalesced
    • torch.cuda.comm.reduce_add
    • torch.cuda.comm.reduce_add_coalesced
    • torch.cuda.comm.scatter
    • torch.cuda.comm.gather
    • Stream
    • ExternalStream
    • Event
    • torch.cuda.is_current_stream_capturing
    • torch.cuda.graph_pool_handle
    • CUDAGraph
    • graph
    • torch.cuda.make_graphed_callables
    • torch.cuda.empty_cache
    • torch.cuda.get_per_process_memory_fraction
    • torch.cuda.list_gpu_processes
    • torch.cuda.mem_get_info
    • torch.cuda.memory_stats
    • torch.cuda.host_memory_stats
    • torch.cuda.memory_summary
    • torch.cuda.memory_snapshot
    • torch.cuda.memory_allocated
    • torch.cuda.max_memory_allocated
    • torch.cuda.reset_max_memory_allocated
    • torch.cuda.memory_reserved
    • torch.cuda.max_memory_reserved
    • torch.cuda.set_per_process_memory_fraction
    • torch.cuda.memory_cached
    • torch.cuda.max_memory_cached
    • torch.cuda.reset_max_memory_cached
    • torch.cuda.reset_peak_memory_stats
    • torch.cuda.reset_peak_host_memory_stats
    • torch.cuda.caching_allocator_alloc
    • torch.cuda.caching_allocator_delete
    • torch.cuda.get_allocator_backend
    • CUDAPluggableAllocator
    • torch.cuda.change_current_allocator
    • MemPool
    • MemPoolContext
    • torch.cuda.memory.caching_allocator_enable
    • torch.cuda.nvtx.mark
    • torch.cuda.nvtx.range_push
    • torch.cuda.nvtx.range_pop
    • torch.cuda.nvtx.range
    • torch.cuda.jiterator._create_jit_fn
    • torch.cuda.jiterator._create_multi_output_jit_fn
    • TunableOp
    • CUDA Stream Sanitizer
    • torch.cuda.gds.gds_register_buffer
    • torch.cuda.gds.gds_deregister_buffer
    • GdsFile
  • Understanding CUDA Memory Usage
  • torch.mps
    • torch.mps.device_count
    • torch.mps.synchronize
    • torch.mps.get_rng_state
    • torch.mps.set_rng_state
    • torch.mps.manual_seed
    • torch.mps.seed
    • torch.mps.empty_cache
    • torch.mps.set_per_process_memory_fraction
    • torch.mps.current_allocated_memory
    • torch.mps.driver_allocated_memory
    • torch.mps.recommended_max_memory
    • torch.mps.compile_shader
    • torch.mps.profiler.start
    • torch.mps.profiler.stop
    • torch.mps.profiler.profile
    • torch.mps.profiler.is_capturing_metal
    • torch.mps.profiler.is_metal_capture_enabled
    • torch.mps.profiler.metal_capture
    • Event
  • torch.xpu
    • StreamContext
    • torch.xpu.current_device
    • torch.xpu.current_stream
    • device
    • torch.xpu.device_count
    • device_of
    • torch.xpu.get_arch_list
    • torch.xpu.get_device_capability
    • torch.xpu.get_device_name
    • torch.xpu.get_device_properties
    • torch.xpu.get_gencode_flags
    • torch.xpu.get_stream_from_external
    • torch.xpu.init
    • torch.xpu.is_available
    • torch.xpu.is_initialized
    • torch.xpu.set_device
    • torch.xpu.set_stream
    • torch.xpu.stream
    • torch.xpu.synchronize
    • torch.xpu.get_rng_state
    • torch.xpu.get_rng_state_all
    • torch.xpu.initial_seed
    • torch.xpu.manual_seed
    • torch.xpu.manual_seed_all
    • torch.xpu.seed
    • torch.xpu.seed_all
    • torch.xpu.set_rng_state
    • torch.xpu.set_rng_state_all
    • Event
    • Stream
    • torch.xpu.empty_cache
    • torch.xpu.max_memory_allocated
    • torch.xpu.max_memory_reserved
    • torch.xpu.mem_get_info
    • torch.xpu.memory_allocated
    • torch.xpu.memory_reserved
    • torch.xpu.memory_stats
    • torch.xpu.memory_stats_as_nested_dict
    • torch.xpu.reset_accumulated_memory_stats
    • torch.xpu.reset_peak_memory_stats
  • torch.mtia
    • StreamContext
    • torch.mtia.current_device
    • torch.mtia.current_stream
    • torch.mtia.default_stream
    • torch.mtia.device_count
    • torch.mtia.init
    • torch.mtia.is_available
    • torch.mtia.is_initialized
    • torch.mtia.memory_stats
    • torch.mtia.get_device_capability
    • torch.mtia.empty_cache
    • torch.mtia.record_memory_history
    • torch.mtia.snapshot
    • torch.mtia.attach_out_of_memory_observer
    • torch.mtia.set_device
    • torch.mtia.set_stream
    • torch.mtia.stream
    • torch.mtia.synchronize
    • device
    • torch.mtia.set_rng_state
    • torch.mtia.get_rng_state
    • torch.mtia.DeferredMtiaCallError
    • Event
    • Stream
  • torch.mtia.memory
    • torch.mtia.memory.memory_stats
  • Meta device
  • torch.backends
  • torch.export
    • torch.export Programming Model
    • torch.export IR Specification
    • Draft Export
      • Fake tensor
      • Dynamic shapes
      • AOTInductor: Ahead-Of-Time Compilation for Torch.Export-ed Models
        • torch._logging
        • AOTInductor Minifier
    • Writing Graph Transformations on ATen IR
    • IRs
    • ExportDB
      • torch.escape-hatch
      • torch.dynamic-shape
      • torch.cond
      • python.closure
      • torch.dynamic-value
      • python.data-structure
      • python.assert
      • python.control-flow
      • torch.map
      • python.builtin
      • python.object-model
      • python.context-manager
      • torch.operator
      • torch.mutation
    • Control Flow - Cond
    • Dynamo Overview
    • Dynamo Deep-Dive
    • Dynamic shapes
    • Fake tensor
  • torch.distributed
  • torch.distributed.tensor
  • torch.distributed.algorithms.join
  • torch.distributed.elastic
    • Quickstart
    • Train script
    • Examples
    • torchrun (Elastic Launch)
    • Elastic Agent
    • Multiprocessing
    • Error Propagation
    • Rendezvous
    • Expiration Timers
    • Metrics
    • Events
    • Subprocess Handling
    • Control Plane
    • Customization
    • TorchElastic Kubernetes
  • torch.distributed.fsdp
  • torch.distributed.fsdp.fully_shard
  • torch.distributed.tensor.parallel
  • torch.distributed.optim
  • torch.distributed.pipelining
  • torch.distributed.checkpoint
  • torch.distributions
  • torch.compiler
    • Getting Started
    • torch.compiler API reference
      • torch.compiler.compile
      • torch.compiler.reset
      • torch.compiler.allow_in_graph
      • torch.compiler.substitute_in_graph
      • torch.compiler.assume_constant_result
      • torch.compiler.list_backends
      • torch.compiler.disable
      • torch.compiler.set_stance
      • torch.compiler.cudagraph_mark_step_begin
      • torch.compiler.is_compiling
      • torch.compiler.is_dynamo_compiling
      • torch.compiler.is_exporting
    • torch.compiler.config
    • TorchDynamo APIs for fine-grained tracing
    • AOTInductor: Ahead-Of-Time Compilation for Torch.Export-ed Models
      • torch._logging
        • torch._logging.set_logs
      • AOTInductor Minifier
    • TorchInductor GPU Profiling
    • Profiling to understand torch.compile performance
    • Frequently Asked Questions
    • torch.compile Troubleshooting
    • PyTorch 2.0 Performance Dashboard
    • TorchInductor and AOTInductor Provenance Tracking
    • Dynamo Overview
    • Dynamo Deep-Dive
    • Dynamic shapes
    • PyTorch 2.0 NNModule Support
    • Best Practices for Backends
    • CUDAGraph Trees
    • Fake tensor
    • Custom Backends
    • Writing Graph Transformations on ATen IR
    • IRs
  • torch.fft
    • torch.fft.fft
    • torch.fft.ifft
    • torch.fft.fft2
    • torch.fft.ifft2
    • torch.fft.fftn
    • torch.fft.ifftn
    • torch.fft.rfft
    • torch.fft.irfft
    • torch.fft.rfft2
    • torch.fft.irfft2
    • torch.fft.rfftn
    • torch.fft.irfftn
    • torch.fft.hfft
    • torch.fft.ihfft
    • torch.fft.hfft2
    • torch.fft.ihfft2
    • torch.fft.hfftn
    • torch.fft.ihfftn
    • torch.fft.fftfreq
    • torch.fft.rfftfreq
    • torch.fft.fftshift
    • torch.fft.ifftshift
  • torch.func
    • torch.func Whirlwind Tour
    • torch.func API Reference
      • torch.func.vmap
      • torch.func.grad
      • torch.func.grad_and_value
      • torch.func.vjp
      • torch.func.jvp
      • torch.func.linearize
      • torch.func.jacrev
      • torch.func.jacfwd
      • torch.func.hessian
      • torch.func.functionalize
      • torch.func.functional_call
      • torch.func.stack_module_state
      • torch.func.replace_all_batch_norm_modules_
      • Patching Batch Norm
      • torch.func.debug_unwrap
    • UX Limitations
    • Migrating from functorch to torch.func
  • torch.futures
  • torch.fx
  • torch.fx.experimental
    • ShapeEnv
    • DimDynamic
    • StrictMinMaxConstraint
    • RelaxedUnspecConstraint
    • EqualityConstraint
    • SymbolicContext
    • StatelessSymbolicContext
    • StatefulSymbolicContext
    • SubclassSymbolicContext
    • DimConstraints
    • ShapeEnvSettings
    • ConvertIntKey
    • CallMethodKey
    • PropagateUnbackedSymInts
    • DivideByKey
    • InnerTensorKey
    • torch.fx.experimental.symbolic_shapes.hint_int
    • torch.fx.experimental.symbolic_shapes.is_concrete_int
    • torch.fx.experimental.symbolic_shapes.is_concrete_bool
    • torch.fx.experimental.symbolic_shapes.is_concrete_float
    • torch.fx.experimental.symbolic_shapes.has_free_symbols
    • torch.fx.experimental.symbolic_shapes.has_free_unbacked_symbols
    • torch.fx.experimental.symbolic_shapes.guard_or_true
    • torch.fx.experimental.symbolic_shapes.guard_or_false
    • torch.fx.experimental.symbolic_shapes.guard_size_oblivious
    • torch.fx.experimental.symbolic_shapes.sym_and
    • torch.fx.experimental.symbolic_shapes.sym_eq
    • torch.fx.experimental.symbolic_shapes.sym_or
    • torch.fx.experimental.symbolic_shapes.constrain_range
    • torch.fx.experimental.symbolic_shapes.constrain_unify
    • torch.fx.experimental.symbolic_shapes.canonicalize_bool_expr
    • torch.fx.experimental.symbolic_shapes.statically_known_true
    • torch.fx.experimental.symbolic_shapes.has_static_value
    • torch.fx.experimental.symbolic_shapes.lru_cache
    • torch.fx.experimental.symbolic_shapes.check_consistent
    • torch.fx.experimental.symbolic_shapes.compute_unbacked_bindings
    • torch.fx.experimental.symbolic_shapes.rebind_unbacked
    • torch.fx.experimental.symbolic_shapes.resolve_unbacked_bindings
    • torch.fx.experimental.symbolic_shapes.is_accessor_node
    • torch.fx.experimental.proxy_tensor.make_fx
    • torch.fx.experimental.proxy_tensor.handle_sym_dispatch
    • torch.fx.experimental.proxy_tensor.get_proxy_mode
    • torch.fx.experimental.proxy_tensor.maybe_enable_thunkify
    • torch.fx.experimental.proxy_tensor.maybe_disable_thunkify
  • torch.hub
  • torch.jit
    • torch.jit.supported_ops
    • TorchScript Language Reference
    • TorchScript Language Reference
    • torch.jit.script
    • torch.jit.trace
    • torch.jit.script_if_tracing
    • torch.jit.trace_module
    • torch.jit.fork
    • torch.jit.wait
    • ScriptModule
    • ScriptFunction
    • torch.jit.freeze
    • torch.jit.optimize_for_inference
    • torch.jit.enable_onednn_fusion
    • torch.jit.onednn_fusion_enabled
    • torch.jit.set_fusion_strategy
    • strict_fusion
    • torch.jit.save
    • torch.jit.load
    • torch.jit.ignore
    • torch.jit.unused
    • torch.jit.interface
    • torch.jit.isinstance
    • Attribute
    • torch.jit.annotate
    • Python Language Reference Coverage
    • TorchScript Unsupported PyTorch Constructs
  • torch.linalg
    • torch.linalg.norm
    • torch.linalg.vector_norm
    • torch.linalg.matrix_norm
    • torch.linalg.diagonal
    • torch.linalg.det
    • torch.linalg.slogdet
    • torch.linalg.cond
    • torch.linalg.matrix_rank
    • torch.linalg.cholesky
    • torch.linalg.qr
    • torch.linalg.lu
    • torch.linalg.lu_factor
    • torch.linalg.eig
    • torch.linalg.eigvals
    • torch.linalg.eigh
    • torch.linalg.eigvalsh
    • torch.linalg.svd
    • torch.linalg.svdvals
    • torch.linalg.solve
    • torch.linalg.solve_triangular
    • torch.linalg.lu_solve
    • torch.linalg.lstsq
    • torch.linalg.inv
    • torch.linalg.pinv
    • torch.linalg.matrix_exp
    • torch.linalg.matrix_power
    • torch.linalg.cross
    • torch.linalg.matmul
    • torch.linalg.vecdot
    • torch.linalg.multi_dot
    • torch.linalg.householder_product
    • torch.linalg.tensorinv
    • torch.linalg.tensorsolve
    • torch.linalg.vander
    • torch.linalg.cholesky_ex
    • torch.linalg.inv_ex
    • torch.linalg.solve_ex
    • torch.linalg.lu_factor_ex
    • torch.linalg.ldl_factor
    • torch.linalg.ldl_factor_ex
    • torch.linalg.ldl_solve
  • torch.monitor
  • torch.signal
    • torch.signal.windows.bartlett
    • torch.signal.windows.blackman
    • torch.signal.windows.cosine
    • torch.signal.windows.exponential
    • torch.signal.windows.gaussian
    • torch.signal.windows.general_cosine
    • torch.signal.windows.general_hamming
    • torch.signal.windows.hamming
    • torch.signal.windows.hann
    • torch.signal.windows.kaiser
    • torch.signal.windows.nuttall
  • torch.special
  • torch.overrides
  • torch.package
  • torch.profiler
  • torch.nn.init
  • torch.nn.attention
    • torch.nn.attention.sdpa_kernel
    • SDPBackend
    • torch.nn.attention.flex_attention
    • torch.nn.attention.bias
    • torch.nn.attention.experimental
  • torch.onnx
    • TorchDynamo-based ONNX Exporter
      • Understanding TorchDynamo-based ONNX Exporter Memory Usage
    • torch.onnx.ops
    • torch.onnx.verification
    • ONNX Backend for TorchDynamo
    • TorchScript-based ONNX Exporter
      • JitScalarType
  • torch.optim
    • torch.optim.Optimizer.add_param_group
    • torch.optim.Optimizer.load_state_dict
    • torch.optim.Optimizer.register_load_state_dict_pre_hook
    • torch.optim.Optimizer.register_load_state_dict_post_hook
    • torch.optim.Optimizer.state_dict
    • torch.optim.Optimizer.register_state_dict_pre_hook
    • torch.optim.Optimizer.register_state_dict_post_hook
    • torch.optim.Optimizer.step
    • torch.optim.Optimizer.register_step_pre_hook
    • torch.optim.Optimizer.register_step_post_hook
    • torch.optim.Optimizer.zero_grad
    • Adadelta
    • Adafactor
    • Adagrad
    • Adam
    • AdamW
    • SparseAdam
    • Adamax
    • ASGD
    • LBFGS
    • NAdam
    • RAdam
    • RMSprop
    • Rprop
    • SGD
    • LRScheduler
    • LambdaLR
    • MultiplicativeLR
    • StepLR
    • MultiStepLR
    • ConstantLR
    • LinearLR
    • ExponentialLR
    • PolynomialLR
    • CosineAnnealingLR
    • ChainedScheduler
    • SequentialLR
    • ReduceLROnPlateau
    • CyclicLR
    • OneCycleLR
    • CosineAnnealingWarmRestarts
    • AveragedModel
    • SWALR
  • Complex Numbers
  • DDP Communication Hooks
  • Quantization
    • Quantization API Reference
      • quantize
      • quantize_dynamic
      • quantize_qat
      • prepare
      • prepare_qat
      • convert
      • fuse_modules
      • QuantStub
      • DeQuantStub
      • QuantWrapper
      • add_quant_dequant
      • swap_module
      • propagate_qconfig
      • default_eval_fn
      • prepare_fx
      • prepare_qat_fx
      • convert_fx
      • fuse_fx
      • QConfigMapping
      • get_default_qconfig_mapping
      • get_default_qat_qconfig_mapping
      • BackendConfig
      • BackendPatternConfig
      • DTypeConfig
      • DTypeWithConstraints
      • ObservationType
      • FuseCustomConfig
      • PrepareCustomConfig
      • ConvertCustomConfig
      • StandaloneModuleConfigEntry
      • model_is_exported
      • lower_pt2e_quantized_to_x86
      • generate_numeric_debug_handle
      • CUSTOM_KEY
      • NUMERIC_DEBUG_HANDLE_KEY
      • prepare_for_propagation_comparison
      • extract_results_from_loggers
      • compare_results
      • torch.quantize_per_tensor
      • torch.quantize_per_channel
      • torch.dequantize
      • torch.Tensor.view
      • torch.Tensor.as_strided
      • torch.Tensor.expand
      • torch.Tensor.flatten
      • torch.Tensor.select
      • torch.Tensor.ne
      • torch.Tensor.eq
      • torch.Tensor.ge
      • torch.Tensor.le
      • torch.Tensor.gt
      • torch.Tensor.lt
      • torch.Tensor.copy_
      • torch.Tensor.clone
      • torch.Tensor.dequantize
      • torch.Tensor.equal
      • torch.Tensor.int_repr
      • torch.Tensor.max
      • torch.Tensor.mean
      • torch.Tensor.min
      • torch.Tensor.q_scale
      • torch.Tensor.q_zero_point
      • torch.Tensor.q_per_channel_scales
      • torch.Tensor.q_per_channel_zero_points
      • torch.Tensor.q_per_channel_axis
      • torch.Tensor.resize_
      • torch.Tensor.sort
      • torch.Tensor.topk
      • ObserverBase
      • MinMaxObserver
      • MovingAverageMinMaxObserver
      • PerChannelMinMaxObserver
      • MovingAveragePerChannelMinMaxObserver
      • HistogramObserver
      • PlaceholderObserver
      • RecordingObserver
      • NoopObserver
      • get_observer_state_dict
      • load_observer_state_dict
      • default_observer
      • default_placeholder_observer
      • default_debug_observer
      • default_weight_observer
      • default_histogram_observer
      • default_per_channel_weight_observer
      • default_dynamic_quant_observer
      • default_float_qparams_observer
      • AffineQuantizedObserverBase
      • Granularity
      • MappingType
      • PerAxis
      • PerBlock
      • PerGroup
      • PerRow
      • PerTensor
      • PerToken
      • TorchAODType
      • ZeroPointDomain
      • get_block_size
      • FakeQuantizeBase
      • FakeQuantize
      • FixedQParamsFakeQuantize
      • FusedMovingAvgObsFakeQuantize
      • default_fake_quant
      • default_weight_fake_quant
      • default_per_channel_weight_fake_quant
      • default_histogram_fake_quant
      • default_fused_act_fake_quant
      • default_fused_wt_fake_quant
      • default_fused_per_channel_wt_fake_quant
      • disable_fake_quant
      • enable_fake_quant
      • disable_observer
      • enable_observer
      • QConfig
      • default_qconfig
      • default_debug_qconfig
      • default_per_channel_qconfig
      • default_dynamic_qconfig
      • float16_dynamic_qconfig
      • float16_static_qconfig
      • per_channel_dynamic_qconfig
      • float_qparams_weight_only_qconfig
      • default_qat_qconfig
      • default_weight_only_qconfig
      • default_activation_only_qconfig
      • default_qat_qconfig_v2
      • ConvReLU1d
      • ConvReLU2d
      • ConvReLU3d
      • LinearReLU
      • ConvBn1d
      • ConvBn2d
      • ConvBn3d
      • ConvBnReLU1d
      • ConvBnReLU2d
      • ConvBnReLU3d
      • BNReLU2d
      • BNReLU3d
      • LinearReLU
      • ConvBn1d
      • ConvBnReLU1d
      • ConvBn2d
      • ConvBnReLU2d
      • ConvReLU2d
      • ConvBn3d
      • ConvBnReLU3d
      • ConvReLU3d
      • update_bn_stats
      • freeze_bn_stats
      • BNReLU2d
      • BNReLU3d
      • ConvReLU1d
      • ConvReLU2d
      • ConvReLU3d
      • LinearReLU
      • LinearReLU
      • Conv2d
      • Conv3d
      • Linear
      • Linear
      • ReLU6
      • Hardswish
      • ELU
      • LeakyReLU
      • Sigmoid
      • BatchNorm2d
      • BatchNorm3d
      • Conv1d
      • Conv2d
      • Conv3d
      • ConvTranspose1d
      • ConvTranspose2d
      • ConvTranspose3d
      • Embedding
      • EmbeddingBag
      • FloatFunctional
      • FXFloatFunctional
      • QFunctional
      • Linear
      • LayerNorm
      • GroupNorm
      • InstanceNorm1d
      • InstanceNorm2d
      • InstanceNorm3d
      • avg_pool2d
      • avg_pool3d
      • adaptive_avg_pool2d
      • adaptive_avg_pool3d
      • conv1d
      • conv2d
      • conv3d
      • interpolate
      • linear
      • max_pool1d
      • max_pool2d
      • celu
      • leaky_relu
      • hardtanh
      • hardswish
      • threshold
      • elu
      • hardsigmoid
      • clamp
      • upsample
      • upsample_bilinear
      • upsample_nearest
      • LSTM
      • MultiheadAttention
      • Linear
      • LSTM
      • GRU
      • RNNCell
      • LSTMCell
      • GRUCell
    • Quantization Backend Configuration
    • Quantization Accuracy Debugging
  • Distributed RPC Framework
    • Remote Reference Protocol
    • Distributed Autograd Design
  • torch.random
  • torch.masked
    • torch.abs
    • torch.absolute
    • torch.acos
    • torch.arccos
    • torch.acosh
    • torch.arccosh
    • torch.angle
    • torch.asin
    • torch.arcsin
    • torch.asinh
    • torch.arcsinh
    • torch.atan
    • torch.arctan
    • torch.atanh
    • torch.arctanh
    • torch.bitwise_not
    • torch.ceil
    • torch.clamp
    • torch.clip
    • torch.conj_physical
    • torch.cos
    • torch.cosh
    • torch.deg2rad
    • torch.digamma
    • torch.erf
    • torch.erfc
    • torch.erfinv
    • torch.exp
    • torch.exp2
    • torch.expm1
    • torch.fix
    • torch.floor
    • torch.frac
    • torch.lgamma
    • torch.log
    • torch.log10
    • torch.log1p
    • torch.log2
    • torch.logit
    • torch.i0
    • torch.isnan
    • torch.nan_to_num
    • torch.neg
    • torch.negative
    • torch.positive
    • torch.pow
    • torch.rad2deg
    • torch.reciprocal
    • torch.round
    • torch.rsqrt
    • torch.sigmoid
    • torch.sign
    • torch.sgn
    • torch.signbit
    • torch.sin
    • torch.sinc
    • torch.sinh
    • torch.sqrt
    • torch.square
    • torch.tan
    • torch.tanh
    • torch.trunc
    • torch.angle
    • torch.positive
    • torch.signbit
    • torch.isnan
    • torch.add
    • torch.atan2
    • torch.arctan2
    • torch.bitwise_and
    • torch.bitwise_or
    • torch.bitwise_xor
    • torch.bitwise_left_shift
    • torch.bitwise_right_shift
    • torch.div
    • torch.divide
    • torch.floor_divide
    • torch.fmod
    • torch.logaddexp
    • torch.logaddexp2
    • torch.mul
    • torch.multiply
    • torch.nextafter
    • torch.remainder
    • torch.sub
    • torch.subtract
    • torch.true_divide
    • torch.eq
    • torch.ne
    • torch.le
    • torch.ge
    • torch.greater
    • torch.greater_equal
    • torch.gt
    • torch.less_equal
    • torch.lt
    • torch.less
    • torch.maximum
    • torch.minimum
    • torch.fmax
    • torch.fmin
    • torch.not_equal
    • torch.logaddexp
    • torch.logaddexp2
    • torch.equal
    • torch.fmin
    • torch.minimum
    • torch.fmax
    • torch.sum
    • torch.mean
    • torch.amin
    • torch.amax
    • torch.argmin
    • torch.argmax
    • torch.prod
    • torch.all
    • torch.norm
    • torch.var
    • torch.std
    • torch.atleast_1d
    • torch.broadcast_tensors
    • torch.broadcast_to
    • torch.cat
    • torch.chunk
    • torch.column_stack
    • torch.dsplit
    • torch.flatten
    • torch.hsplit
    • torch.hstack
    • torch.kron
    • torch.meshgrid
    • torch.narrow
    • torch.nn.functional.unfold
    • torch.ravel
    • torch.select
    • torch.split
    • torch.stack
    • torch.t
    • torch.transpose
    • torch.vsplit
    • torch.vstack
    • torch.Tensor.expand
    • torch.Tensor.expand_as
    • torch.Tensor.reshape
    • torch.Tensor.reshape_as
    • torch.Tensor.unfold
    • torch.Tensor.view
  • torch.nested
  • torch.Size
  • torch.sparse
    • torch.Tensor.is_sparse
    • torch.Tensor.is_sparse_csr
    • torch.Tensor.dense_dim
    • torch.Tensor.sparse_dim
    • torch.Tensor.sparse_mask
    • torch.Tensor.to_sparse
    • torch.Tensor.to_sparse_coo
    • torch.Tensor.to_sparse_csr
    • torch.Tensor.to_sparse_csc
    • torch.Tensor.to_sparse_bsr
    • torch.Tensor.to_sparse_bsc
    • torch.Tensor.to_dense
    • torch.Tensor.values
    • torch.Tensor.coalesce
    • torch.Tensor.sparse_resize_
    • torch.Tensor.sparse_resize_and_clear_
    • torch.Tensor.is_coalesced
    • torch.Tensor.indices
    • torch.Tensor.crow_indices
    • torch.Tensor.col_indices
    • torch.Tensor.row_indices
    • torch.Tensor.ccol_indices
    • torch.sparse_coo_tensor
    • torch.sparse_csr_tensor
    • torch.sparse_csc_tensor
    • torch.sparse_bsr_tensor
    • torch.sparse_bsc_tensor
    • torch.sparse_compressed_tensor
    • torch.sparse.sum
    • torch.sparse.addmm
    • torch.sparse.sampled_addmm
    • torch.sparse.mm
    • torch.sspaddmm
    • torch.hspmm
    • torch.smm
    • torch.sparse.softmax
    • torch.sparse.spsolve
    • torch.sparse.log_softmax
    • torch.sparse.spdiags
    • check_sparse_tensor_invariants
    • torch.sparse.as_sparse_gradcheck
  • torch.Storage
  • torch.testing
  • torch.utils
    • torch.utils.rename_privateuse1_backend
    • torch.utils.generate_methods_for_privateuse1_backend
    • torch.utils.get_cpp_backtrace
    • torch.utils.set_module
    • torch.utils.swap_tensors
  • torch.utils.benchmark
  • torch.utils.bottleneck
  • torch.utils.checkpoint
  • torch.utils.cpp_extension
  • torch.utils.data
  • torch.utils.deterministic
  • torch.utils.jit
  • torch.utils.dlpack
  • torch.utils.mobile_optimizer
  • torch.utils.model_zoo
  • torch.utils.tensorboard
  • torch.utils.module_tracker
  • Type Info
  • Named Tensors
  • Named Tensors operator coverage
  • torch.__config__
  • torch.__future__
  • torch._logging
    • torch._logging.set_logs
  • Torch Environment Variables
    • Threading Environment Variables
    • CUDA Environment Variables
    • MPS Environment Variables
    • Debugging Environment Variables
    • Miscellaneous Environment Variables
    • torch._logging
      • torch._logging.set_logs
    • PYTORCH ProcessGroupNCCL Environment Variables
  • Python API
  • Automatic differentiation package - torch.autograd
  • torch.autogr...
Rate this Page
★ ★ ★ ★ ★

torch.autograd.function.once_differentiable#

torch.autograd.function.once_differentiable(fn)[source]#
Rate this Page
★ ★ ★ ★ ★

previous

torch.autograd.function.FunctionCtx.set_materialize_grads

next

BackwardCFunction

© Copyright PyTorch Contributors.

Built with the PyData Sphinx Theme 0.15.4.

previous

torch.autograd.function.FunctionCtx.set_materialize_grads

next

BackwardCFunction

On this page
  • once_differentiable()
Show Source
PyTorch Community
  • PyTorch Governance
  • PyTorch Design Philosophy
  • The Ultimate Guide to PyTorch Contributions
Language Bindings
  • C++
  • Javadoc
  • torch.multiply
PyTorch Libraries
  • torchaudio
  • torchao
  • ExecuTorch
  • torchrec
  • torchserve
  • torchdata
  • torchvision
  • PyTorch on XLA Devices

Docs

Access comprehensive developer documentation for PyTorch

View Docs

Tutorials

Get in-depth tutorials for beginners and advanced developers

View Tutorials

Resources

Find development resources and get your questions answered

View Resources
  • PyTorch
  • Get Started
  • Features
  • Ecosystem
  • Blog
  • Contributing
  • Resources
  • Tutorials
  • Docs
  • Discuss
  • Github Issues
  • Brand Guidelines
  • Stay up to date
  • Facebook
  • Twitter
  • YouTube
  • LinkedIn
  • PyTorch Podcasts
  • Spotify
  • Apple
  • Google
  • Amazon
  • Terms
  • |
  • Privacy

© Copyright The Linux Foundation. The PyTorch Foundation is a project of The Linux Foundation. For web site terms of use, trademark policy and other policies applicable to The PyTorch Foundation please see www.linuxfoundation.org/policies/. The PyTorch Foundation supports the PyTorch open source project, which has been established as PyTorch Project a Series of LF Projects, LLC. For policies applicable to the PyTorch Project a Series of LF Projects, LLC, please see www.lfprojects.org/policies/.

To analyze traffic and optimize your experience, we serve cookies on this site. By clicking or navigating, you agree to allow our usage of cookies. As the current maintainers of this site, Facebook’s Cookies Policy applies. Learn more, including about available controls: Cookies Policy.

To analyze traffic and optimize your experience, we serve cookies on this site. By clicking or navigating, you agree to allow our usage of cookies. As the current maintainers of this site, Facebook’s Cookies Policy applies. Learn more, including about available controls: Cookies Policy.

© Copyright PyTorch Contributors.

Created using Sphinx 5.3.0.

Built with the PyData Sphinx Theme 0.15.4.