List of all items
Structs
- AdaptiveAvgPool2dLayer
- AdaptiveMaxPool2dLayer
- AllReduceAggregator
- AnchorFreeHead
- ArchitectureConfig
- AvgPool2dLayer
- Batch
- BatchCollector
- BatchIterOptions
- BatchNorm2dLayer
- BestModelCheckpoint
- CenterCrop
- CnnTrainConfig
- Compose
- CompressedGradient
- Conv1dLayer
- Conv2dLayer
- Conv3dLayer
- ConvTranspose2dLayer
- CrossAttention
- CutMixConfig
- DataLoader
- DataLoaderBatch
- DataLoaderConfig
- DataLoaderIter
- DataParallelConfig
- DatasetSplit
- DeformableConv2dLayer
- DepthwiseConv2dLayer
- DistributedConfig
- DropoutLayer
- DynamicBatchConfig
- DynamicLossScaler
- EarlyStopping
- EmbeddingLayer
- EpochMetrics
- EpochTrainOptions
- ExponentialMovingAverage
- FeedForward
- FeedForwardLayer
- FlattenLayer
- FpnNeck
- GELULayer
- GaussianBlur
- GlobalAvgPool2dLayer
- GroupNormLayer
- GruCell
- GruLayer
- HubEntry
- ImageAugmentationPipeline
- InProcessTransport
- InferencePipeline
- InstanceNormLayer
- LayerNormLayer
- LeakyReLULayer
- LinearLayer
- LocalAggregator
- LoraConfig
- LoraLinear
- LrFinderConfig
- LrFinderResult
- LstmCell
- LstmLayer
- MaskHead
- MaxPool2dLayer
- MbConvBlock
- MetricsLogger
- MiniBatchIter
- MishLayer
- MixUpConfig
- MixedPrecisionConfig
- ModelHub
- ModelZoo
- MultiHeadAttention
- MultiHeadAttentionConfig
- MultiHeadAttentionLayer
- Normalize
- PReLULayer
- ParameterServer
- PatchEmbedding
- PerChannelQuantResult
- PermuteDims
- PipelineParallelConfig
- PipelineStage
- PixelShuffleLayer
- PrunedTensor
- QuantizedTensor
- RandomHorizontalFlip
- RandomSampler
- ReLULayer
- ResidualBlock
- Resize
- RnnCell
- RnnLayer
- SafeTensorFile
- ScaleValues
- ScheduledEpochMetrics
- SchedulerTrainOptions
- SeparableConv2dLayer
- SequentialCheckpoint
- SequentialModel
- SequentialSampler
- SiLULayer
- SigmoidLayer
- SoftmaxLayer
- SqueezeExciteBlock
- StreamingDataLoader
- SupervisedCsvConfig
- SupervisedDataset
- SupervisedImageFolderConfig
- SupervisedImageFolderLoadResult
- SupervisedImageManifestConfig
- SupervisedJsonlConfig
- TanhLayer
- TcpAllReduceAggregator
- TcpTransport
- TensorBoardCallback
- TensorBoardWriter
- TensorInfo
- TensorSnapshot
- TopKCompressor
- TrainResult
- Trainer
- TrainerConfig
- TrainingLog
- TransformerDecoder
- TransformerDecoderBlock
- TransformerEncoderBlock
- TransformerEncoderLayer
- UNetDecoderStage
- UNetEncoderStage
- UpsampleLayer
- VisionTransformer
- WeightedRandomSampler
- tcp_transport::TcpAllReduceAggregator
- tcp_transport::TcpTransport
Enums
- ImageAugmentationOp
- ImageFolderTargetMode
- LayerCheckpoint
- LossKind
- ModelArchitecture
- ModelError
- ModelLayer
- MonitorMode
- NodeRole
- OptimizerKind
- OptimizerType
- QuantMode
- SafeTensorDType
- SamplingPolicy
- SupervisedLoss
- tcp_transport::NodeRole
Traits
Functions
- accumulate_gradients
- adam_state_from_map
- adam_state_to_map
- add_bottleneck_block
- add_residual_block
- apply_pruning_mask
- batched_inference
- bce_loss
- bilstm_forward_sequence
- build_alexnet
- build_classifier
- build_feature_extractor
- build_mobilenet_v2
- build_resnet
- build_resnet_custom
- build_resnet_feature_extractor
- build_simple_cnn_classifier
- build_vgg
- cast_params_for_forward
- cast_to_master
- checkpoint_from_json
- checkpoint_to_json
- collect_gradients
- compress_gradients
- constant
- contrastive_loss
- cosine_embedding_loss
- cross_entropy_loss
- ctc_loss
- decompress_gradients
- default_cache_dir
- dequantize_weights
- dice_loss
- distillation_loss
- distributed_train_step
- export_sequential_to_onnx
- export_sequential_to_onnx_file
- focal_loss
- fuse_conv_bn
- gather_shards
- generate_causal_mask
- generate_padding_mask
- gru_forward_sequence
- hinge_loss
- huber_loss
- infer_batch
- infer_batch_graph
- inspect_weights
- kaiming_normal
- kaiming_uniform
- kl_div_loss
- label_smoothing_cross_entropy
- load_state_dict
- load_supervised_dataset_csv_file
- load_supervised_dataset_jsonl_file
- load_supervised_image_folder_dataset
- load_supervised_image_folder_dataset_with_classes
- load_supervised_image_manifest_csv_file
- load_training_checkpoint
- load_weights
- loopback_pair
- lr_range_test
- lstm_forward_sequence
- mae_loss
- mixed_precision_train_step
- mse_loss
- nll_loss
- optimize_sequential
- orthogonal
- parse_supervised_dataset_csv
- parse_supervised_dataset_jsonl
- parse_supervised_image_manifest_csv
- prune_magnitude
- quantize_per_channel
- quantize_weights
- quantized_matmul
- remap_state_dict
- rnn_forward_sequence
- save_training_checkpoint
- save_weights
- scale_gradients
- scaled_dot_product_attention
- sgd_state_from_map
- sgd_state_to_map
- shard_tensor
- smooth_l1_loss
- split_into_stages
- tcp_transport::loopback_pair
- timm_to_yscv_name
- train_cnn_epoch_adam
- train_cnn_epoch_adamw
- train_cnn_epoch_sgd
- train_cnn_epochs
- train_epoch_adam
- train_epoch_adam_with_loss
- train_epoch_adam_with_options
- train_epoch_adam_with_options_and_loss
- train_epoch_adamw
- train_epoch_adamw_with_loss
- train_epoch_adamw_with_options
- train_epoch_adamw_with_options_and_loss
- train_epoch_distributed
- train_epoch_distributed_sgd
- train_epoch_rmsprop
- train_epoch_rmsprop_with_loss
- train_epoch_rmsprop_with_options
- train_epoch_rmsprop_with_options_and_loss
- train_epoch_sgd
- train_epoch_sgd_with_loss
- train_epoch_sgd_with_options
- train_epoch_sgd_with_options_and_loss
- train_epochs_adam_with_scheduler
- train_epochs_adam_with_scheduler_and_loss
- train_epochs_adamw_with_scheduler
- train_epochs_adamw_with_scheduler_and_loss
- train_epochs_rmsprop_with_scheduler
- train_epochs_rmsprop_with_scheduler_and_loss
- train_epochs_sgd_with_scheduler
- train_epochs_sgd_with_scheduler_and_loss
- train_epochs_with_callbacks
- train_step_adam
- train_step_adam_with_accumulation
- train_step_adam_with_loss
- train_step_adamw
- train_step_adamw_with_accumulation
- train_step_adamw_with_loss
- train_step_rmsprop
- train_step_rmsprop_with_accumulation
- train_step_rmsprop_with_loss
- train_step_sgd
- train_step_sgd_with_accumulation
- train_step_sgd_with_loss
- triplet_loss
- xavier_normal
- xavier_uniform