TorchOpt
stable
Documentation
Basics
Optimizers
Explicit Gradient Differentiation
Implicit Gradient Differentiation
Zero-order Gradient Differentiation
Distributed Training
Visualization
Tutorial Notebooks
Get Started with Jupyter Notebook
Examples
Model-Agnostic Meta-Learning
Developer Documentation
Contributing to TorchOpt
Contributor
API Documentation
TorchOpt Optimizer
Functional Optimizers
Classic Optimizers
Differentiable Meta-Optimizers
Implicit Differentiation
Linear System Solvers
Zero-Order Differentiation
Optimizer Hooks
Gradient Transformation
Optimizer Schedules
Apply Parameter Updates
Combining Optimizers
Distributed Utilities
General Utilities
Visualizing Gradient Flow
TorchOpt
Index
Edit on GitHub
Index
_
|
A
|
B
|
C
|
D
|
E
|
F
|
G
|
I
|
L
|
M
|
N
|
O
|
P
|
R
|
S
|
Z
_
__abstractmethods__ (torchopt.diff.implicit.nn.ImplicitMetaGradientModule attribute)
(torchopt.diff.zero_order.nn.ZeroOrderGradientModule attribute)
__annotations__ (torchopt.diff.implicit.nn.ImplicitMetaGradientModule attribute)
(torchopt.diff.zero_order.nn.ZeroOrderGradientModule attribute)
__init__() (torchopt.FuncOptimizer method)
(torchopt.MetaOptimizer method)
(torchopt.Optimizer method)
__init_subclass__() (torchopt.diff.implicit.nn.ImplicitMetaGradientModule class method)
(torchopt.diff.zero_order.nn.ZeroOrderGradientModule class method)
__parameters__ (torchopt.diff.zero_order.nn.ZeroOrderGradientModule attribute)
__subclasshook__() (torchopt.diff.zero_order.nn.ZeroOrderGradientModule method)
A
AdaDelta (class in torchopt)
adadelta() (in module torchopt)
AdaGrad (class in torchopt)
adagrad() (in module torchopt)
Adam (class in torchopt)
adam() (in module torchopt)
AdaMax (class in torchopt)
adamax() (in module torchopt)
AdamW (class in torchopt)
adamw() (in module torchopt)
add_param_group() (torchopt.MetaOptimizer method)
(torchopt.Optimizer method)
apply_updates() (in module torchopt)
auto_init_rpc() (in module torchopt.distributed)
B
backward() (in module torchopt.distributed.autograd)
barrier() (in module torchopt.distributed)
batch_partitioner() (in module torchopt.distributed)
C
chain() (in module torchopt.combine)
clip_grad_norm() (in module torchopt)
context() (in module torchopt.distributed.autograd)
custom_root() (in module torchopt.diff.implicit)
D
dim_partitioner() (in module torchopt.distributed)
E
extract_state_dict() (in module torchopt)
F
forward() (torchopt.diff.zero_order.nn.ZeroOrderGradientModule method)
FuncOptimizer (class in torchopt)
G
get_gradients() (in module torchopt.distributed.autograd)
get_local_rank() (in module torchopt.distributed)
get_local_world_size() (in module torchopt.distributed)
get_rank() (in module torchopt.distributed)
get_worker_id() (in module torchopt.distributed)
get_world_info() (in module torchopt.distributed)
get_world_rank() (in module torchopt.distributed)
get_world_size() (in module torchopt.distributed)
grad() (in module torchopt.distributed.autograd)
I
ImplicitMetaGradientModule (class in torchopt.diff.implicit.nn)
L
linear_schedule() (in module torchopt.schedule)
linear_solve (torchopt.diff.implicit.nn.ImplicitMetaGradientModule attribute)
load_state_dict() (torchopt.FuncOptimizer method)
(torchopt.MetaOptimizer method)
(torchopt.Optimizer method)
M
make_dot() (in module torchopt.visual)
mean_reducer() (in module torchopt.distributed)
MetaAdaDelta (class in torchopt)
MetaAdaGrad (class in torchopt)
MetaAdam (class in torchopt)
MetaAdaMax (class in torchopt)
MetaAdamW (class in torchopt)
MetaOptimizer (class in torchopt)
MetaRAdam (class in torchopt)
MetaRMSProp (class in torchopt)
MetaSGD (class in torchopt)
N
nan_to_num() (in module torchopt)
nan_to_num_hook() (in module torchopt.hook)
not_on_rank() (in module torchopt.distributed)
O
objective() (torchopt.diff.implicit.nn.ImplicitMetaGradientModule method)
on_rank() (in module torchopt.distributed)
optimality() (torchopt.diff.implicit.nn.ImplicitMetaGradientModule method)
Optimizer (class in torchopt)
P
parallelize() (in module torchopt.distributed)
parallelize_async() (in module torchopt.distributed)
parallelize_sync() (in module torchopt.distributed)
polynomial_schedule() (in module torchopt.schedule)
Python Enhancement Proposals
PEP 599
R
RAdam (class in torchopt)
radam() (in module torchopt)
rank_non_zero_only() (in module torchopt.distributed)
rank_zero_only() (in module torchopt.distributed)
recover_state_dict() (in module torchopt)
register_hook() (in module torchopt.hook)
remote_async_call() (in module torchopt.distributed)
remote_sync_call() (in module torchopt.distributed)
RMSProp (class in torchopt)
rmsprop() (in module torchopt)
S
sample() (torchopt.diff.zero_order.nn.ZeroOrderGradientModule method)
SGD (class in torchopt)
sgd() (in module torchopt)
solve() (torchopt.diff.implicit.nn.ImplicitMetaGradientModule method)
solve_cg() (in module torchopt.linear_solve)
solve_inv() (in module torchopt.linear_solve)
solve_normal_cg() (in module torchopt.linear_solve)
state_dict() (torchopt.FuncOptimizer method)
(torchopt.MetaOptimizer method)
(torchopt.Optimizer method)
step() (torchopt.FuncOptimizer method)
(torchopt.MetaOptimizer method)
(torchopt.Optimizer method)
stop_gradient() (in module torchopt)
sum_reducer() (in module torchopt.distributed)
Z
zero_grad() (torchopt.Optimizer method)
zero_nan_hook() (in module torchopt.hook)
zero_order() (in module torchopt.diff.zero_order)
ZeroOrderGradientModule (class in torchopt.diff.zero_order.nn)
Read the Docs
v: stable
Versions
latest
stable
Downloads
On Read the Docs
Project Home
Builds