I have a tensor of shape (number_of rays, number_of_points_per_ray, 3), let’s call it input. input is passed through a model and some processing (all ...
I have a tensor of shape (number_of rays, number_of_points_per_ray, 3), let’s call it input. input is passed through a model and some processing (all ...
Autograd doesn't seem to be working reliably for a dataset I'm currently working with and I would like to use manually computed gradients with the Tor ...
I am trying to fine-tune GPT J, but I have this error. I think it's related to the activation function and it's in-place but I don't know how to code ...
I'm making a model mixing Fine-tuning CLIP model & Freezing clip model. And I make a custom loss using kl_loss and CEE But when I have train mo ...
In the pytorch autograd profiler documentation, it says that the profiler is a "Context manager that manages autograd profiler state and holds a summa ...
what would be the equivalent in Pytorch of the following in tensorflow, where loss is the calculated loss in the iteration of the network and net is t ...
I'm new to pytorch and I'm having a problem with some code to train a a neural network to solve a control problem. I use the following code to solve a ...
I am having trouble understanding the usage of the inputs keyword in the .backward() call. The Documentation says the following: inputs (sequence ...
I can't fix the runtime error "one of the variables needed for gradient computation has been modified by an inplace operation. I know, that if I comm ...
jax.nn.softmax is defined as: def softmax(x: Array, axis: Optional[Union[int, Tuple[int, ...]]] = -1, where: Optional[Array] ...
I was reading this blog from PyTorch. Just before the AutoGrad in training Section , it is mentioned Be aware that only leaf nodes of the computat ...
I would like to use pytorch to optimize a objective function which makes use of an operation that cannot be tracked by torch.autograd. I wrapped such ...
A brief description of my model: Consists of a single parameter X of dtype ComplexDouble and shape (20, 20, 20, 3). For reference, this must be co ...
I'm trying to compute the gradient of a lambda function that involves other gradients of functions, but the computation is hanging and I do not unders ...
I've been trying to understand how automatic differentiation (autodiff) works. There are several implementations of this that can be found in Tensorfl ...
I'm new to pytorch and had no luck following similar threads. I'm trying to jointly train two models in the same loop, and the model updates involve a ...
I am trying to compute a gradient of y_hat to x (y_hat is the sum of gradients of model output to x) but it gives me the error: One of the differentia ...
I tried to use gradient accumulation in my project. To my understanding, the gradient accumulation is the same as increasing the batch size by x times ...
After executing codes, the a.grad is None although a.requires_grad is True. But if the code a = a.cuda() is removed, a.grad is available after the l ...
I am training a model to predict pose using a custom Pytorch model. However, V1 below never learns (params don't change). The output is connected to t ...