Akshay
Akshay
I am facing this same issue , I deleted a lot of the tasks but it still shows the storage is exceeded , @oren-allegro could you check and let me...
@rwightman Thanks for the detailed reply , I guess the openAI team through empirical analysis might have seen that there is no difference is adding bias or making the mlp...
Thanks @rom1504 Thanks for the logs , isnt the logit_scale going towards 100 as the loss decreases in this case , not to 1
this works for me sometimes , In the server delete “~/.vscode-server/“ folder and reconnect via vscode
so based on experiments it was found that GELU has a significantly smoother gradient transition and its not abrupt or sharp like relu , if u look at both the...
It also works if u do this ``` import math import torch from torch import nn, GradScaler from torch.utils.data import TensorDataset, DataLoader class TestModule(nn.Module): def __init__(self, in_dim=512, out_dim=16): super().__init__() self.in_dim...
Thanks but ideally i would want to know the prompt that i am sending before hand , before sending it to the llm even for tracing , currently editing the...
The issue lies here and it gets solved when u do highres_in_channels = encoder_channels[-3] and high_res_features = self.block1(features[-3]) not sure if its a good workaround , would love to hear...