Comments (10)
Question 1: For each epoch during training, the model is trained with seen data and then evaluate the model (from this epoch) with test set. After that, the accuracy is obtained after evaluation in this epoch. After 30 epochs, you pick the best accuracy as the final result. Am I right?
Ans - Yes, your understanding is correct.
Question 2: How about the argparse parameter: --syn_num? I can see that the default is set to 600 (means generating 600 visual representations for each action class?). Is there any experiment to indicate that this parameter can significantly influence in the model performance? Any suggestions for that?
Ans - syn_num parameters controls how many unseen samples per class you want to synthesis from the generator. Yes this parameter influences the model performance because further after synthesis you also train a classifier. The classifier training input data is a combination of these synthesised unseen class samples along with the real samples.
from tfvaegan.
In addition, I tried a different class embedding (dimension is 2048), but got the error RuntimeError: CUDA error: device-side assert triggered'
. I set argparse as: --nz 2048 --attSize 2048
Also, this error comes from optimizerE.step()
in train_action.py
. Is this error about tensor size mismatch? I have no idea right now. Thanks
from tfvaegan.
In addition, I tried a different class embedding (dimension is 2048), but got the error
RuntimeError: CUDA error: device-side assert triggered'
. I set argparse as:--nz 2048 --attSize 2048
Also, this error comes from
optimizerE.step()
intrain_action.py
. Is this error about tensor size mismatch? I have no idea right now. Thanks
For the issue above, there are more details as follows.
'/pytorch/aten/src/ATen/native/cuda/Loss.cu:111: operator(): block: [1267,0,0], thread: [31,0,0] Assertion input_val >= zero && input_val <= one
failed.' `
File "/content/kg_gnn_gan/train_tfvaegan.py", line 274, in <module>', ' vae_loss_seen = loss_fn(recon_x, input_resv, means, log_var)', ' File "/content/kg_gnn_gan/train_tfvaegan.py", line 72, in loss_fn', ' BCE = torch.nn.functional.binary_cross_entropy(recon_x + 1e-12, x.detach(), size_average=False)', ' File "/usr/local/lib/python3.7/dist-packages/torch/nn/functional.py", line 2893, in binary_cross_entropy', ' return torch._C._nn.binary_cross_entropy(input, target, weight, reduction_enum)', 'RuntimeError: CUDA error: device-side assert triggered']
It seems like the issue is about the calculation of loss function.
from tfvaegan.
In addition, I tried a different class embedding (dimension is 2048), but got the error
RuntimeError: CUDA error: device-side assert triggered'
. I set argparse as:--nz 2048 --attSize 2048
Also, this error comes from
optimizerE.step()
intrain_action.py
. Is this error about tensor size mismatch? I have no idea right now. Thanks
Some updates for the issue above.
I tested another semantic embedding, which has 1024 dimensions. It works.
But this failed -> the case with using 2048 dimensional vector as semantic embedding.
Notes: When I apply different semantic embeddings, I set proper parameters according to the size of embedding in the script. Also, I do not change other codes or settings (only changing semantic embedding).
Looking forward to hearing from you. Thanks.
Kind regards.
Kai
from tfvaegan.
Hi Akshitac,
Sorry to open a new issue since I get more questions about training.
Question 1: For each epoch during training, the model is trained with seen data and then evaluate the model (from this epoch) with test set. After that, the accuracy is obtained after evaluation in this epoch. After 30 epochs, you pick the best accuracy as the final result. Am I right?
Question 2: How about the argparse parameter:
--syn_num
? I can see that the default is set to 600 (means generating 600 visual representations for each action class?). Is there any experiment to indicate that this parameter can significantly influence in the model performance? Any suggestions for that?Thanks in advance.
Kind regards.
Kaiqiang
Based on my understanding of this paper: Your understanding is correct. For the issue of syn_num parameter, please refer to FREE: Feature Refinement for Generalized Zero-Shot Learning。
from tfvaegan.
Hi Akshitac,
Sorry to open a new issue since I get more questions about training.
Question 1: For each epoch during training, the model is trained with seen data and then evaluate the model (from this epoch) with test set. After that, the accuracy is obtained after evaluation in this epoch. After 30 epochs, you pick the best accuracy as the final result. Am I right?
Question 2: How about the argparse parameter:
--syn_num
? I can see that the default is set to 600 (means generating 600 visual representations for each action class?). Is there any experiment to indicate that this parameter can significantly influence in the model performance? Any suggestions for that?Thanks in advance.
Kind regards.
Kaiqiang
Based on my understanding of this paper: Your understanding is correct. For the issue of syn_num parameter, please refer to FREE: Feature Refinement for Generalized Zero-Shot Learning。
Thank you so much. Nice paper recommendation.
from tfvaegan.
Hi Akshitac,
Sorry to open a new issue since I get more questions about training.
Question 1: For each epoch during training, the model is trained with seen data and then evaluate the model (from this epoch) with test set. After that, the accuracy is obtained after evaluation in this epoch. After 30 epochs, you pick the best accuracy as the final result. Am I right?
Question 2: How about the argparse parameter:
--syn_num
? I can see that the default is set to 600 (means generating 600 visual representations for each action class?). Is there any experiment to indicate that this parameter can significantly influence in the model performance? Any suggestions for that?Thanks in advance.
Kind regards.
Kaiqiang
Based on my understanding of this paper: Your understanding is correct. For the issue of syn_num parameter, please refer to FREE: Feature Refinement for Generalized Zero-Shot Learning。
Thank you so much. Nice paper recommendation.
You bet! Do you know this question? #24 (comment)
from tfvaegan.
Hi Akshitac,
Sorry to open a new issue since I get more questions about training.
Question 1: For each epoch during training, the model is trained with seen data and then evaluate the model (from this epoch) with test set. After that, the accuracy is obtained after evaluation in this epoch. After 30 epochs, you pick the best accuracy as the final result. Am I right?
Question 2: How about the argparse parameter:
--syn_num
? I can see that the default is set to 600 (means generating 600 visual representations for each action class?). Is there any experiment to indicate that this parameter can significantly influence in the model performance? Any suggestions for that?Thanks in advance.
Kind regards.
Kaiqiang
Based on my understanding of this paper: Your understanding is correct. For the issue of syn_num parameter, please refer to FREE: Feature Refinement for Generalized Zero-Shot Learning。
Thank you so much. Nice paper recommendation.
You bet! Do you know this question? #24 (comment)
Sorry, I did not test on image ZSL. I just focus on action recognition. I guess the issue may be caused by hyper-parameter settings.
from tfvaegan.
Hi Akshitac,
Sorry to open a new issue since I get more questions about training.
Question 1: For each epoch during training, the model is trained with seen data and then evaluate the model (from this epoch) with test set. After that, the accuracy is obtained after evaluation in this epoch. After 30 epochs, you pick the best accuracy as the final result. Am I right?
Question 2: How about the argparse parameter:
--syn_num
? I can see that the default is set to 600 (means generating 600 visual representations for each action class?). Is there any experiment to indicate that this parameter can significantly influence in the model performance? Any suggestions for that?Thanks in advance.
Kind regards.
Kaiqiang
Based on my understanding of this paper: Your understanding is correct. For the issue of syn_num parameter, please refer to FREE: Feature Refinement for Generalized Zero-Shot Learning。
Thank you so much. Nice paper recommendation.
You bet! Do you know this question? #24 (comment)
Sorry, I did not test on image ZSL. I just focus on action recognition. I guess the issue may be caused by hyper-parameter settings.
Oh, it doesn't matter.
from tfvaegan.
Question 1: For each epoch during training, the model is trained with seen data and then evaluate the model (from this epoch) with test set. After that, the accuracy is obtained after evaluation in this epoch. After 30 epochs, you pick the best accuracy as the final result. Am I right?
Ans - Yes, your understanding is correct.
Question 2: How about the argparse parameter: --syn_num? I can see that the default is set to 600 (means generating 600 visual representations for each action class?). Is there any experiment to indicate that this parameter can significantly influence in the model performance? Any suggestions for that?
Ans - syn_num parameters controls how many unseen samples per class you want to synthesis from the generator. Yes this parameter influences the model performance because further after synthesis you also train a classifier. The classifier training input data is a combination of these synthesised unseen class samples along with the real samples.
Thanks for your clarification. Much appreciated.
from tfvaegan.
Related Issues (20)
- There is problem with AWA2!! HOT 3
- ModuleNotFoundError: No module named 'util' HOT 1
- Transductive Setting HOT 6
- Recreating ResNet visual features HOT 3
- performance about CUB dataset HOT 7
- epoch 0 HOT 7
- classify HOT 5
- 'Variable' object has no attribute 'getLayersOutDet' HOT 7
- Transductive setting's hyperparameters HOT 1
- Model selection criteria HOT 4
- There is a gap in performance than in the paper HOT 7
- Question about argparse HOT 2
- What is the training unseen feature under transductive setting?
- Question about training on action recognition HOT 1
- About the loss in the training process HOT 3
- Question regarding fine-tuned features HOT 1
- role of parameter
- _test_unseen_feature = scaler.transform(feature[test_unseen_loc]) IndexError: index 382 is out of bounds for axis 0 with size 382
- Finetuned features MISS HOT 2
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from tfvaegan.