TSTPlus __init__ adds extraneous layers with MVP Training? #120
Unanswered
xanderdunn
asked this question in
Q&A
Replies: 1 comment 1 reply
-
Hi @xanderdunn, Thanks for keeping a close eye on MVP :)
I'm not sure why there's such a big change in your numbers. It's logical that there's some difference, but not that big. Are you using any custom_head in TST that could explain this large drop? PS: I've continued to use MVP with both InceptionTimePlus and TSTPlus and have had very good results with a good speed. Did you raise this issue for any performance issue? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
The
TSTPlus
init creates its head like this:However, when using it for
MVP
training, theMVP
before_fit
replaces the model's.head
:With the default code, my TSTPlus model:
When I remove the TSTPlus code that creates a head so that only MVP is creating the head, I see:
Where the number of parameters in the graph is calculated with this:
It appears even though MVP replaced the TSTPlus head with its own
Sequential(Dropout, Conv1d)
head, the original TSTPlus head is still lingering in the torch graph. This is undesirable because it's still using VRAM and it may be computed on every iteration even though it shouldn't be affecting the outcome.Am I misunderstanding how the TSTPlus
__init__
head creation and theMVP
head creation interact? Thanks!Beta Was this translation helpful? Give feedback.
All reactions