arthurdouillard / dytox

Dynamic Token Expansion with Continual Transformers, accepted at CVPR 2022

Home Page:https://arxiv.org/abs/2111.11326

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Why joint accuracy differs in different incremental settings?

ChenJunzhi-buaa opened this issue · comments

Thanky for your good work! I have some question about the upper bound.
As below picture from your papper, joint accuracy differs in different incremental settings. However, joint accuracy has nothing to do with incremental setting?
What's more, it seems that joint accuracy with resnet18 on CIFAR100 is about 70. In some other papers such as DER, when saving all old data, the average of all learning steps is about 80, so the acc of the last step is lower than 80.
Can you give some explains? Thank you very much!
image

For the joint, I'm reporting DER's metrics. Which indeed should have been the same, but i choose to report those for consistency.

I'm not sure about you mean in the second sentence. There is no "avg" for joint model, as it does only one phase. DER puts it as "avg" but it's a mistake from them.