HazyResearch / safari

Convolutions for Sequence Modeling

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

H3 / LongConvKernel - l_max=None isn't working

grazder opened this issue · comments

In H3 model here noted that

l_max: the maximum kernel length, also denoted by L. Set l_max=None to always use a global kernel

But it isn't working, because of

return torch.randn(self.channels, self.H, self.L) * 0.002

Which is leading to torch.randn(int, int, None) error:

TypeError: randn(): argument 'size' must be tuple of ints, but found element of type NoneType at pos 3

So, are global kernels supported now? Or, to make it more global, should I use large l_max value?

commented

ok thank!
it would be clearer if there was an assert or something like that.
this would be helpful for users like me who use just the model implementation without any other framework things

commented