yuqinie98 / PatchTST

An offical implementation of PatchTST: "A Time Series is Worth 64 Words: Long-term Forecasting with Transformers." (ICLR 2023) https://arxiv.org/abs/2211.14730

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

how to use learner.distributed(), in self supervised pretrain code ?

lileishitou opened this issue · comments

how to use self supervised pretrain code to train on multiple GPUS or multi-node ?

I want to user revised the code for multinode or multiple GPUS for large dataset and large parameters.

But not successed