YingLong

YingLong model is introduced in this paper. This version is pre-trained on 78B time points. More details can be found at our github.

Quickstart

pip install xformers transformers
pip install flash-attn --no-build-isolation
git clone https://github.com/Dao-AILab/flash-attention && cd flash-attention
cd csrc/rotary && pip install .
cd ../layer_norm && pip install .

The flash attention is not required. If you use V100 or other GPU doesn't support flash attention, just change the FlashAttention2Available = RequirementCache("flash-attn>=2.0.0.post1") to FlashAttention2Available = False in the model.py file. It should be able to run.

import torch
from transformers import AutoModelForCausalLM

# load pretrain model
model = AutoModelForCausalLM.from_pretrained('qcw2333/YingLong_6m', trust_remote_code=True,torch_dtype=torch.bfloat16).cuda()

# prepare input
batch_size, lookback_length = 1, 2880
seqs = torch.randn(batch_size, lookback_length).bfloat16().cuda()

# generate forecast
prediction_length = 96
output = model.generate(seqs, future_token=prediction_length)

print(output.shape)

A notebook example is also provided here. The sample codes for long-term forecasting tasks and gift-eval tasks are provided at link.

Citation

Coming soon...

Contact

If you have any questions or want to use the code, feel free to contact:

Xue Wang ([email protected])

Tian Zhou ([email protected])

License

This model is licensed under the cc-by-4.0 License.

Downloads last month
113
Safetensors
Model size
7.32M params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support
OSZAR »