You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Moirai is a large pre-trained Time Series Model based on the Masked Encoder architecture. It is a universal time series forecasting model capable of addressing diverse forecasting tasks across multiple domains, frequencies, and variables in a zero-shot manner.

This is a version of Moirai small trained by Faculty AI. It was pre-trained on the LOTSA data using the codebase provided by Woo et al. (2024). Both the dataset and codebase are licensed under the Apache License 2.0. For more details on the model architecture, training, and results, please refer to the paper.

Usage

Please follow the Installation instructions and Getting Started section provided in the uni2ts repo. To use the model trained by Faculty AI simply use FacultyAI/moirai-small when fetching the model weights.

model = MoiraiForecast(
    module=MoiraiModule.from_pretrained("FacultyAI/moirai-small"),
    ...
    )

References

Woo, G., Liu, C., Kumar, A., Xiong, C., Savarese, S., & Sahoo, D. (2024). Unified Training of Universal Time Series Forecasting Transformers. arXiv preprint arXiv:2402.02592.
Downloads last month
6
Safetensors
Model size
13.8M params
Tensor type
F32
·
Inference API
Unable to determine this model's library. Check the docs .