Skip to content

Latest commit

 

History

History
49 lines (33 loc) · 3.11 KB

project.md

File metadata and controls

49 lines (33 loc) · 3.11 KB

Research project

We propose a model that leverages the state of the art in foundation models trained on massive datasets, along with another simpler model trained on a smaller dataset specific to the domain of the time series to be forecasted. This combination aims to bring out the best of both worlds: increasing prediction accuracy compared to individual model results, as well as reducing computational costs by avoiding expensive training and fine-tuning routines.

Table of Contents

Research problem

The research problem revolves around the high computational costs associated with extensive training routines and fine-tuning of foundation models. These models, trained on massive datasets, offer state-of-the-art performance but require significant computational resources and time for training and optimization. The challenge aims to finding efficient methods to leverage the benefits of these foundation models while mitigating the computational complexity, especially in domains such as time series forecasting where computational efficiency is crucial for real-world applications.

Related work

Ongoing Tasks

Experiments

Proposal

Proposal

Initial Results

Proof of Concept

Initial paper (Bracis)

  • Distillations:
    • TimeGPT without tuning + residualMLP
    • Chronos without tuning + residualMLP
    • (TimeGPT + Chronos) without tuning + residualMLP

Future works

  • How to fine-tune the Chronos model?
  • Canonical correlation analysis (Similarity from model's internal representation)
  • Feature Distillation: Create a distillation related to internal know-how (Initializing Models with Larger Ones).