🚩News(Jan 16, 2025): After the meeting, we will upload this paper to arXiv.
In real-world scenarios, different types of missing mechanisms, like MAR (Missing At Random), and MNAR (Missing Not At Random) can occur in time series data. However, existing methods often overlook the difference among the aforementioned missing mechanisms and use a single model for time series imputation, which can easily lead to misleading results due to mechanism mismatching. In this paper, we propose a framework for time series imputation problem by exploring Different Missing Mechanisms (DMM in short) as shown in Figure 1 and tailoring solutions accordingly.
Figure 1. Data generation processes of time series data under different missing mechanisms.
Figure 2. Illustration of the DMM framework.
- Python 3.8
- torch == 2.3.1
- reformer-pytorch==1.4.4
- scikit-learn==1.2.2
- einops == 0.4.0
- tqdm == 4.64.1
Dependencies can be installed using the following command:
pip install -r requirements.txt
You can obtain all datasets from Google Drive.
To easily reproduce the results you can run the following commands:
python run_DMM.py -data Exchange -mask_type MAR -mask_rate 0.2 -train_mode 0 -DMM_type MAR
Multiple seeds and datasets can be run at one time. The important parameters are in file DMM_config.py and you can go inside to change the parameters you want.
And we provide explanations for the important parameters:
Parameter name | Description of parameter |
---|---|
data | The dataset name |
root_path | The root path of the data file (defaults to ./data/exchange_rate/ ) |
data_path | The data file name (defaults to exchange_rate.csv ) |
seq_len | Input sequence length (defaults to 96) |
mask_type | Dataset missing mechanism |
mask_rate | Dataset missing rate |
train_mode | Training methods |
d_model | Dimension of model |
train_epochs | Epochs in train |
learning_rate | Optimizer learning rate |
More parameter information please refer to run.py
.
The main results are shown in table 1 and table 2.
If you find this repository useful in your research, please consider citing the following papers:
To be continued...