-
Notifications
You must be signed in to change notification settings - Fork 119
Difference in implementation and paper #12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I totally agree with you.. |
I agree, is there any reason for these differences? |
I think the authors use Chebyshev Polynomials to reduce time complexity (laplacian(n^3)). |
Output layers using GLU and fully connected layers are also not implemented. Is there anyone who reproduced the paper's performance? |
I think the current pytorch implementation is not following the paper's defined architecture. Perhaps we can try to replicate the exact architecture to see if the results will be closer? |
I tried but this paper lacks detail about implementation. |
I've also tried to reproduce the results on PEMS 3,4 and 7 but without success. |
Hi, I'm reading the paper and code rencently and also find the difference like the loss F you've described.Have you found a solution to the problem? |
Yes...Some concepts of StemGNN are really attractive, but they seem to be missing in the code... |
why not do some ablation experiment about single Y1 and single Y2? |
it seems the author made a few mistakes on the task they performed. For example pems07 is a traffic speed data but the author use it as a traffic flow task, meanwhile the result of pems04 experiment is definitely a traffic flow task but the author run it as a traffic speed task. |
@superarthurlx I think there are 2 different pems07. For example the one released by STSGCN(AAAI-2020) is traffic flow. |
@superarthurlx I think these concepts are getting blurred, maybe because they are the same to the model. For example the STGNN[1] also use METR-LA and PEMS-BAY as traffic flow prediction datasets, however they record traffic speed. [1] Wang X, Ma Y, Wang Y, et al. Traffic flow prediction via spatial temporal graph neural network[C]//Proceedings of The Web Conference 2020. 2020: 1082-1092. |
look at Table 2, the author used the same result presented in ST-GCN[38] which is actually called STGCN(Cheb), so the datasets should be same as well which is a speed data. |
i had loaded the dataset pems07 from the github of STSGCN,and proved that it was traffic speed.
…------------------ Original ------------------
From: superarthurlx ***@***.***>
Date: Wed,Oct 27,2021 11:04 PM
To: microsoft/StemGNN ***@***.***>
Cc: lixus7 ***@***.***>, Comment ***@***.***>
Subject: Re: [microsoft/StemGNN] Difference in implementation and paper (#12)
@superarthurlx I think there are 2 different pems07. For example the one released by STSGCN(AAAI-2020) is traffic flow.
look at Table 2, the author used the same result presented in ST-GCN[38] which is actually called STGCN(Cheb), so the datasets should be same as well which is a speed data.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub, or unsubscribe.
Triage notifications on the go with GitHub Mobile for iOS or Android.
|
Check table2 or pems07.csv file you can find the different. Flow data should be all int number while speed data are float numbers and usually less than 70 which is the speed limit in pems area. here are some of my reproductions, there are still many differences like mape values. ==! |
Thanks for explaining the difference. |
|
Hello, So, no one has succeeded in reproducing the results on their paper? |
Also interested in whether or not anyone has successfully reproduced output? Sounds like there are some outstanding discrepancies between the paper and the code? |
I found that the calculation of metrics in StemGNN seems to be unfair and inconsistent with baselines, such as the calculation in GWNet. |
Good day,
Thank you for the paper. I noticed that there are some difference in implementation described in the paper as well as the source code,
For example, the spectral sequential cell defined in the paper uses a convo 1D operation but i cant seem to see it in the code (unless i am missing something )
In the training loop, the backcast are not used in the loss function
Another question i have is that i dont understand how the following:
is actually the IGFT operation. I understand that as opposed to the paper, the Chebyshev Polynomial is used to directly compute the GFT operation without the need for the eigenvectors of L.
Will be good to explain! thanks!
The text was updated successfully, but these errors were encountered: