You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, first at all thanks for sharing the code of your AFM implementation. We tried to replicate the results from the original paper. We executed the code exactly as was given here in this github repository.
However after 20 epochs of training the difference between FM and AFM for the MovieLens dataset was not as large as shown in Figure 5 in the paper. Here are the training results for FM and AFM:
In Figure 5 in the paper the RMSE of AFM after 20 epochs are below 0.45.
We also tried to reimplement the AFM model. With the help of the details given in this github we were already able to improve the performance of the AFM model significantly. Did we miss any other details or is the code in this github repository different from the code used in the paper?
We would be very thankful if you can help us since we currently try to include the AFM model into our production pipeline.
The text was updated successfully, but these errors were encountered:
As mentioned in one of the github issues, we also tried to set batch_size to 2048 and lamda_attention to 32 but with no success. FM and AFM are almost identical: FM RMSE is 0.47 and AFM RMSE is 0.471.
As mentioned in one of the github issues, we also tried to set batch_size to 2048 and lamda_attention to 32 but with no success. FM and AFM are almost identical: FM RMSE is 0.47 and AFM RMSE is 0.471.
I had the same problem, how do you solve it finally?
As mentioned in one of the github issues, we also tried to set batch_size to 2048 and lamda_attention to 32 but with no success. FM and AFM are almost identical: FM RMSE is 0.47 and AFM RMSE is 0.471.
I had the same problem, how do you solve it finally?
We haven't solved it yet. If you find the solution we would be very interested in it.
Hi, first at all thanks for sharing the code of your AFM implementation. We tried to replicate the results from the original paper. We executed the code exactly as was given here in this github repository.
However after 20 epochs of training the difference between FM and AFM for the MovieLens dataset was not as large as shown in Figure 5 in the paper. Here are the training results for FM and AFM:
python FM.py --dataset ml-tag --epoch 20 --pretrain -1 --batch_size 4096 --hidden_factor 256 --lr 0.01 --keep 0.7
Best Iter(validation)= 17 train = 0.0838, valid = 0.4855 [891.8 s]
python AFM.py --dataset ml-tag --epoch 20 --pretrain 1 --batch_size 4096 --hidden_factor [8,256] --keep [1.0,0.5] --lamda_attention 2.0 --lr 0.1
Best Iter(validation)= 20 train = 0.1038, valid = 0.4780 [1587.2 s]
In Figure 5 in the paper the RMSE of AFM after 20 epochs are below 0.45.
We also tried to reimplement the AFM model. With the help of the details given in this github we were already able to improve the performance of the AFM model significantly. Did we miss any other details or is the code in this github repository different from the code used in the paper?
We would be very thankful if you can help us since we currently try to include the AFM model into our production pipeline.
The text was updated successfully, but these errors were encountered: