Replies: 1 comment 1 reply
-
@korchi - Could you figure it out? This seems like a big issue in getting the online performance at par with offline one |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi. First of all, thank you for the great tool you are developing. However, I am puzzled already for a week about how can I replicate
trainer.evaluation()
results with inferencing the model.My initial idea was to truncate every sessions by one (removing last item_id), call
trainer.predict(truncated_sessions)
, and then computerecall(last_item_ids, predictions[:20])
. However, I am getting different recall metric.The only way I managed to "replicate" evaluate() results is by: (1) providing not-truncated inputs to the
trainer.predict()
and (2) changing-1
into-2
inTransformers4Rec/transformers4rec/torch/model/prediction_task.py
Line 460 in 348c963
I am puzzled why, but this was the only way I could ensure that the
x
inTransformers4Rec/transformers4rec/torch/model/prediction_task.py
Line 464 in 348c963
x
inTransformers4Rec/transformers4rec/torch/model/prediction_task.py
Line 444 in 348c963
Is it because
trainer.evaluate()
shifts the inputs to the left by one position? Or what am I doing incorrectly? Could any provide me insights how to do it "correctly", please?Thanks a lot.
Beta Was this translation helpful? Give feedback.
All reactions