Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[QUESTION] Issue with Evaluating Decision Transformer Using Evaluators in d3rlpy #406

Open
XiudingCai opened this issue Jul 21, 2024 · 1 comment
Labels
enhancement New feature or request

Comments

@XiudingCai
Copy link

XiudingCai commented Jul 21, 2024

I have encountered an issue while trying to evaluate the performance of the Decision Transformer (DT) using the d3rlpy library. Unlike other methods such as CQL, it seems that DT does not support passing evaluators like:

evaluators={
    'action_diff': d3rlpy.metrics.ContinuousActionDiffEvaluator(test_episodes),
}

This limitation is problematic, especially in settings where an environment is not available for evaluation. It hinders the ability to compare the performance of DT with other methods under these conditions.

Is there a workaround or a recommended approach for this situation?

Thanks so much! :>

@XiudingCai XiudingCai added the enhancement New feature or request label Jul 21, 2024
@takuseno
Copy link
Owner

@XiudingCai Hi, sorry for the late response. This is a little tricky issue because Q-learning and Decision Transformer are completely different algorithms and it's difficult to share some interface. One possible workaround is to use callback option at fit method. You can add arbitrary logics at every steps.

callback: Optional[Callable[[Self, int, int], None]] = None,

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants