You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The code runs with the modification, and I get SR=18.03%, which matches the val-unseen score for "without template assumption" (Table 2 of the FILM paper). I am using the best_model_multi.pt as the semantic search policy.
What I wasn't quite sure was what language processing modules is being used. I think the predicted templates are read from the models/instructions_processed_LP/instruction2_params*.p files, and I was wondering whether they are generated with/without the template assumption.
Thanks,
The text was updated successfully, but these errors were encountered:
I looked at my results in more details and have one update to the original post.
The code runs with the modification, and I get SR=18.03%
The score I report here is actually a result of splitting the evaluation among 3 PCs (--from_idx and --to_idx were split among 3 PCs), so I doubt the random number generation is the same as the one used in the paper. So please ignore about 18.03% matching the paper result.
I am still interested in knowing what language processing modules is being used- i.e. whether models/instructions_processed_LP/instruction2_params*.p files are generated with/without the template assumption.
I am currently trying to run the evaluation code for valid-unseen dataset, with the available pretrained models, and I have some questions.
First, this is the code I am using to run the evaluation:
The code fails to run, however, complaining that
rewards.json
is missing. So I pulled it from the [ALFRED repo](https://github.com/askforalfred/alfred/blob/master/models/config/rewards.json), and added--reward_config
flag toarguments.py
.The code runs with the modification, and I get SR=18.03%, which matches the val-unseen score for "without template assumption" (Table 2 of the FILM paper). I am using the
best_model_multi.pt
as the semantic search policy.What I wasn't quite sure was what language processing modules is being used. I think the predicted templates are read from the
models/instructions_processed_LP/instruction2_params*.p
files, and I was wondering whether they are generated with/without the template assumption.Thanks,
The text was updated successfully, but these errors were encountered: