Hi, thanks for releasing this impressive work.
I have a question about the Embodied Interactive Tasks(ALFWorld) experiments for baselines such as A-MEM and MemoryOS. Since these methods were originally designed more for dialogue / QA settings, I was wondering how you adapted them to ALFWorld in practice.
In particular:
How did you build their memory bank from the ALFWorld training trajectories? What did you use as the retrieval query during evaluation (objective only, or also observation/history)?
Was the adaptation mainly prompt-level, or did you also modify some code structure / memory pipeline for these baselines?
I’m trying to understand whether the ALFWorld setup for these baselines was mostly a unified prompting interface, or whether some method-specific implementation changes were needed.
Thanks again, the work is very impressive.
Hi, thanks for releasing this impressive work.
I have a question about the Embodied Interactive Tasks(ALFWorld) experiments for baselines such as A-MEM and MemoryOS. Since these methods were originally designed more for dialogue / QA settings, I was wondering how you adapted them to ALFWorld in practice.
In particular:
How did you build their memory bank from the ALFWorld training trajectories? What did you use as the retrieval query during evaluation (objective only, or also observation/history)?
Was the adaptation mainly prompt-level, or did you also modify some code structure / memory pipeline for these baselines?
I’m trying to understand whether the ALFWorld setup for these baselines was mostly a unified prompting interface, or whether some method-specific implementation changes were needed.
Thanks again, the work is very impressive.