Git Product home page Git Product logo

timechamber's People

Contributors

zeldahuang avatar ziyiliubird avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

timechamber's Issues

How much time costs on training?

In Humanoid strike task config training file, training process seems epoch has to reach until 100000, which I assume, may take about 10days or more..
Could you give the specfic training days for this?

Bug

if self.player_pool_type == 'multi_thread':

return PFSPPlayerProcessPool(max_length=self.max_his_player_num,

elif self.player_pool_type == 'multi_process':

return PFSPPlayerThreadPool(max_length=self.max_his_player_num,

Maybe line 58, and 59 should exchange with line 61, and 62 ?

Maybe a Bug

It occurred:

NotADirectoryError: [Errno 20] Not a directory: '/home/lzy/lzy/MARL/self-play/TimeChamber/timechamber/models/ant_sumo/policy.pth/../elo.jpg'

when I run policy evaluation:

python train.py task=MA_Ant_Sumo test=True headless=True checkpoint='models/ant_sumo/policy.pth'

Maybe this line code has some problem: TimeChamber/timechamber/learning/ppo_sp_player.py", line 286,

And the bug can be fixed by replacing

plt.savefig(self.params['load_path'] + '/../elo.jpg')

to

parent_path = os.path.dirname(self.params['load_path'])

plt.savefig(os.path.join(parent_path, 'elo.jpg'))

How much time costs on training?

In Humanoid strike task config training file, it seems that epoch has to reach until 100000, which I assume, may take about 10days or more..
Could you give the specfic training days for this?

Can't find isaacgym module

Hello,
When I try to run train.py, the following message appears:

Traceback (most recent call last):
File "train.py", line 33, in
import isaacgym
ModuleNotFoundError: No module named 'isaacgym'

But the thing is, IsaacGym is installed and works on my computer, I can run IsaacGymEnvs without problem. Thanks a lot for any help.

Suggestion to use RNN for multi-agent tasks

Hello,
Good job for the amazing work. I noticed that for the task of ma humanoid strike, you used a similar reward design to the one used for boxing agents in this paper:

https://dl.acm.org/doi/abs/10.1145/3450626.3459761

I was thinking that maybe strategic behavior could emerge with the sword fighters and the results could be better if you add a memory module (lstm, gru, transformer) like in the paper. Also, as far as I understand, in the literature of multi-agent partially observable mdp, each agent should take the history of its observations when taking actions to have more accurate belief about the global state and also to account for the non-stationary environment.

Thanks

ModuleNotFoundError: No module named 'timechamber.ase'

when I run the code, there is an error:
Traceback (most recent call last): File "train.py", line 47, in <module> from timechamber.ase import ase_agent ModuleNotFoundError: No module named 'timechamber.ase'
I find that there is not 'ase' directory in timechamber, how can i get ase?

Issue with contact forces when computing r damage in strike task

Hello,

When trying to see the value of each reward term for the task of striking, I noticed that the r_damage can be triggered even when the agents are not hitting each other. After further investigation, I found that the contact buffers can also include self collision in addition to contact between different agents; for example, if an agent's hand collide with its own head it will be included in the contact buffer, therefore it will be used in the reward as if it got hit by the opponent which was not the case.

This could be solved by filtering the contact buffer by specifying the bodies involved in the contact, but I saw that this functionality is not available right now in isaacgym. The workaround I can think of is to detect first if there was contact between specific bodies by measuring the distance between them, once the distance is lower than some threshold, you assume there was contact, then you check the magnitude of the contact force.

Hope I was clear.

numEnvs

How long does it take to train humanoid strike? I calculated that my computer takes about a month to train, is that normal? My GPU is RTX 4080.
Why is training one epoch slower than before after setting numEnvs=8192?
What parameters should I change to make training faster?
Please let me know, thank you very much.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.