inspirai / timechamber Goto Github PK
View Code? Open in Web Editor NEWA Massively Parallel Large Scale Self-Play Framework
License: MIT License
A Massively Parallel Large Scale Self-Play Framework
License: MIT License
In Humanoid strike task config training file, training process seems epoch has to reach until 100000, which I assume, may take about 10days or more..
Could you give the specfic training days for this?
Maybe line 58, and 59 should exchange with line 61, and 62 ?
I have a rtx 4090 and when I use watch-smi and train Strike it uses less than half the memory and only 50% of the the GPU. How can I change the settings to use the rest of the GPU?
It occurred:
NotADirectoryError: [Errno 20] Not a directory: '/home/lzy/lzy/MARL/self-play/TimeChamber/timechamber/models/ant_sumo/policy.pth/../elo.jpg'
when I run policy evaluation:
python train.py task=MA_Ant_Sumo test=True headless=True checkpoint='models/ant_sumo/policy.pth'
Maybe this line code has some problem: TimeChamber/timechamber/learning/ppo_sp_player.py", line 286,
And the bug can be fixed by replacing
plt.savefig(self.params['load_path'] + '/../elo.jpg')
to
parent_path = os.path.dirname(self.params['load_path'])
plt.savefig(os.path.join(parent_path, 'elo.jpg'))
I want to see the video of policy evaluation, so I set headless=False
. But I found that the code run very slow in https://github.com/Jackory/TimeChamber/blob/37c2b40b8306921b748b318828769abafb90f347/timechamber/tasks/ma_ant_sumo.py#L63.
Could you give me some instructions to reproduce the README demo? Thanks
In Humanoid strike task config training file, it seems that epoch has to reach until 100000, which I assume, may take about 10days or more..
Could you give the specfic training days for this?
Hello,
When I try to run train.py, the following message appears:
Traceback (most recent call last):
File "train.py", line 33, in
import isaacgym
ModuleNotFoundError: No module named 'isaacgym'
But the thing is, IsaacGym is installed and works on my computer, I can run IsaacGymEnvs without problem. Thanks a lot for any help.
Hello,
Good job for the amazing work. I noticed that for the task of ma humanoid strike, you used a similar reward design to the one used for boxing agents in this paper:
https://dl.acm.org/doi/abs/10.1145/3450626.3459761
I was thinking that maybe strategic behavior could emerge with the sword fighters and the results could be better if you add a memory module (lstm, gru, transformer) like in the paper. Also, as far as I understand, in the literature of multi-agent partially observable mdp, each agent should take the history of its observations when taking actions to have more accurate belief about the global state and also to account for the non-stationary environment.
Thanks
when I run the code, there is an error:
Traceback (most recent call last): File "train.py", line 47, in <module> from timechamber.ase import ase_agent ModuleNotFoundError: No module named 'timechamber.ase'
I find that there is not 'ase' directory in timechamber, how can i get ase?
Hello,
When trying to see the value of each reward term for the task of striking, I noticed that the r_damage can be triggered even when the agents are not hitting each other. After further investigation, I found that the contact buffers can also include self collision in addition to contact between different agents; for example, if an agent's hand collide with its own head it will be included in the contact buffer, therefore it will be used in the reward as if it got hit by the opponent which was not the case.
This could be solved by filtering the contact buffer by specifying the bodies involved in the contact, but I saw that this functionality is not available right now in isaacgym. The workaround I can think of is to detect first if there was contact between specific bodies by measuring the distance between them, once the distance is lower than some threshold, you assume there was contact, then you check the magnitude of the contact force.
Hope I was clear.
How long does it take to train humanoid strike? I calculated that my computer takes about a month to train, is that normal? My GPU is RTX 4080.
Why is training one epoch slower than before after setting numEnvs=8192?
What parameters should I change to make training faster?
Please let me know, thank you very much.
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.