探索flutter框架开发的app在移动应用市场的潜力与挑战
1082
2022-10-22
Huskarl:并行深度强化学习框架
Huskarl
Huskarl is a framework for deep reinforcement learning focused on research and fast prototyping. It's built on TensorFlow 2.0 and uses the tf.keras API when possible for conciseness and readability.
Huskarl makes it easy to parallelize computation of environment dynamics across multiple CPUs. This is useful for speeding up on-policy learning algorithms that benefit from multiple concurrent sources of experience such as A2C or PPO. It is specially useful for computationally intensive environments such as physics-based ones.
Huskarl works seamlessly with OpenAI Gym environments.
There are plans to support multi-agent environments and Unity3D environments.
Algorithms
Several algorithms are implemented already and many more are planned.
Deep Q-Learning Network (DQN) Multi-step DQN Double DQN Dueling Architecture DQN Advantage Actor-Critic (A2C) Deep Deterministic Policy Gradient (DDPG) Proximal Policy Optimization (PPO) Prioritized Experience Replay Curiosity-Driven Exploration
Installation
Since TensorFlow 2.0 is not officially out yet you need to install it and other dependencies manually for now:
pip install tf-nightly-2.0-previewpip install cloudpicklepip install scipypip install huskarl --no-deps
Citing
If you use Huskarl in your research, you can cite it as follows:
@misc{salvadori2019huskarl, author = {Daniel Salvadori}, title = {huskarl}, year = {2019}, publisher = {GitHub}, journal = {GitHub repository}, howpublished = {\url{https://github.com/danaugrs/huskarl}},}
About
hùskarl in Old Norse means a warrior who works in his/her lord's service.
版权声明:本文内容由网络用户投稿,版权归原作者所有,本站不拥有其著作权,亦不承担相应法律责任。如果您发现本站中有涉嫌抄袭或描述失实的内容,请联系我们jiasou666@gmail.com 处理,核实后本网站将在24小时内删除侵权内容。
发表评论
暂时没有评论,来抢沙发吧~