dart_rl 0.1.0-alpha.1
dart_rl: ^0.1.0-alpha.1 copied to clipboard
A Dart package implementing reinforcement learning algorithms (SARSA, Q-Learning, Expected-SARSA) for both Dart and Flutter applications.
Changelog #
All notable changes to this project will be documented in this file.
The format is based on Keep a Changelog, and this project adheres to Semantic Versioning.
0.1.0-alpha.1 2024-01-XX #
Added #
- Initial alpha release of dart_rl package
- Q-Learning algorithm implementation (
QLearningAgent) - SARSA algorithm implementation (
SarsaAgent) - Expected-SARSA algorithm implementation (
ExpectedSarsaAgent) Environmentinterface for creating custom RL environmentsAgentbase class with epsilon-greedy exploration strategyState,Action, andStateActionclasses for representing RL componentsStepResultclass for environment step results- Grid World example environment
- Frozen Lake example environment
- Comprehensive unit tests
- Documentation and README with usage examples
Features #
- Support for discrete state and action spaces
- Configurable learning rate (α), discount factor (γ), and epsilon (ε)
- Epsilon-greedy exploration with decay functionality
- Q-table access for inspection and debugging
- Training methods for single episodes and multiple episodes
- Compatible with both Dart and Flutter applications
0.1.0 2024-01-XX #
Added #
- Initial release of dart_rl package
- Q-Learning algorithm implementation (
QLearningAgent) - SARSA algorithm implementation (
SarsaAgent) - Expected-SARSA algorithm implementation (
ExpectedSarsaAgent) Environmentinterface for creating custom RL environmentsAgentbase class with epsilon-greedy exploration strategyState,Action, andStateActionclasses for representing RL componentsStepResultclass for environment step results- Grid World example environment
- Frozen Lake example environment
- Comprehensive unit tests
- Documentation and README with usage examples
Features #
- Support for discrete state and action spaces
- Configurable learning rate (?), discount factor (?), and epsilon (?)
- Epsilon-greedy exploration with decay functionality
- Q-table access for inspection and debugging
- Training methods for single episodes and multiple episodes
- Compatible with both Dart and Flutter applications