Release Notes
Next release plan
The next release should deliver the LunarLander environment benchamrk (almost done) and SAIL (IJCNN 2023).
v0.6.0
Support for Visual environments
This version adds support for visual environments (e.g., Atari) to all methods, the BaselineDataset, and the benchmark feature.
Related issues:
Known bugs:
- For some reason, even though packaging builds the benchmark folder, it is not packaging it. So, for now, if users want to use the benchmark feature, they should install it from the source.
git clone https://github.com/NathanGavenski/IL-Datasets.git
cd IL-Datasets
pip install -e .
v0.4.0
Benchmarking
Now IL-Datasets has its own benchmarking! We are adding new methods and environments to the repository. For a full list of the methods and environments planned for release, please check the repository readme.md file.
Support for benchmark requirements
We split the imitation_datasets
and benchmark
modules requirements.
pip install il-datasets
will only install requirements regarding the imitation_datasets
module. For using benchmark
please use:
pip install "il-datasets[benchmark]"
Full Changelog: https://github.com/NathanGavenski/IL-Datasets/compare/0.3.0...0.4.0
v0.3.0
This version adds another point from the TODO list, Datasets!
Now, if you use the baseline_enjoy
and baseline_collate
functions, you can use the BaselineDataset
.
The datasets will load the generated numpy file and organize all entries to be (s_t, a_t, s_{t+1}), provide the average reward for all episodes and also allow for fewer episodes with the parameter n_episodes
.
Alongside the dataset, I've implemented a HuggingFace solution as well as utility functions that allow users to upload their datasets to the HuggingFace website. There is already an example at: https://huggingface.co/datasets/NathanGavenski/CartPole-v1 In the future, these datasets will be used for benchmarking, but for now, it allows for storing outside drivers (such as Google's and Microsoft's)
This version also comes with some QoL improvements, such as pylint, and unit tests, so the code is more readable and also more stable.
Finally, with this release, I've implemented some metrics: performance
, average episodic reward
and accuracy
.
Future release sneak peek
It is my plan that the future release will introduce benchmarking to IL-Datasets. With benchmarking, we will host a set of different datasets for common environments in the IL literature. This should help all researchers (including myself) to stop running different methods for each experiment.
Full Changelog: https://github.com/NathanGavenski/IL-Datasets/compare/0.2.0...0.3.0
V0.2.0
New Features
- Added support for Gymnasium and Gym version 0.26.0.
- Created template functions for
enjoy
andcollate
for a simple and one following the original dictionary from StableBaselines.
v0.1.0
FIx:
- Sometimes, when the policy did not reach the goal and the enjoy function returned
False
, theController
would not execute the enjoy function again
First release
First release for IL-Datasets. The missing features (such as documentation) list is in the README.md. If you have any issues with this release be sure to open an issue or contact me 😄