From 4feb4bef6658f6c749fc782f3b8890dd99e2b737 Mon Sep 17 00:00:00 2001 From: Moritz Gunz Date: Fri, 13 Dec 2024 11:51:14 +0100 Subject: [PATCH 1/9] Bump required Python 3.7 -> 3.8 Closes #1326 --- .github/workflows/black.yml | 2 +- .github/workflows/main.yml | 18 +++++++++--------- .github/workflows/publish.yml | 2 +- CHANGELOG.md | 2 ++ CONTRIBUTING.md | 2 +- pyproject.toml | 2 +- .../inspectionProfiles/Project_Default.xml | 7 +++---- 7 files changed, 18 insertions(+), 17 deletions(-) diff --git a/.github/workflows/black.yml b/.github/workflows/black.yml index 15791da38a..c0e16a6c9b 100644 --- a/.github/workflows/black.yml +++ b/.github/workflows/black.yml @@ -17,7 +17,7 @@ jobs: - uses: actions/checkout@v4 - uses: actions/setup-python@v5 with: - python-version: 3.7 + python-version: 3.8 cache: 'pip' cache-dependency-path: '.github/workflows/black.yml' - run: pip install black==22.3.0 diff --git a/.github/workflows/main.yml b/.github/workflows/main.yml index df540aed4b..b4b2f45680 100644 --- a/.github/workflows/main.yml +++ b/.github/workflows/main.yml @@ -92,7 +92,7 @@ jobs: fail-fast: false matrix: python-version: - - 3.7 + - 3.8 action: - TEST=Dataset - TEST=fork_exec @@ -158,7 +158,7 @@ jobs: fail-fast: false matrix: python-version: - - 3.7 + - 3.8 tf-version: - 2.10.0 action: @@ -175,22 +175,22 @@ jobs: - TEST=tools include: - action: TEST=TFUtil - python-version: 3.7 + python-version: 3.8 tf-version: 1.15.3 - action: TEST=TFEngine - python-version: 3.7 + python-version: 3.8 tf-version: 1.15.3 - action: TEST=TFEngine - python-version: 3.7 + python-version: 3.8 tf-version: 2.3.0 - action: TEST=TFNativeOp - python-version: 3.7 + python-version: 3.8 tf-version: 1.14.0 - action: TEST=TFNetworkSigProcLayer - python-version: 3.7 + python-version: 3.8 tf-version: 1.15.3 - action: TEST=demos RETURNN_DISABLE_TORCH=1 - python-version: 3.7 + python-version: 3.8 tf-version: 2.3.0 steps: @@ -358,7 +358,7 @@ jobs: include: # Some selected tests (but not all) with some other versions. - action: TEST=rf_base - python-version: 3.7 + python-version: 3.8 torch-version: 1.13.1 tf-version: 2.10.0 diff --git a/.github/workflows/publish.yml b/.github/workflows/publish.yml index 439117a59d..f144bb229c 100644 --- a/.github/workflows/publish.yml +++ b/.github/workflows/publish.yml @@ -26,7 +26,7 @@ jobs: - uses: actions/setup-python@v5 with: - python-version: 3.7 + python-version: 3.8 - name: Install Python deps run: | diff --git a/CHANGELOG.md b/CHANGELOG.md index c312078f81..557467a893 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -6,6 +6,8 @@ or any changes which could potentially break or change the behavior of existing This is intentionally kept short. For a full change log, just see the Git log. +## 2023-12-13: Bump min Python version from 3.7 to 3.8 ([issue #1326](https://github.com/rwth-i6/returnn/issues/1326)) + ## 2024-06-07: `VariableDataset` Custom subdataset per subepoch based on user-provided function. diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index b9699abcaf..f9936e237a 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -29,7 +29,7 @@ General rules when contributing to the code of RETURNN: and in general to use newer TF features (even maybe TF >=2), as long as your new feature (new layer or so) is just optional. For older TF support, `TFCompat.py` might be helpful. (See code for examples.) -* We support Python >=3.7 ([discussion](https://github.com/rwth-i6/returnn/issues/487)). +* We support Python >=3.8 ([discussion](https://github.com/rwth-i6/returnn/issues/1326)). About new features: diff --git a/pyproject.toml b/pyproject.toml index a14bd4aae0..14c8c02ca3 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [tool.black] line-length = 120 -target-version = ["py37"] # https://github.com/rwth-i6/returnn/issues/487 +target-version = ["py38"] # https://github.com/rwth-i6/returnn/issues/1326 extend-exclude = ''' /( # All Git submodules, see .gitmodules. diff --git a/tests/PyCharm.idea/inspectionProfiles/Project_Default.xml b/tests/PyCharm.idea/inspectionProfiles/Project_Default.xml index 88cda7840b..6be6cae45f 100644 --- a/tests/PyCharm.idea/inspectionProfiles/Project_Default.xml +++ b/tests/PyCharm.idea/inspectionProfiles/Project_Default.xml @@ -4,10 +4,9 @@ From 1b79a6a5751d983dcd48b6700845787de60564b0 Mon Sep 17 00:00:00 2001 From: Moritz Gunz Date: Fri, 13 Dec 2024 12:12:41 +0100 Subject: [PATCH 2/9] fix year lol --- CHANGELOG.md | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 557467a893..35add63ea3 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -6,7 +6,9 @@ or any changes which could potentially break or change the behavior of existing This is intentionally kept short. For a full change log, just see the Git log. -## 2023-12-13: Bump min Python version from 3.7 to 3.8 ([issue #1326](https://github.com/rwth-i6/returnn/issues/1326)) +## 2024-12-13: Bump min Python version from 3.7 to 3.8 ([issue #1326](https://github.com/rwth-i6/returnn/issues/1326)) + +This also drops support for TF 1.x. ## 2024-06-07: `VariableDataset` From c00d1f58ecc788b63265908b5ba1285b9bd15567 Mon Sep 17 00:00:00 2001 From: Moritz Gunz Date: Fri, 13 Dec 2024 12:27:11 +0100 Subject: [PATCH 3/9] convert readme to markdown, add note about min python versions Markdown is easier to edit for many folks as it has broader usage outside the python docs ecosystem. --- README.md | 72 ++++++++++++++++++++++++++++++++++++++++++++++++++++ README.rst | 74 ------------------------------------------------------ setup.py | 4 +-- 3 files changed, 74 insertions(+), 76 deletions(-) create mode 100644 README.md delete mode 100644 README.rst diff --git a/README.md b/README.md new file mode 100644 index 0000000000..3abd31c11f --- /dev/null +++ b/README.md @@ -0,0 +1,72 @@ +# Welcome to RETURNN + +[GitHub repository](https://github.com/rwth-i6/returnn), +[RETURNN paper 2016](https://arxiv.org/abs/1608.00895), +[RETURNN paper 2018](https://arxiv.org/abs/1805.05225). + +RETURNN - RWTH extensible training framework for universal recurrent neural networks, +is a PyTorch/TensorFlow-based implementation of modern neural network architectures. +It is optimized for fast and reliable training of neural networks in a multi-GPU environment. + +The high-level features and goals of RETURNN are: + +- **Simplicity** + - Writing config / code is simple & straight-forward (setting up experiment, defining model) + - Debugging in case of problems is simple + - Reading config / code is simple (defined model, training, decoding all becomes clear) + +- **Flexibility** + - Allow for many different kinds of experiments / models + +- **Efficiency** + - Training speed + - Decoding speed + +All items are important for research, decoding speed is esp. important for production. + +See our [Interspeech 2020 tutorial "Efficient and Flexible Implementation of Machine Learning for ASR and MT" video](https://www.youtube.com/watch?v=wPKdYqSOlAY) +([slides](https://www-i6.informatik.rwth-aachen.de/publications/download/1154/Zeyer--2020.pdf)) +with an introduction of the core concepts. + +More specific features include: + +- Mini-batch training of feed-forward neural networks +- Sequence-chunking based batch training for recurrent neural networks +- Long short-term memory recurrent neural networks + including our own fast CUDA kernel +- Multidimensional LSTM (GPU only, there is no CPU version) +- Memory management for large data sets +- Work distribution across multiple devices +- Flexible and fast architecture which allows all kinds of encoder-attention-decoder models + +See [documentation](https://returnn.readthedocs.io/). +See [basic usage](https://returnn.readthedocs.io/en/latest/basic_usage.html) and [technological overview](https://returnn.readthedocs.io/en/latest/tech_overview.html). + +[Here is the video recording of a RETURNN overview talk](https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.recording.cut.mp4) +([slides](https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.returnn-overview.session1.handout.v1.pdf), +[exercise sheet](https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.exercise_sheet.pdf); hosted by eBay). + +There are [many example demos](https://github.com/rwth-i6/returnn/blob/master/demos/) +which work on artificially generated data, +i.e. they should work as-is. + +There are [some real-world examples](https://github.com/rwth-i6/returnn-experiments) +such as setups for speech recognition on the Switchboard or LibriSpeech corpus. + +Some benchmark setups against other frameworks +can be found [here](https://github.com/rwth-i6/returnn-benchmarks). +The results are in the [RETURNN paper 2016](https://arxiv.org/abs/1608.00895). +Performance benchmarks of our LSTM kernel vs CuDNN and other TensorFlow kernels +are in [TensorFlow LSTM benchmark](https://returnn.readthedocs.io/en/latest/tf_lstm_benchmark.html). + +There is also [a wiki](https://github.com/rwth-i6/returnn/wiki). +Questions can also be asked on +[StackOverflow using the RETURNN tag](https://stackoverflow.com/questions/tagged/returnn). + +[![CI](https://github.com/rwth-i6/returnn/workflows/CI/badge.svg)](https://github.com/rwth-i6/returnn/actions) + +## Dependencies + +pip dependencies are listed in `requirements.txt` and `requirements-dev`, although some parts of the code may require additional dependencies (e.g. `librosa`, `resampy`) on-demand. + +RETURNN supports Python >= 3.8. Bumps to the minimum Python version are listed in [`CHANGELOG.md`](https://github.com/rwth-i6/returnn/blob/master/CHANGELOG.md). diff --git a/README.rst b/README.rst deleted file mode 100644 index 4793b3fc0f..0000000000 --- a/README.rst +++ /dev/null @@ -1,74 +0,0 @@ -================== -Welcome to RETURNN -================== - -`GitHub repository `__. -`RETURNN paper 2016 `_, -`RETURNN paper 2018 `_. - -RETURNN - RWTH extensible training framework for universal recurrent neural networks, -is a Theano/TensorFlow-based implementation of modern recurrent neural network architectures. -It is optimized for fast and reliable training of recurrent neural networks in a multi-GPU environment. - -The high-level features and goals of RETURNN are: - -* **Simplicity** - - * Writing config / code is simple & straight-forward (setting up experiment, defining model) - * Debugging in case of problems is simple - * Reading config / code is simple (defined model, training, decoding all becomes clear) - -* **Flexibility** - - * Allow for many different kinds of experiments / models - -* **Efficiency** - - * Training speed - * Decoding speed - -All items are important for research, decoding speed is esp. important for production. - -See our `Interspeech 2020 tutorial "Efficient and Flexible Implementation of Machine Learning for ASR and MT" video `__ -(`slides `__) -with an introduction of the core concepts. - -More specific features include: - -- Mini-batch training of feed-forward neural networks -- Sequence-chunking based batch training for recurrent neural networks -- Long short-term memory recurrent neural networks - including our own fast CUDA kernel -- Multidimensional LSTM (GPU only, there is no CPU version) -- Memory management for large data sets -- Work distribution across multiple devices -- Flexible and fast architecture which allows all kinds of encoder-attention-decoder models - -See `documentation `__. -See `basic usage `__ -and `technological overview `__. - -`Here is the video recording of a RETURNN overview talk `_ -(`slides `__, -`exercise sheet `__; -hosted by eBay). - -There are `many example demos `_ -which work on artificially generated data, -i.e. they should work as-is. - -There are `some real-world examples `_ -such as setups for speech recognition on the Switchboard or LibriSpeech corpus. - -Some benchmark setups against other frameworks -can be found `here `_. -The results are in the `RETURNN paper 2016 `_. -Performance benchmarks of our LSTM kernel vs CuDNN and other TensorFlow kernels -are in `TensorFlow LSTM benchmark `__. - -There is also `a wiki `_. -Questions can also be asked on -`StackOverflow using the RETURNN tag `_. - -.. image:: https://github.com/rwth-i6/returnn/workflows/CI/badge.svg - :target: https://github.com/rwth-i6/returnn/actions diff --git a/setup.py b/setup.py index d76ee99652..6d62de6553 100644 --- a/setup.py +++ b/setup.py @@ -87,8 +87,8 @@ def main(): author_email="albzey@gmail.com", url="https://github.com/rwth-i6/returnn/", license="RETURNN license", - long_description=open("README.rst").read(), - long_description_content_type="text/x-rst", + long_description=open("README.md").read(), + long_description_content_type="text/markdown", # https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Development Status :: 5 - Production/Stable", From 62134da092bfcefe686dfd26cd80cd2275c7fe79 Mon Sep 17 00:00:00 2001 From: Moritz Gunz Date: Fri, 13 Dec 2024 12:30:36 +0100 Subject: [PATCH 4/9] drop TF1 from CI --- .github/workflows/main.yml | 14 +------------- 1 file changed, 1 insertion(+), 13 deletions(-) diff --git a/.github/workflows/main.yml b/.github/workflows/main.yml index b4b2f45680..cefecd2bb2 100644 --- a/.github/workflows/main.yml +++ b/.github/workflows/main.yml @@ -174,21 +174,9 @@ jobs: - TEST=TFUtil - TEST=tools include: - - action: TEST=TFUtil - python-version: 3.8 - tf-version: 1.15.3 - - action: TEST=TFEngine - python-version: 3.8 - tf-version: 1.15.3 - action: TEST=TFEngine python-version: 3.8 tf-version: 2.3.0 - - action: TEST=TFNativeOp - python-version: 3.8 - tf-version: 1.14.0 - - action: TEST=TFNetworkSigProcLayer - python-version: 3.8 - tf-version: 1.15.3 - action: TEST=demos RETURNN_DISABLE_TORCH=1 python-version: 3.8 tf-version: 2.3.0 @@ -221,7 +209,7 @@ jobs: pip install --user pytest - if [[ "${{matrix.tf-version}}" == 2.[0123].* || "${{matrix.tf-version}}" == 1.* ]]; then + if [[ "${{matrix.tf-version}}" == 2.[0123].* ]]; then # Older TF needs older NumPy version. # https://github.com/rwth-i6/returnn/pull/1160#issuecomment-1284537803 pip install --user numpy==1.19.5 From 2f242b20d77032101f893fb939bb40848e6b8b7d Mon Sep 17 00:00:00 2001 From: Moritz Gunz Date: Fri, 13 Dec 2024 13:23:58 +0100 Subject: [PATCH 5/9] Revert "convert readme to markdown, add note about min python versions" This reverts commit c00d1f58ecc788b63265908b5ba1285b9bd15567. --- README.md | 72 ---------------------------------------------------- README.rst | 74 ++++++++++++++++++++++++++++++++++++++++++++++++++++++ setup.py | 4 +-- 3 files changed, 76 insertions(+), 74 deletions(-) delete mode 100644 README.md create mode 100644 README.rst diff --git a/README.md b/README.md deleted file mode 100644 index 3abd31c11f..0000000000 --- a/README.md +++ /dev/null @@ -1,72 +0,0 @@ -# Welcome to RETURNN - -[GitHub repository](https://github.com/rwth-i6/returnn), -[RETURNN paper 2016](https://arxiv.org/abs/1608.00895), -[RETURNN paper 2018](https://arxiv.org/abs/1805.05225). - -RETURNN - RWTH extensible training framework for universal recurrent neural networks, -is a PyTorch/TensorFlow-based implementation of modern neural network architectures. -It is optimized for fast and reliable training of neural networks in a multi-GPU environment. - -The high-level features and goals of RETURNN are: - -- **Simplicity** - - Writing config / code is simple & straight-forward (setting up experiment, defining model) - - Debugging in case of problems is simple - - Reading config / code is simple (defined model, training, decoding all becomes clear) - -- **Flexibility** - - Allow for many different kinds of experiments / models - -- **Efficiency** - - Training speed - - Decoding speed - -All items are important for research, decoding speed is esp. important for production. - -See our [Interspeech 2020 tutorial "Efficient and Flexible Implementation of Machine Learning for ASR and MT" video](https://www.youtube.com/watch?v=wPKdYqSOlAY) -([slides](https://www-i6.informatik.rwth-aachen.de/publications/download/1154/Zeyer--2020.pdf)) -with an introduction of the core concepts. - -More specific features include: - -- Mini-batch training of feed-forward neural networks -- Sequence-chunking based batch training for recurrent neural networks -- Long short-term memory recurrent neural networks - including our own fast CUDA kernel -- Multidimensional LSTM (GPU only, there is no CPU version) -- Memory management for large data sets -- Work distribution across multiple devices -- Flexible and fast architecture which allows all kinds of encoder-attention-decoder models - -See [documentation](https://returnn.readthedocs.io/). -See [basic usage](https://returnn.readthedocs.io/en/latest/basic_usage.html) and [technological overview](https://returnn.readthedocs.io/en/latest/tech_overview.html). - -[Here is the video recording of a RETURNN overview talk](https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.recording.cut.mp4) -([slides](https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.returnn-overview.session1.handout.v1.pdf), -[exercise sheet](https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.exercise_sheet.pdf); hosted by eBay). - -There are [many example demos](https://github.com/rwth-i6/returnn/blob/master/demos/) -which work on artificially generated data, -i.e. they should work as-is. - -There are [some real-world examples](https://github.com/rwth-i6/returnn-experiments) -such as setups for speech recognition on the Switchboard or LibriSpeech corpus. - -Some benchmark setups against other frameworks -can be found [here](https://github.com/rwth-i6/returnn-benchmarks). -The results are in the [RETURNN paper 2016](https://arxiv.org/abs/1608.00895). -Performance benchmarks of our LSTM kernel vs CuDNN and other TensorFlow kernels -are in [TensorFlow LSTM benchmark](https://returnn.readthedocs.io/en/latest/tf_lstm_benchmark.html). - -There is also [a wiki](https://github.com/rwth-i6/returnn/wiki). -Questions can also be asked on -[StackOverflow using the RETURNN tag](https://stackoverflow.com/questions/tagged/returnn). - -[![CI](https://github.com/rwth-i6/returnn/workflows/CI/badge.svg)](https://github.com/rwth-i6/returnn/actions) - -## Dependencies - -pip dependencies are listed in `requirements.txt` and `requirements-dev`, although some parts of the code may require additional dependencies (e.g. `librosa`, `resampy`) on-demand. - -RETURNN supports Python >= 3.8. Bumps to the minimum Python version are listed in [`CHANGELOG.md`](https://github.com/rwth-i6/returnn/blob/master/CHANGELOG.md). diff --git a/README.rst b/README.rst new file mode 100644 index 0000000000..4793b3fc0f --- /dev/null +++ b/README.rst @@ -0,0 +1,74 @@ +================== +Welcome to RETURNN +================== + +`GitHub repository `__. +`RETURNN paper 2016 `_, +`RETURNN paper 2018 `_. + +RETURNN - RWTH extensible training framework for universal recurrent neural networks, +is a Theano/TensorFlow-based implementation of modern recurrent neural network architectures. +It is optimized for fast and reliable training of recurrent neural networks in a multi-GPU environment. + +The high-level features and goals of RETURNN are: + +* **Simplicity** + + * Writing config / code is simple & straight-forward (setting up experiment, defining model) + * Debugging in case of problems is simple + * Reading config / code is simple (defined model, training, decoding all becomes clear) + +* **Flexibility** + + * Allow for many different kinds of experiments / models + +* **Efficiency** + + * Training speed + * Decoding speed + +All items are important for research, decoding speed is esp. important for production. + +See our `Interspeech 2020 tutorial "Efficient and Flexible Implementation of Machine Learning for ASR and MT" video `__ +(`slides `__) +with an introduction of the core concepts. + +More specific features include: + +- Mini-batch training of feed-forward neural networks +- Sequence-chunking based batch training for recurrent neural networks +- Long short-term memory recurrent neural networks + including our own fast CUDA kernel +- Multidimensional LSTM (GPU only, there is no CPU version) +- Memory management for large data sets +- Work distribution across multiple devices +- Flexible and fast architecture which allows all kinds of encoder-attention-decoder models + +See `documentation `__. +See `basic usage `__ +and `technological overview `__. + +`Here is the video recording of a RETURNN overview talk `_ +(`slides `__, +`exercise sheet `__; +hosted by eBay). + +There are `many example demos `_ +which work on artificially generated data, +i.e. they should work as-is. + +There are `some real-world examples `_ +such as setups for speech recognition on the Switchboard or LibriSpeech corpus. + +Some benchmark setups against other frameworks +can be found `here `_. +The results are in the `RETURNN paper 2016 `_. +Performance benchmarks of our LSTM kernel vs CuDNN and other TensorFlow kernels +are in `TensorFlow LSTM benchmark `__. + +There is also `a wiki `_. +Questions can also be asked on +`StackOverflow using the RETURNN tag `_. + +.. image:: https://github.com/rwth-i6/returnn/workflows/CI/badge.svg + :target: https://github.com/rwth-i6/returnn/actions diff --git a/setup.py b/setup.py index 6d62de6553..d76ee99652 100644 --- a/setup.py +++ b/setup.py @@ -87,8 +87,8 @@ def main(): author_email="albzey@gmail.com", url="https://github.com/rwth-i6/returnn/", license="RETURNN license", - long_description=open("README.md").read(), - long_description_content_type="text/markdown", + long_description=open("README.rst").read(), + long_description_content_type="text/x-rst", # https://pypi.python.org/pypi?%3Aaction=list_classifiers classifiers=[ "Development Status :: 5 - Production/Stable", From 51cf45559ddf0806107939b857f57caadafe17ca Mon Sep 17 00:00:00 2001 From: Moritz Gunz Date: Fri, 13 Dec 2024 13:29:14 +0100 Subject: [PATCH 6/9] add version notes to readme --- README.rst | 7 +++++++ 1 file changed, 7 insertions(+) diff --git a/README.rst b/README.rst index 4793b3fc0f..2835ececac 100644 --- a/README.rst +++ b/README.rst @@ -72,3 +72,10 @@ Questions can also be asked on .. image:: https://github.com/rwth-i6/returnn/workflows/CI/badge.svg :target: https://github.com/rwth-i6/returnn/actions + +Dependencies +============ + +pip dependencies are listed in ``requirements.txt`` and ``requirements-dev``, although some parts of the code may require additional dependencies (e.g. ``librosa``, ``resampy``) on-demand. + +RETURNN supports Python >= 3.8. Bumps to the minimum Python version are listed in `CHANGELOG.md `__. From 71a120968e4a1915e9b1df966c32b64707f47da5 Mon Sep 17 00:00:00 2001 From: Moritz Gunz Date: Fri, 13 Dec 2024 15:21:23 +0100 Subject: [PATCH 7/9] skip IEW TF logs --- tests/test_Log.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/test_Log.py b/tests/test_Log.py index 68b987f302..d504286a2b 100644 --- a/tests/test_Log.py +++ b/tests/test_Log.py @@ -52,7 +52,7 @@ def filter_out(ls): i = 0 while i < len(ls): s = ls[i] - if "tensorflow/core/" in s or "tensorflow/stream_executor/" in s: # some TF warnings + if any(f"{l} tensorflow/" in s for l in "IWE"): # some TF warnings i += 1 continue # RuntimeWarning|FutureWarning are warnings and they include the code-line in the next output line From b60fe240cbd416e6aa3c0a3bb4df281e9d973932 Mon Sep 17 00:00:00 2001 From: Moritz Gunz Date: Tue, 17 Dec 2024 10:11:08 +0100 Subject: [PATCH 8/9] skip syntax warnings --- tests/test_Log.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/tests/test_Log.py b/tests/test_Log.py index d504286a2b..2f8163ec00 100644 --- a/tests/test_Log.py +++ b/tests/test_Log.py @@ -55,6 +55,9 @@ def filter_out(ls): if any(f"{l} tensorflow/" in s for l in "IWE"): # some TF warnings i += 1 continue + elif "SyntaxWarning" in s: + i += 2 + continue # RuntimeWarning|FutureWarning are warnings and they include the code-line in the next output line if i + 1 < len(ls) and ls[i + 1].startswith(" "): if re.match(".*:\\d+: RuntimeWarning: numpy.*", s) or re.match(".*:\\d+: FutureWarning: .*", s): From 6aba446d47ee1e32008fc6bb33a7fc7797347b22 Mon Sep 17 00:00:00 2001 From: Moritz Gunz Date: Tue, 17 Dec 2024 10:26:32 +0100 Subject: [PATCH 9/9] extend readme --- README.rst | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/README.rst b/README.rst index 2835ececac..e6463de131 100644 --- a/README.rst +++ b/README.rst @@ -76,6 +76,11 @@ Questions can also be asked on Dependencies ============ -pip dependencies are listed in ``requirements.txt`` and ``requirements-dev``, although some parts of the code may require additional dependencies (e.g. ``librosa``, ``resampy``) on-demand. +pip dependencies are listed in ``requirements.txt`` and ``requirements-dev``, +although some parts of the code may require additional dependencies (e.g. ``librosa``, ``resampy``) on-demand. RETURNN supports Python >= 3.8. Bumps to the minimum Python version are listed in `CHANGELOG.md `__. + +TensorFlow-based setups require TensorFlow >= 2.2. + +PyTorch-based setups require Torch >= 1.0.