-
Notifications
You must be signed in to change notification settings - Fork 0
Do Not Merge: Integration Branch for GT4Py Next #1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Draft
philip-paul-mueller
wants to merge
27
commits into
main
Choose a base branch
from
gt4py-next-integration
base: main
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This was referenced Apr 30, 2025
philip-paul-mueller
added a commit
to GridTools/gt4py
that referenced
this pull request
Apr 30, 2025
Instead of pulling directly from the official DaCe repo, we now (for the time being) pull from [this PR](GridTools/dace#1). This became necessary as we have a lot of open PR in DaCe and need some custom fixes (that can in their current form not be merged into DaCe). In the long term however, we should switch back to the main DaCe repo.
7fcf8f9 to
8b9b674
Compare
8b9b674 to
268fc18
Compare
964e84b to
2d85437
Compare
2d85437 to
9f72250
Compare
88c99f4 to
d779cd1
Compare
d779cd1 to
4f40029
Compare
4f40029 to
09dfda3
Compare
09dfda3 to
87c77ef
Compare
87c77ef to
c2a4e42
Compare
c2a4e42 to
0deba99
Compare
Merged
2 tasks
178037a to
9114985
Compare
33b63a1 to
2417e09
Compare
3472895 to
bed3b0e
Compare
Increased pytest timeout from 300 to 600 seconds.
## Refactor `dace/data.py` into `dace/data/` package
### Summary
This PR refactors the monolithic `dace/data.py` file into a modular
`dace/data/` package with separate files for different functionality,
improving code organization and maintainability.
### Changes
- [x] **`dace/data/core.py`**: Core data descriptor classes (`Data`,
`Scalar`, `Array`, `ContainerArray`, `Stream`, `Structure`, `View`,
`Reference` and their subclasses)
- [x] **`dace/data/tensor.py`**: Tensor/sparse tensor support (`Tensor`,
`TensorIndex*` classes)
- [x] **`dace/data/creation.py`**: Data descriptor creation functions
(`create_datadescriptor`, `make_array_from_descriptor`,
`make_reference_from_descriptor`)
- [x] **`dace/data/ctypes_interop.py`**: Ctypes interoperability
(`make_ctypes_argument`)
- [x] **`dace/data/ml.py`**: ML-related descriptors (`ParameterArray`)
- [x] **`dace/data/__init__.py`**: Re-exports all public API for
backward compatibility
- [x] **`dace/utils.py`**: Utility functions (`find_new_name`,
`deduplicate`, `prod`)
- [x] **`dace/properties.py`**: Updated to handle circular import
gracefully
- [x] **`dace/autodiff/library/library.py`**: Updated to import
`ParameterArray` from the new location
- [x] **Deleted** old `dace/data.py` file
- [x] **Removed** `Number` and `ArrayLike` from `dace/data/__init__.py`
(other places import directly)
- [x] **Moved** `_prod` to `dace/utils.py` as `prod` (kept `_prod`
export for backward compat)
- [x] **Fixed** broken imports in `data_report.py`,
`data_layout_tuner.py`, and `cutout.py`
### Backward Compatibility
All public APIs are re-exported from `dace.data`, ensuring backward
compatibility with existing code.
<!-- START COPILOT CODING AGENT SUFFIX -->
<details>
<summary>Original prompt</summary>
>
> ----
>
> *This section details on the original issue you should resolve*
>
> <issue_title>Refactor `dace/data.py`</issue_title>
> <issue_description>`data.py` is a monolithic file containing classes
for core data containers (Data, Scalar, Array, Stream, View, Reference,
and their subclasses `*{View, Reference}`; functionality to get data
descriptors from arbitrary objects; derived objects for Tensors and
sparse tensors; and other functions.
>
> This issue will be resolved once `data.py` is refactored to a
`dace/data/*` folder, which will contain separate files for:
> 1. core descriptor classes
> 2. structures (the Structure class and similar functionality)
> 3. tensors/sparse tensors
> 4. descriptor creation
> 5. ML-related data descriptors, such as parameter arrays (see
`dace/autodiff/library/library.py`)
> 6...N. Other functions and classes categorized by their semantic
meaning.
>
> The code for `dace/data/*` will be refactored out of `data.py` (which
should not exist at the end of this issue), `dtypes.py` (which may exist
but be shorter), and other files that contain data descriptors
(subclasses of Data/Array/Stream/Structure/View/Reference, such as
ParameterArray. Try to find all such subclasses in the codebase barring
tests/* and samples/*).
>
> Lastly, utility functions in `data.py` and `dtypes.py` (only those two
files for this issue), such as `find_new_name` from data.py and
`deduplicate` from dtypes.py, should find themselves in a new
`dace/utils.py` file.</issue_description>
>
> ## Comments on the Issue (you are @copilot in this section)
>
> <comments>
> </comments>
>
</details>
- Fixes spcl#2244
<!-- START COPILOT CODING AGENT TIPS -->
---
💡 You can make Copilot smarter by setting up custom instructions,
customizing its development environment and configuring Model Context
Protocol (MCP) servers. Learn more [Copilot coding agent
tips](https://gh.io/copilot-coding-agent-tips) in the docs.
---------
Co-authored-by: copilot-swe-agent[bot] <198982749+Copilot@users.noreply.github.com>
Co-authored-by: tbennun <8348955+tbennun@users.noreply.github.com>
…to seq. maps inside GPU kernels or gpu dev. maps (spcl#2088) GPU codegen crashes and generates incorrect code with dynamic inputs to seq. maps inside GPU kernels or gpu dev. maps --------- Co-authored-by: alexnick83 <31545860+alexnick83@users.noreply.github.com> Co-authored-by: Tal Ben-Nun <tbennun@users.noreply.github.com>
…pcl#2246) Updated the documentation for proposed pass decomposition, including changes to pass names and descriptions for clarity.
3eec6f6 to
ab9eaef
Compare
e552808 to
d4db8e7
Compare
d4db8e7 to
cd52c4b
Compare
The current version of DaCe only supports CUDA 12 since CUDA 13 has a few breaking changes. This PR updates the runtime headers according to https://nvidia.github.io/cccl/cccl/3.0_migration_guide.html, replacing classes that moved from cub to thrust. As I have no way of testing on CUDA 12, I have conservatively left the old code intact using preprocessor macros. Tested on CUDA 13.1
…spcl#2251) The values retrieved from `Config` are strings. For block_size configuration, the values need to be converted to `int` type. Co-authored-by: Philipp Schaad <schaad.phil@gmail.com> Co-authored-by: Tal Ben-Nun <tbennun@users.noreply.github.com>
As part of bringing our schedule tree branch from `v1/maintenance` to `main` (see spcl#2262), I have fixed a couple of typos. I'd like to pull these out into a separate PR to keep the schedule tree PR as small as possible (which will still be large once done). Co-authored-by: Roman Cattaneo <1116746+romanc@users.noreply.github.com>
The DaCe FPGA backend will be supported out-of-tree. This PR extracts the capabilities to a separate repository: https://github.com/spcl/dace-fpga This PR also reduces circular imports and encourages more object-oriented design for code generation.
- Added NVTX/rocTX markers to SDFGs, states, scopes and copies in the CPU code to analyze and measure easily the DaCe runtime using NVIDIA/AMD profiling tools - Added related test and documentation --------- Co-authored-by: Tal Ben-Nun <tbennun@users.noreply.github.com> Co-authored-by: Philip Müller <147368808+philip-paul-mueller@users.noreply.github.com> Co-authored-by: Philip Mueller <philip.paul.mueller@bluemain.ch> Co-authored-by: Tal Ben-Nun <tbennun@gmail.com>
0362ff3 to
54b3349
Compare
Although the GPU codegen is being reworked, this small change would be useful for the time being. It a an extension to the HIP codegen of the current solution for stream-order memory allocation. It has been tested in GT4Py.
- [x] Remove ArrayInterface - [x] Remove other FPGA-related enumeration entries - [x] Mitigate extraneous arguments in `dace.codegen.targets.cpp` - [x] Remove runtime includes and references to DaCe-FPGA macros - [x] Address TODOs from spcl#2252
Allow the compiler to unroll itself the `LoopRegion`s instead of just having a manual unroll feature with the `LoopUnroll` pass
This enables array-like objects to be used in callbacks and retain the original object.
This new type enables DaCe users to perform calculations with stochastic rounding in single precision. This change is validated with unit tests. --------- Co-authored-by: Tal Ben-Nun <tbennun@users.noreply.github.com>
Allow compilation with latest LLVM versions. Without this change there's
the following error:
```
> raise cgx.CompilationError('Compiler failure:\n' + ex.output)
E dace.codegen.exceptions.CompilationError: Compiler failure:
E [ 20%] Building CXX object CMakeFiles/vertically_implicit_solver_at_predictor_step.dir/capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/dycore_711_res_tmps_rcu/.gt4py_cache/vertically_implicit_solver_at_predictor_step_c2100e7df02fe12899d3acf60eb0b974edf9029c01b12b0414e215257065bc8c/src/cpu/vertically_implicit_solver_at_predictor_step.cpp.o
E [ 40%] Building HIP object CMakeFiles/vertically_implicit_solver_at_predictor_step.dir/capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/dycore_711_res_tmps_rcu/.gt4py_cache/vertically_implicit_solver_at_predictor_step_c2100e7df02fe12899d3acf60eb0b974edf9029c01b12b0414e215257065bc8c/src/cuda/hip/vertically_implicit_solver_at_predictor_step_cuda.cpp.o
E In file included from /capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/dycore_711_res_tmps_rcu/.gt4py_cache/vertically_implicit_solver_at_predictor_step_c2100e7df02fe12899d3acf60eb0b974edf9029c01b12b0414e215257065bc8c/src/cuda/hip/vertically_implicit_solver_at_predictor_step_cuda.cpp:3:
E In file included from /capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/.venv/lib/python3.12/site-packages/dace/codegen/../runtime/include/dace/dace.h:17:
E /capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/.venv/lib/python3.12/site-packages/dace/codegen/../runtime/include/dace/math.h:509:37: warning: 'host' attribute only applies to functions [-Wignored-attributes]
E 509 | static const __attribute__((host)) __attribute__((device)) typeless_pi pi{};
E | ^
E In file included from /capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/dycore_711_res_tmps_rcu/.gt4py_cache/vertically_implicit_solver_at_predictor_step_c2100e7df02fe12899d3acf60eb0b974edf9029c01b12b0414e215257065bc8c/src/cuda/hip/vertically_implicit_solver_at_predictor_step_cuda.cpp:3:
E In file included from /capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/.venv/lib/python3.12/site-packages/dace/codegen/../runtime/include/dace/dace.h:30:
E /capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/.venv/lib/python3.12/site-packages/dace/codegen/../runtime/include/dace/cuda/copy.cuh:772:41: error: a template argument list is expected after a name prefixed by the template keyword [-Wmissing-template-arg-list-after-template-kw]
E 772 | wcr_custom<T>::template reduce(
E | ^
E /capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/.venv/lib/python3.12/site-packages/dace/codegen/../runtime/include/dace/cuda/copy.cuh:779:45: error: a template argument list is expected after a name prefixed by the template keyword [-Wmissing-template-arg-list-after-template-kw]
E 779 | wcr_custom<T>::template reduce(
E | ^
E /capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/.venv/lib/python3.12/site-packages/dace/codegen/../runtime/include/dace/cuda/copy.cuh:796:49: error: a template argument list is expected after a name prefixed by the template keyword [-Wmissing-template-arg-list-after-template-kw]
E 796 | wcr_fixed<REDTYPE, T>::template reduce_atomic(
E | ^
E /capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/.venv/lib/python3.12/site-packages/dace/codegen/../runtime/include/dace/cuda/copy.cuh:803:53: error: a template argument list is expected after a name prefixed by the template keyword [-Wmissing-template-arg-list-after-template-kw]
E 803 | wcr_fixed<REDTYPE, T>::template reduce_atomic(
E | ^
E 1 warning and 4 errors generated when compiling for gfx942.
E gmake[2]: *** [CMakeFiles/vertically_implicit_solver_at_predictor_step.dir/build.make:92: CMakeFiles/vertically_implicit_solver_at_predictor_step.dir/capstor/scratch/cscs/ioannmag/beverin/cycle33/ICON4PY/icon4py/dycore_711_res_tmps_rcu/.gt4py_cache/vertically_implicit_solver_at_predictor_step_c2100e7df02fe12899d3acf60eb0b974edf9029c01b12b0414e215257065bc8c/src/cuda/hip/vertically_implicit_solver_at_predictor_step_cuda.cpp.o] Error 1
E gmake[1]: *** [CMakeFiles/Makefile2:90: CMakeFiles/vertically_implicit_solver_at_predictor_step.dir/all] Error 2
E gmake: *** [Makefile:91: all] Error 2
.venv/lib/python3.12/site-packages/dace/codegen/compiler.py:254: CompilationError
```
This PR refactors how calling an SDFG works. [PR#1467](github.com/spcl/pull/1467) introduced the `fast_call()` API, which allowed to call a compiled SDFG and skipping some tests. This was done to support the use case of calling the same SDFG with the same (as in pointers) multiple times. However, the PR did not introduced a simple way to generate the argument vector that had to be passed to `fast_call()` without relying on internal implementation details of the class. This PR, beside other things, introduces this use case and give access to the all steps needed to call an SDFG: - `construct_arguments()`: It accepts Python arguments, such as `int` or NumPy arrays and turns them into an argument vector in the right order and converted to the required C type. - `fast_call()`: Performs the actual call using the _passed_ argument vectors, if needed it will also run initialization. Note that this function is not new, but was slightly modified and no longer handles the return values, see below. - `convert_return_values()`: This function performs the actual `return` operation, i.e. composes the specified return type, i.e. either a single array or a tuple. Note, before this function was called by `fast_call()` but it was moved outside to reduce the hot path, because usually return values are passed using inout or out arguments directly. Beside these changes the PR also modifies the following things: - It was possible to pass return values, i.e. `__return`, as ordinary arguments, this is still supported, but now a warning is returned. - `CompiledSDFG` was technically able to handle scalar return values, however, due to technical limitation this is [not possible](spcl#1609), thus the feature was removed. However, the feature was not dropped completely and is still used to handle `pyobjects` since they are passed as pointer, see below. - The handling of `pyobject` return values was modified. Before it was not possible to use `pyobject` instances as return values that were manged _outside_, i.e. not allocated by, `CompiledSDFG`, now they are "handled". It is important that an _array_, i.e. multiple instances, of `pyobject`s are handled as a single object (this is the correct behaviour and retained for bug compatibility with the unit tests), however, a warning is generated. - It was possible to pass an argument as named argument and as positional argument, this is now forbidden. - `safe_call()` is not possible to handle return values, if the method is called on such an SDFG an error is generated. - Before it was not possible to return a `tuple` with a single argument, in that case the value was always directly returned, this has been fixed and is correctly handled. - The allocation of return values was inconsistent. If there was no change in size, then `__call__()` would always return the _same_ arrays, which might lead to very sudden bugs. The new behaviour is to always allocate new memory, this is done by `construct_arguments()`. - Shared return values. - Before `CompiledSDFG` had a member `_lastargs` which "cached" the last pointer arguments that were used to call the SDFG. It was updated by `_construct_args()` (old version of `construct_arguments()`), which did not make much sense. The original intention was to remove it, but this proved to be harder and it is thus maintained. However, it is now updated by `__call__()` and `initialize()` to support the use case for `{get, set}_workspace_size()`. --------- Co-authored-by: Philipp Schaad <schaad.phil@gmail.com> Co-authored-by: Tal Ben-Nun <tbennun@gmail.com>
commit 7e05c76 Author: Philip Mueller <philip.mueller@cscs.ch> Date: Wed Jan 7 10:08:42 2026 +0100 Rename of the Index. commit 32224bb Author: Philip Mueller <philip.mueller@cscs.ch> Date: Thu Dec 18 09:53:57 2025 +0100 Updated the workflow file. commit ecb2785 Author: Philip Mueller <philip.mueller@cscs.ch> Date: Wed Dec 17 08:19:40 2025 +0100 Updated the dace updater workflow file. commit f3198ef Author: Philip Mueller <philip.mueller@cscs.ch> Date: Wed Dec 17 07:41:26 2025 +0100 Made the update point to the correct repo. commit 96f963a Merge: 8b7cce5 387f1e8 Author: Philip Mueller <philip.mueller@cscs.ch> Date: Wed Dec 17 07:37:48 2025 +0100 Merge remote-tracking branch 'spcl/main' into automatic_gt4py_deployment commit 8b7cce5 Author: Philip Mueller <philip.mueller@cscs.ch> Date: Mon Dec 1 09:18:22 2025 +0100 Restored the original workflow files. commit 362ab70 Author: Philip Mueller <philip.mueller@cscs.ch> Date: Mon Dec 1 07:41:40 2025 +0100 Now it has run once, so let's make it less runnable. commit 81b8cfa Author: Philip Mueller <philip.mueller@cscs.ch> Date: Mon Dec 1 07:39:09 2025 +0100 Made it run always. commit 6d71466 Author: Philip Mueller <philip.mueller@cscs.ch> Date: Mon Dec 1 07:38:11 2025 +0100 Small update. commit eb31e6c Author: Philip Mueller <philip.mueller@cscs.ch> Date: Fri Nov 21 15:23:33 2025 +0100 Empty commit in the branch containing the workflow file. commit 2970a75 Author: Philip Mueller <philip.mueller@cscs.ch> Date: Fri Nov 21 15:21:09 2025 +0100 Next step. commit f5d3d9d Author: Philip Mueller <philip.mueller@cscs.ch> Date: Fri Nov 21 15:17:56 2025 +0100 Let's disable everything. commit 211e415 Author: Philip Mueller <philip.mueller@cscs.ch> Date: Fri Nov 21 15:10:43 2025 +0100 Disabled the kickstarter. commit d012c26 Author: Philip Mueller <philip.mueller@cscs.ch> Date: Fri Nov 21 15:05:38 2025 +0100 Updated everything.
commit 7c8e1d3 Merge: 5b068e7 a27e416 Author: Afif <37773945+affifboudaoud@users.noreply.github.com> Date: Thu Jan 15 14:40:56 2026 +0100 Merge branch 'main' into fix_map_fusion_vertical_dfs commit 5b068e7 Author: Affifboudaoud <hk_boudaoud@esi.dz> Date: Sun Nov 23 22:50:46 2025 +0100 Add visited set to avoid visiting same node multiple times
513e824 to
139389d
Compare
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This is the PR/branch that GT4Py.Next uses to pull DaCe.
It is essentially DaCe main together with our fixes that, for various reasons have not made it yet into DaCe main.
The process for updating this branch is is as follows there are no exceptions:
version.pyfile in thedace/subfolder. Make sure that there is no new line at the end. Fornextwe are using the epoch 43,cartesianwould use 42. As version number the date is used. Thus the version (fornext) would look something like:'43!YYYY.MM.DD'.gt4py-next-integration).__gt4py-next-integration_YYYY_MM_DDand push it as well.Afterwards you have to update GT4Py's
pyproject.tomlfile.For this you have to update the version requirement of DaCe in the
dace-nextgroup at the beginning of the file to the version you just created, i.e. change it todace==43!YYYY.MM.DD.Then you have to update the the source in the
uvspecific parts of the file, there you have to change the source to the new tag you have just created.Then you have to update the uv look by running
uv sync --extra next --group dace-next, if you have installed the precommit hooks then this will be done automatically.NOTE: Once PR#2423 has been merged the second step, i.e. adapting the tag in the
uvspecific parts is no longer needed.On top of
DaCe/mainwe are using the following PRs:MapFusionVerticalNo Longer Needed
DaCe.ConfigPruneSymbolsscope_tree_recursive()MapFusionother_subsetvalidationstate_fission()SubgraphViewtry_initialize()edges in Map fusionMapFusionVerticalRedundantSecondArrayimportinfast_call()compiled_sdfg_call_hooks_managerself._lastargsMutable No longer needed since we now use GT4Py PR#2353 and DaCe PR#2206.self._lastargsMutable (Should be replaced by a more permanent solution).MapFusion*AddThreadBlockMapapply_transformation_once_everywhere()CompiledSDFGrefactoring (archive):For some reason the original PR has been "taken over" by Tal.
Due to the inherent dependency that GT4Py has on this PR we should use the the archive (liked at the top).