You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
https://doi.org/10.1103/PhysRevE.90.033315 describes some pruning methods to the single-term factorization (aka: the operation minimization of a tensor network), that would probably allow to speed this part of SeQuant's processing up quite a bit.
This might become very relevant in case one ever wants to treat expressions with a very large amount of tensors in a single network or if one wants to step into the direction of some sort of global-ish optimizations of the given expressions (beyond a single tensor network).
The text was updated successfully, but these errors were encountered:
Yes, this will be useful for when we have really large tensor networks. For tensor networks with 10 to 12 tensors, the current implementation is fast enough (very fast). Tree pruning might be implemented in the future.
https://doi.org/10.1103/PhysRevE.90.033315 describes some pruning methods to the single-term factorization (aka: the operation minimization of a tensor network), that would probably allow to speed this part of SeQuant's processing up quite a bit.
This might become very relevant in case one ever wants to treat expressions with a very large amount of tensors in a single network or if one wants to step into the direction of some sort of global-ish optimizations of the given expressions (beyond a single tensor network).
The text was updated successfully, but these errors were encountered: