The following pulls the trigger, defaulting --param vect-force-slp to 1.
I know of no features missing but eventually minor testsuite and
optimization quality fallout.
Bootstrapped and tested on x86_64-unknown-linux-gnu. I'll amend
PR116578 with the list of FAILs this causes (my baseline is outdated,
need to reproduce it).
* params.opt (vect-force-slp): Default to 1.
---
gcc/params.opt | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/gcc/params.opt b/gcc/params.opt
index 7c572774df2..10a2b089279 100644
--- a/gcc/params.opt
+++ b/gcc/params.opt
@@ -1207,7 +1207,7 @@ Common Joined UInteger Var(param_vect_induction_float)
Init(1) IntegerRange(0, 1
Enable loop vectorization of floating point inductions.
-param=vect-force-slp=
-Common Joined UInteger Var(param_vect_force_slp) Init(0) IntegerRange(0, 1)
Param Optimization
+Common Joined UInteger Var(param_vect_force_slp) Init(1) IntegerRange(0, 1)
Param Optimization
Force the use of SLP when vectorizing, fail if not possible.
-param=vrp-block-limit=
--
2.43.0