Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.
Comment: Adding a few suggestions.

...

  • info on how a new rule compares to the full *existing* ruleset
  • overlaps between rules, using "hit-frequencies -o"
  • collated results across all users' corpora, which can be broken down to view each user separately or all together
  • checked rules and their results are kept in a version-control history, so benefits of VC are available
  • ongoing visibility of hit-rates of the existing ruleset, against fresh corpora

(LOAFER: Suggestion: It would be good to know the % runtime figure for a sandbox rule as a missing boundary can take a rule from 1.5% to 0.0n% performance hit easily

No Format

 perl -d:DProf mass-check -j=1 spam:dir:some_reasonable_sample_set_including_hits_and_misses
 dprofpp -O 2000 > perf.log

Someway of scheduling a small run during the development day would be useful, rather than waiting for the nightly. An email of users completed results would be nice to see too.
)

(TODO: migrate the ruleqa CGI onto the SpamAssassin zone so this is still visible, even though the automc stuff is disabled)

...

JustinMason: 'I'd like to see if there's a way to combine the two (that is, nightly and list-driven mass-checks) somehow, so that new SVN commits that update sandbox rules, are immediately mass-checked alone. However, I can't see a way to do that reliably from SVN commits alone, because (for example) meta rules may depend on other rules that were not changed as part of the same commit. So I think the "email with attached rules file" is still a better model.'
LOAFER: There are eval rules to consider too.