1 # Copyright (c) 2013-2014 OPEN CASCADE SAS
3 # This file is part of Open CASCADE Technology software library.
5 # This library is free software; you can redistribute it and/or modify it under
6 # the terms of the GNU Lesser General Public License version 2.1 as published
7 # by the Free Software Foundation, with special exception defined in the file
8 # OCCT_LGPL_EXCEPTION.txt. Consult the file LICENSE_LGPL_21.txt included in OCCT
9 # distribution for complete text of the license and disclaimer of any warranty.
11 # Alternatively, this file may be used under the terms of Open CASCADE
12 # commercial license or contractual agreement.
14 ############################################################################
15 # This file defines scripts for execution of OCCT tests.
16 # It should be loaded automatically when DRAW is started, and provides
17 # top-level commands starting with 'test'. Type 'help test' to get their
19 # See OCCT Tests User Guide for description of the test system.
21 # Note: procedures with names starting with underscore are for internal use
22 # inside the test system.
23 ############################################################################
25 # Default verbose level for command _run_test
28 # regexp for parsing test case results in summary log
29 set _test_case_regexp {^CASE\s+([\w.-]+)\s+([\w.-]+)\s+([\w.-]+)\s*:\s*([\w]+)(.*)}
31 # Basic command to run indicated test case in DRAW
33 Run specified test case
34 Use: test group grid casename [options...]
36 -echo: all commands and results are echoed immediately,
37 but log is not saved and summary is not produced
38 It is also possible to use "1" instead of "-echo"
39 If echo is OFF, log is stored in memory and only summary
40 is output (the log can be obtained with command "dlog get")
41 -outfile filename: set log file (should be non-existing),
42 it is possible to save log file in text file or
43 in html file(with snapshot), for that "filename"
44 should have ".html" extension
45 -overwrite: force writing log in existing file
46 -beep: play sound signal at the end of the test
47 -errors: show all lines from the log report that are recognized as errors
48 This key will be ignored if the "-echo" key is already set.
50 proc test {group grid casename {args {}}} {
51 # set default values of arguments
58 # get test case paths (will raise error if input is invalid)
59 _get_test $group $grid $casename dir gridname casefile
62 for {set narg 0} {$narg < [llength $args]} {incr narg} {
63 set arg [lindex $args $narg]
64 # if echo specified as "-echo", convert it to bool
65 if { $arg == "-echo" || $arg == "1" } {
71 if { $arg == "-outfile" } {
73 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
74 set logfile [lindex $args $narg]
76 error "Option -outfile requires argument"
81 # allow overwrite existing log
82 if { $arg == "-overwrite" } {
87 # sound signal at the end of the test
88 if { $arg == "-beep" } {
93 # if errors specified as "-errors", convert it to bool
94 if { $arg == "-errors" } {
100 error "Error: unsupported option \"$arg\""
103 uplevel _run_test $dir $group $gridname $casefile $echo
107 _check_log $dir $group $gridname $casename $errors [dlog get] summary html_log
110 if { ! $overwrite && [file isfile $logfile] } {
111 error "Error: Specified log file \"$logfile\" exists; please remove it before running test or use -overwrite option"
113 if {$logfile != ""} {
114 if {[file extension $logfile] == ".html"} {
115 if {[regexp {vdump ([^\s\n]+)} $html_log dump snapshot]} {
116 catch {file copy -force $snapshot [file rootname $logfile][file extension $snapshot]}
118 _log_html $logfile $html_log "Test $group $grid $casename"
120 _log_save $logfile "[dlog get]\n$summary" "Test $group $grid $casename"
125 # play sound signal at the end of test
132 # Basic command to run indicated test case in DRAW
134 Run all tests, or specified group, or one grid
135 Use: testgrid [groupmask [gridmask [casemask]]] [options...]
137 -parallel N: run N parallel processes (default is number of CPUs, 0 to disable)
138 -refresh N: save summary logs every N seconds (default 600, minimal 1, 0 to disable)
139 -outdir dirname: set log directory (should be empty or non-existing)
140 -overwrite: force writing logs in existing non-empty directory
141 -xml filename: write XML report for Jenkins (in JUnit-like format)
142 -beep: play sound signal at the end of the tests
143 -regress dirname: re-run only a set of tests that have been detected as regressions on some previous run.
144 Here "dirname" is path to directory containing results of previous run.
145 Groups, grids, and test cases to be executed can be specified by list of file
146 masks, separated by spaces or comma; default is all (*).
148 proc testgrid {args} {
149 global env tcl_platform _tests_verbose
151 ######################################################
153 ######################################################
155 # check that environment variable defining paths to test scripts is defined
156 if { ! [info exists env(CSF_TestScriptsPath)] ||
157 [llength $env(CSF_TestScriptsPath)] <= 0 } {
158 error "Error: Environment variable CSF_TestScriptsPath is not defined"
162 set parallel [_get_nb_cpus]
170 for {set narg 0} {$narg < [llength $args]} {incr narg} {
171 set arg [lindex $args $narg]
174 if { $arg == "-parallel" } {
176 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
177 set parallel [expr [lindex $args $narg]]
179 error "Option -parallel requires argument"
185 if { $arg == "-refresh" } {
187 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
188 set refresh [expr [lindex $args $narg]]
190 error "Option -refresh requires argument"
196 if { $arg == "-outdir" } {
198 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
199 set logdir [lindex $args $narg]
201 error "Option -outdir requires argument"
206 # allow overwrite logs
207 if { $arg == "-overwrite" } {
213 if { $arg == "-xml" } {
215 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
216 set xmlfile [lindex $args $narg]
218 if { $xmlfile == "" } {
219 set xmlfile TESTS-summary.xml
224 # sound signal at the end of the test
225 if { $arg == "-beep" } {
230 # re-run only a set of tests that have been detected as regressions on some previous run
231 if { $arg == "-regress" } {
233 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
234 set prev_logdir [lindex $args $narg]
237 error "Option -regress requires argument"
243 if { [regexp {^-} $arg] } {
244 error "Error: unsupported option \"$arg\""
247 # treat arguments not recognized as options as group and grid names
248 if { ! [info exists groupmask] } {
249 set groupmask [split $arg ,]
250 } elseif { ! [info exists gridmask] } {
251 set gridmask [split $arg ,]
252 } elseif { ! [info exists casemask] } {
253 set casemask [split $arg ,]
255 error "Error: cannot interpret argument $narg ($arg)"
259 # check that target log directory is empty or does not exist
260 set logdir [file normalize [string trim $logdir]]
261 set prev_logdir [file normalize [string trim $prev_logdir]]
262 if { $logdir == "" } {
263 # if specified logdir is empty string, generate unique name like
264 # results/<branch>_<timestamp>
266 if { ! [catch {exec git branch} gitout] &&
267 [regexp {[*] ([\w-]+)} $gitout res branch] } {
268 set prefix "${branch}_"
270 set logdir "results/${prefix}[clock format [clock seconds] -format {%Y-%m-%dT%H%M}]"
272 set logdir [file normalize $logdir]
274 if { [file isdirectory $logdir] && ! $overwrite && ! [catch {glob -directory $logdir *}] } {
275 error "Error: Specified log directory \"$logdir\" is not empty; please clean it before running tests"
277 if { [catch {file mkdir $logdir}] || ! [file writable $logdir] } {
278 error "Error: Cannot create directory \"$logdir\", or it is not writable"
281 # masks for search of test groups, grids, and cases
282 if { ! [info exists groupmask] } { set groupmask * }
283 if { ! [info exists gridmask ] } { set gridmask * }
284 if { ! [info exists casemask ] } { set casemask * }
286 # Find test cases with FAILED and IMPROVEMENT statuses in previous run
287 # if option "regress" is given
288 set rerun_group_grid_case {}
290 if { ${regress} > 0 } {
291 if { "${groupmask}" != "*"} {
292 lappend rerun_group_grid_case [list $groupmask $gridmask $casemask]
295 lappend rerun_group_grid_case [list $groupmask $gridmask $casemask]
298 if { ${regress} > 0 } {
299 if { [file exists ${prev_logdir}/tests.log] } {
300 set fd [open ${prev_logdir}/tests.log]
301 while { [gets $fd line] >= 0 } {
302 if {[regexp {CASE ([^\s]+) ([^\s]+) ([^\s]+): FAILED} $line dump group grid casename] ||
303 [regexp {CASE ([^\s]+) ([^\s]+) ([^\s]+): IMPROVEMENT} $line dump group grid casename]} {
304 lappend rerun_group_grid_case [list $group $grid $casename]
309 error "Error: file ${prev_logdir}/tests.log is not found, check your input arguments!"
313 ######################################################
314 # prepare list of tests to be performed
315 ######################################################
317 # list of tests, each defined by a list of:
318 # test scripts directory
319 # group (subfolder) name
320 # grid (subfolder) name
322 # path to test case file
325 foreach group_grid_case ${rerun_group_grid_case} {
326 set groupmask [lindex $group_grid_case 0]
327 set gridmask [lindex $group_grid_case 1]
328 set casemask [lindex $group_grid_case 2]
330 # iterate by all script paths
331 foreach dir [lsort -unique [_split_path $env(CSF_TestScriptsPath)]] {
332 # protection against empty paths
333 set dir [string trim $dir]
334 if { $dir == "" } { continue }
336 if { $_tests_verbose > 0 } { _log_and_puts log "Examining tests directory $dir" }
338 # check that directory exists
339 if { ! [file isdirectory $dir] } {
340 _log_and_puts log "Warning: directory $dir listed in CSF_TestScriptsPath does not exist, skipped"
344 # search all directories in the current dir with specified mask
345 if [catch {glob -directory $dir -tail -types d {*}$groupmask} groups] { continue }
348 if { $_tests_verbose > 0 } { _log_and_puts log "Groups to be executed: $groups" }
349 foreach group [lsort -dictionary $groups] {
350 if { $_tests_verbose > 0 } { _log_and_puts log "Examining group directory $group" }
352 # file grids.list must exist: it defines sequence of grids in the group
353 if { ! [file exists $dir/$group/grids.list] } {
354 _log_and_puts log "Warning: directory $dir/$group does not contain file grids.list, skipped"
358 # read grids.list file and make a list of grids to be executed
360 set fd [open $dir/$group/grids.list]
362 while { [gets $fd line] >= 0 } {
365 # skip comments and empty lines
366 if { [regexp "\[ \t\]*\#.*" $line] } { continue }
367 if { [string trim $line] == "" } { continue }
369 # get grid id and name
370 if { ! [regexp "^\(\[0-9\]+\)\[ \t\]*\(\[A-Za-z0-9_.-\]+\)\$" $line res gridid grid] } {
371 _log_and_puts log "Warning: cannot recognize line $nline in file $dir/$group/grids.list as \"gridid gridname\"; ignored"
375 # check that grid fits into the specified mask
376 foreach mask $gridmask {
377 if { $mask == $gridid || [string match $mask $grid] } {
378 lappend gridlist $grid
384 # iterate by all grids
385 foreach grid $gridlist {
387 # check if this grid is aliased to another one
388 set griddir $dir/$group/$grid
389 if { [file exists $griddir/cases.list] } {
390 set fd [open $griddir/cases.list]
391 if { [gets $fd line] >= 0 } {
392 set griddir [file normalize $dir/$group/$grid/[string trim $line]]
397 # check if grid directory actually exists
398 if { ! [file isdirectory $griddir] } {
399 _log_and_puts log "Error: tests directory for grid $grid ($griddir) is missing; skipped"
403 # create directory for logging test results
404 if { $logdir != "" } { file mkdir $logdir/$group/$grid }
406 # iterate by all tests in the grid directory
407 if { [catch {glob -directory $griddir -type f {*}$casemask} testfiles] } { continue }
408 foreach casefile [lsort -dictionary $testfiles] {
409 # filter out files with reserved names
410 set casename [file tail $casefile]
411 if { $casename == "begin" || $casename == "end" ||
412 $casename == "parse.rules" } {
416 lappend tests_list [list $dir $group $grid $casename $casefile]
422 if { [llength $tests_list] < 1 } {
423 error "Error: no tests are found, check your input arguments and variable CSF_TestScriptsPath!"
425 puts "Running tests (total [llength $tests_list] test cases)..."
428 ######################################################
430 ######################################################
432 # log command arguments and environment
433 lappend log "Command: testgrid $args"
434 lappend log "Host: [info hostname]"
435 lappend log "Started on: [clock format [clock seconds] -format {%Y-%m-%d %H:%M:%S}]"
436 catch {lappend log "DRAW build:\n[dversion]" }
437 lappend log "Environment:"
438 foreach envar [lsort [array names env]] {
439 lappend log "$envar=\"$env($envar)\""
443 set refresh_timer [clock seconds]
444 uplevel dchrono _timer reset
445 uplevel dchrono _timer start
447 # if parallel execution is requested, allocate thread pool
448 if { $parallel > 0 } {
449 if { ! [info exists tcl_platform(threaded)] || [catch {package require Thread}] } {
450 _log_and_puts log "Warning: Tcl package Thread is not available, running in sequential mode"
453 set worker [tpool::create -minworkers $parallel -maxworkers $parallel]
454 # suspend the pool until all jobs are posted, to prevent blocking of the process
455 # of starting / processing jobs by running threads
456 catch {tpool::suspend $worker}
457 if { $_tests_verbose > 0 } { _log_and_puts log "Executing tests in (up to) $parallel threads" }
458 # limit number of jobs in the queue by reasonable value
459 # to prevent slowdown due to unnecessary queue processing
461 set nbpooled_max [expr 10 * $parallel]
462 set nbpooled_ok [expr 5 * $parallel]
468 foreach test_def $tests_list {
469 # check for user break
470 if { $userbreak || "[info commands dbreak]" == "dbreak" && [catch dbreak] } {
475 set dir [lindex $test_def 0]
476 set group [lindex $test_def 1]
477 set grid [lindex $test_def 2]
478 set casename [lindex $test_def 3]
479 set casefile [lindex $test_def 4]
481 # command to set tests for generation of image in results directory
483 if { $logdir != "" } { set imgdir_cmd "set imagedir $logdir/$group/$grid" }
485 # prepare command file for running test case in separate instance of DRAW
486 set fd_cmd [open $logdir/$group/$grid/${casename}.tcl w]
487 puts $fd_cmd "$imgdir_cmd"
488 puts $fd_cmd "set test_image $casename"
489 puts $fd_cmd "_run_test $dir $group $grid $casefile t"
491 # use dlog command to obtain complete output of the test when it is absent (i.e. since OCCT 6.6.0)
492 # note: this is not needed if echo is set to 1 in call to _run_test above
493 if { ! [catch {dlog get}] } {
494 puts $fd_cmd "puts \[dlog get\]"
496 # else try to use old-style QA_ variables to get more output...
499 set env(QA_print_command) 1
502 # final 'exit' is needed when running on Linux under VirtualGl
506 # commant to run DRAW with a command file;
507 # note that empty string is passed as standard input to avoid possible
508 # hang-ups due to waiting for stdin of the launching process
509 set command "exec <<{} DRAWEXE -f $logdir/$group/$grid/${casename}.tcl"
511 # alternative method to run without temporary file; disabled as it needs too many backslashes
513 # set command "exec <<\"\" DRAWEXE -c $imgdir_cmd\\\; set test_image $casename\\\; \
514 # _run_test $dir $group $grid $casefile\\\; \
515 # puts \\\[dlog get\\\]\\\; exit"
518 # run test case, either in parallel or sequentially
519 if { $parallel > 0 } {
521 set job [tpool::post -nowait $worker "catch \"$command\" output; return \$output"]
522 set job_def($job) [list $logdir $dir $group $grid $casename]
524 if { $nbpooled > $nbpooled_max } {
525 _testgrid_process_jobs $worker $nbpooled_ok
528 # sequential execution
529 catch {eval $command} output
530 _log_test_case $output $logdir $dir $group $grid $casename log
532 # update summary log with requested period
533 if { $logdir != "" && $refresh > 0 && [expr [clock seconds] - $refresh_timer > $refresh] } {
534 # update and dump summary
535 _log_summarize $logdir $log
536 set refresh_timer [clock seconds]
541 # get results of started threads
542 if { $parallel > 0 } {
543 _testgrid_process_jobs $worker
544 # release thread pool
545 if { $nbpooled > 0 } {
546 tpool::cancel $worker [array names job_def]
548 catch {tpool::resume $worker}
549 tpool::release $worker
552 uplevel dchrono _timer stop
553 set time [lindex [split [uplevel dchrono _timer show] "\n"] 0]
556 _log_and_puts log "*********** Stopped by user break ***********"
557 set time "${time} \nNote: the process is not finished, stopped by user break!"
560 ######################################################
561 # output summary logs and exit
562 ######################################################
564 _log_summarize $logdir $log $time
565 if { $logdir != "" } {
566 puts "Detailed logs are saved in $logdir"
568 if { $logdir != "" && $xmlfile != "" } {
569 # XML output file is assumed relative to log dir unless it is absolute
570 if { [ file pathtype $xmlfile] == "relative" } {
571 set xmlfile [file normalize $logdir/$xmlfile]
573 _log_xml_summary $logdir $xmlfile $log 0
574 puts "XML summary is saved to $xmlfile"
576 # play sound signal at the end of test
583 # Procedure to regenerate summary log from logs of test cases
585 Regenerate summary log in the test directory from logs of test cases.
586 This can be necessary if test grids are executed separately (e.g. on
587 different stations) or some grids have been re-executed.
588 Use: testsummarize dir
590 proc testsummarize {dir} {
591 global _test_case_regexp
593 if { ! [file isdirectory $dir] } {
594 error "Error: \"$dir\" is not a directory"
597 # get summary statements from all test cases in one log
600 # to avoid huge listing of logs, first find all subdirectories and iterate
601 # by them, parsing log files in each subdirectory independently
602 foreach grid [glob -directory $dir -types d -tails */*] {
603 foreach caselog [glob -nocomplain -directory [file join $dir $grid] -types f -tails *.log] {
604 set file [file join $dir $grid $caselog]
606 set fd [open $file r]
607 while { [gets $fd line] >= 0 } {
608 if { [regexp $_test_case_regexp $line res grp grd cas status message] } {
609 if { "[file join $grid $caselog]" != "[file join $grp $grd ${cas}.log]" } {
610 puts "Error: $file contains status line for another test case ($line)"
618 if { $nbfound != 1 } {
619 puts "Error: $file contains $nbfound status lines, expected 1"
624 _log_summarize $dir $log "Summary regenerated from logs at [clock format [clock seconds]]"
628 # Procedure to compare results of two runs of test cases
630 Compare results of two executions of tests (CPU times, ...)
631 Use: testdiff dir1 dir2 [groupname [gridname]] [options...]
632 Where dir1 and dir2 are directories containing logs of two test runs.
634 -image [filename]: compare only images and save its in specified file (default
635 name is <dir1>/diffimage-<dir2>.log)
636 -cpu [filename]: compare only CPU and save it in specified file (default
637 name is <dir1>/diffcpu-<dir2>.log)
638 -memory [filename]: compare only memory and save it in specified file (default
639 name is <dir1>/diffmemory-<dir2>.log)
640 -save filename: save resulting log in specified file (default name is
641 <dir1>/diff-<dir2>.log); HTML log is saved with same name
643 -status {same|ok|all}: filter cases for comparing by their status:
644 same - only cases with same status are compared (default)
645 ok - only cases with OK status in both logs are compared
646 all - results are compared regardless of status
648 1 - output only differences
649 2 - output also list of logs and directories present in one of dirs only
650 3 - (default) output also progress messages
651 -highlight_percent value: highlight considerable (>value in %) deviations
652 of CPU and memory (default value is 5%)
654 proc testdiff {dir1 dir2 args} {
655 if { "$dir1" == "$dir2" } {
656 error "Input directories are the same"
659 ######################################################
661 ######################################################
664 set logfile [file join $dir1 "diff-[file tail $dir2].log"]
667 set logfile_memory ""
675 set highlight_percent 5
676 for {set narg 0} {$narg < [llength $args]} {incr narg} {
677 set arg [lindex $args $narg]
679 if { $arg == "-save" } {
681 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
682 set logfile [lindex $args $narg]
684 error "Error: Option -save must be followed by log file name"
691 if { $arg == "-image" } {
693 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
694 set logfile_image [lindex $args $narg]
696 set logfile_image [file join $dir1 "diffimage-[file tail $dir2].log"]
704 if { $arg == "-cpu" } {
706 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
707 set logfile_cpu [lindex $args $narg]
709 set logfile_cpu [file join $dir1 "diffcpu-[file tail $dir2].log"]
716 # memory compared log
717 if { $arg == "-memory" } {
719 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
720 set logfile_memory [lindex $args $narg]
722 set logfile_memory [file join $dir1 "diffmemory-[file tail $dir2].log"]
730 if { $arg == "-status" } {
732 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
733 set status [lindex $args $narg]
737 if { "$status" != "same" && "$status" != "all" && "$status" != "ok" } {
738 error "Error: Option -status must be followed by one of \"same\", \"all\", or \"ok\""
744 if { $arg == "-verbose" } {
746 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
747 set verbose [expr [lindex $args $narg]]
749 error "Error: Option -verbose must be followed by integer verbose level"
755 if { $arg == "-highlight_percent" } {
757 if { $narg < [llength $args] && ! [regexp {^-} [lindex $args $narg]] } {
758 set highlight_percent [expr [lindex $args $narg]]
760 error "Error: Option -highlight_percent must be followed by integer value"
765 if { [regexp {^-} $arg] } {
766 error "Error: unsupported option \"$arg\""
769 # non-option arguments form a subdirectory path
770 set basename [file join $basename $arg]
773 if {$image != false || $cpu != false || $memory != false} {
774 if {$save != false} {
775 error "Error: Option -save can not be used with image/cpu/memory options"
779 # run diff procedure (recursive)
780 _test_diff $dir1 $dir2 $basename $image $cpu $memory $status $verbose log log_image log_cpu log_memory
782 # save result to log file
783 if {$image == false && $cpu == false && $memory == false} {
784 if { "$logfile" != "" } {
785 _log_save $logfile [join $log "\n"]
786 _log_html_diff "[file rootname $logfile].html" $log $dir1 $dir2 ${highlight_percent}
787 puts "Log is saved to $logfile (and .html)"
790 foreach mode {image cpu memory} {
791 if {"[set logfile_${mode}]" != ""} {
792 _log_save "[set logfile_${mode}]" [join "[set log_${mode}]" "\n"]
793 _log_html_diff "[file rootname [set logfile_${mode}]].html" "[set log_${mode}]" $dir1 $dir2 ${highlight_percent}
794 puts "Log (${mode}) is saved to [set logfile_${mode}] (and .html)"
801 # Procedure to check data file before adding it to repository
803 Checks specified data files for putting them into the test data files repository.
805 Use: testfile filelist
808 - data file (non-binary) is in DOS encoding (CR/LF)
809 - same data file (with same or another name) already exists in the repository
810 - another file with the same name already exists
811 Note that names are considered to be case-insensitive (for compatibility
814 Unless the file is already in the repository, tries to load it, reports
815 the recognized file format, file size, number of faces and edges in the
816 loaded shape (if any), information contained its triangulation, and makes
817 snapshot (in the temporary directory).
819 Finally it advises whether the file should be put to public section of the
824 If "-check" is given as an argument, then procedure will check files already
825 located in the repository (for possible duplicates and for DOS encoding).
827 proc testfile {filelist} {
830 # check that CSF_TestDataPath is defined
831 if { ! [info exists env(CSF_TestDataPath)] } {
832 error "Environment variable CSF_TestDataPath must be defined!"
836 if { "$filelist" == "-check" } { set checkrepo t }
838 # build registry of existing data files (name -> path) and (size -> path)
839 puts "Collecting info on test data files repository..."
840 foreach dir [_split_path $env(CSF_TestDataPath)] {
841 while {[llength $dir] != 0} {
842 set curr [lindex $dir 0]
843 set dir [lrange $dir 1 end]
844 eval lappend dir [glob -nocomplain -directory $curr -type d *]
845 foreach file [glob -nocomplain -directory $curr -type f *] {
846 set name [file tail $file]
847 set name_lower [string tolower $name]
848 set size [file size $file]
850 # check that the file is not in DOS encoding
852 if { [_check_dos_encoding $file] } {
853 puts "Warning: file $file is in DOS encoding; was this intended?"
855 _check_file_format $file
857 # check if file with the same name is present twice or more
858 if { [info exists names($name_lower)] } {
859 puts "Error: more than one file with name $name is present in the repository:"
860 if { [_diff_files $file $names($name_lower)] } {
861 puts "(files are different by content)"
863 puts "(files are same by content)"
866 puts "--> $names($name_lower)"
870 # check if file with the same content exists
871 if { [info exists sizes($size)] } {
872 foreach other $sizes($size) {
873 if { ! [_diff_files $file $other] } {
874 puts "Warning: two files with the same content found:"
882 # add the file to the registry
883 lappend names($name_lower) $file
884 lappend sizes($size) $file
888 if { $checkrepo || [llength $filelist] <= 0 } { return }
890 # check the new files
892 puts "Checking new file(s)..."
893 foreach file $filelist {
894 set name [file tail $file]
895 set name_lower [string tolower $name]
898 # check for presence of the file with same name
899 if { [info exists names($name_lower)] } {
901 foreach other $names($name_lower) {
902 # avoid comparing the file with itself
903 if { [file normalize $file] == [file normalize $other] } {
907 if { [_diff_files $file $other] } {
908 puts "\n* $file: error\n name is already used by existing file\n --> $other"
910 puts "\n* $file: already present \n --> $other"
915 if { $found } { continue }
918 # get size of the file; if it is in DOS encoding and less than 1 MB,
919 # estimate also its size in UNIX encoding to be able to find same
920 # file if already present but in UNIX encoding
921 set sizeact [file size $file]
923 set isdos [_check_dos_encoding $file]
924 if { $isdos && $sizeact < 10000000 } {
925 set fd [open $file r]
926 fconfigure $fd -translation crlf
927 set sizeunx [string length [read $fd]]
931 # check if file with the same content exists
932 foreach size "$sizeact $sizeunx" {
933 if { [info exists sizes($size)] } {
934 foreach other $sizes($size) {
935 # avoid comparing the file with itself
936 if { [file normalize $file] == [file normalize $other] } {
940 if { ! [_diff_files $file $other] } {
941 puts "\n* $file: duplicate \n already present under name [file tail $other]\n --> $other"
946 if { $found } { break }
949 if { $found } { continue }
951 # file is not present yet, so to be analyzed
952 puts "\n* $file: new file"
954 # add the file to the registry as if it were added to the repository,
955 # to report possible duplicates among the currently processed files
956 lappend names($name_lower) $file
957 if { "$sizeunx" != "" } {
958 lappend sizes($sizeunx) $file
960 lappend sizes($sizeact) $file
963 # first of all, complain if it is in DOS encoding
965 puts " Warning: DOS encoding detected, consider converting to"
966 puts " UNIX unless DOS line ends are needed for the test"
969 # try to read the file
970 set format [_check_file_format $file]
971 if { [catch {uplevel load_data_file $file $format a}] } {
972 puts " Warning: Cannot read as $format file"
976 # warn if shape contains triangulation
978 if { "$format" != "STL" &&
979 [regexp {contains\s+([0-9]+)\s+triangles} [uplevel trinfo a] res nbtriangles] &&
980 $nbtriangles != 0 } {
981 puts " Warning: shape contains triangulation ($nbtriangles triangles),"
982 puts " consider removing them unless they are needed for the test!"
985 # get number of faces and edges
988 set nbs [uplevel nbshapes a]
989 regexp {EDGE[ \t:]*([0-9]+)} $nbs res edges
990 regexp {FACE[ \t:]*([0-9]+)} $nbs res faces
992 # classify; first check file size and number of faces and edges
993 if { $size < 95000 && $faces < 20 && $edges < 100 } {
1000 puts " $format size=[expr $size / 1024] KiB, nbfaces=$faces, nbedges=$edges -> $dir"
1002 set tmpdir [_get_temp_dir]
1003 file mkdir $tmpdir/$dir
1008 uplevel vsetdispmode 1
1011 uplevel vdump $tmpdir/$dir/[file rootname [file tail $file]].png
1014 if { $has_images } {
1015 puts "Snapshots are saved in subdirectory [_get_temp_dir]"
1019 # Procedure to locate data file for test given its name.
1020 # The search is performed assuming that the function is called
1021 # from the test case script; the search order is:
1022 # - subdirectory "data" of the test script (grid) folder
1023 # - subdirectories in environment variable CSF_TestDataPath
1024 # - subdirectory set by datadir command
1025 # If file is not found, raises Tcl error.
1026 proc locate_data_file {filename} {
1027 global env groupname gridname casename
1029 # check if the file is located in the subdirectory data of the script dir
1030 set scriptfile [info script]
1031 if { $scriptfile != "" } {
1032 set path [file join [file dirname $scriptfile] data $filename]
1033 if { [file exists $path] } {
1034 return [file normalize $path]
1038 # check sub-directories in paths indicated by CSF_TestDataPath
1039 if { [info exists env(CSF_TestDataPath)] } {
1040 foreach dir [_split_path $env(CSF_TestDataPath)] {
1041 while {[llength $dir] != 0} {
1042 set name [lindex $dir 0]
1043 set dir [lrange $dir 1 end]
1044 # skip directories starting with dot
1045 if { [regexp {^[.]} $name] } { continue }
1046 if { [file exists $name/$filename] } {
1047 return [file normalize $name/$filename]
1049 eval lappend dir [glob -nocomplain -directory $name -type d *]
1054 # check current datadir
1055 if { [file exists [uplevel datadir]/$filename] } {
1056 return [file normalize [uplevel datadir]/$filename]
1060 error [join [list "File $filename could not be found" \
1061 "(should be in paths indicated by CSF_TestDataPath environment variable, " \
1062 "or in subfolder data in the script directory)"] "\n"]
1065 # Internal procedure to find test case indicated by group, grid, and test case names;
1067 # - dir: path to the base directory of the tests group
1068 # - gridname: actual name of the grid
1069 # - casefile: path to the test case script
1070 # if no such test is found, raises error with appropriate message
1071 proc _get_test {group grid casename _dir _gridname _casefile} {
1073 upvar $_gridname gridname
1074 upvar $_casefile casefile
1078 # check that environment variable defining paths to test scripts is defined
1079 if { ! [info exists env(CSF_TestScriptsPath)] ||
1080 [llength $env(CSF_TestScriptsPath)] <= 0 } {
1081 error "Error: Environment variable CSF_TestScriptsPath is not defined"
1084 # iterate by all script paths
1085 foreach dir [_split_path $env(CSF_TestScriptsPath)] {
1086 # protection against empty paths
1087 set dir [string trim $dir]
1088 if { $dir == "" } { continue }
1090 # check that directory exists
1091 if { ! [file isdirectory $dir] } {
1092 puts "Warning: directory $dir listed in CSF_TestScriptsPath does not exist, skipped"
1096 # check if test group with given name exists in this dir
1097 # if not, continue to the next test dir
1098 if { ! [file isdirectory $dir/$group] } { continue }
1100 # check that grid with given name (possibly alias) exists; stop otherwise
1102 if { ! [file isdirectory $dir/$group/$gridname] } {
1103 # check if grid is named by alias rather than by actual name
1104 if { [file exists $dir/$group/grids.list] } {
1105 set fd [open $dir/$group/grids.list]
1106 while { [gets $fd line] >= 0 } {
1107 if { [regexp "\[ \t\]*\#.*" $line] } { continue }
1108 if { [regexp "^$grid\[ \t\]*\(\[A-Za-z0-9_.-\]+\)\$" $line res gridname] } {
1115 if { ! [file isdirectory $dir/$group/$gridname] } { continue }
1117 # get actual file name of the script; stop if it cannot be found
1118 set casefile $dir/$group/$gridname/$casename
1119 if { ! [file exists $casefile] } {
1120 # check if this grid is aliased to another one
1121 if { [file exists $dir/$group/$gridname/cases.list] } {
1122 set fd [open $dir/$group/$gridname/cases.list]
1123 if { [gets $fd line] >= 0 } {
1124 set casefile [file normalize $dir/$group/$gridname/[string trim $line]/$casename]
1129 if { [file exists $casefile] } {
1135 # coming here means specified test is not found; report error
1136 error [join [list "Error: test case $group / $grid / $casename is not found in paths listed in variable" \
1137 "CSF_TestScriptsPath (current value is \"$env(CSF_TestScriptsPath)\")"] "\n"]
1140 # Internal procedure to run test case indicated by base directory,
1141 # grid and grid names, and test case file path.
1142 # The log can be obtained by command "dlog get".
1143 proc _run_test {scriptsdir group gridname casefile echo} {
1147 uplevel dchrono _timer reset
1148 uplevel dchrono _timer start
1149 catch {uplevel meminfo h} membase
1151 # enable commands logging; switch to old-style mode if dlog command is not present
1153 if { [catch {dlog reset}] } {
1155 } elseif { $echo } {
1160 rename puts puts-saved
1162 global _tests_verbose
1164 # log only output to stdout and stderr, not to file!
1165 if {[llength $args] > 1} {
1166 set optarg [lindex $args end-1]
1167 if { $optarg == "stdout" || $optarg == "stderr" || $optarg == "-newline" } {
1168 dlog add [lindex $args end]
1170 eval puts-saved $args
1173 dlog add [lindex $args end]
1178 # evaluate test case
1181 # set variables identifying test case
1182 uplevel set casename [file tail $casefile]
1183 uplevel set groupname $group
1184 uplevel set gridname $gridname
1185 uplevel set dirname $scriptsdir
1187 # set path for saving of log and images (if not yet set) to temp dir
1188 if { ! [uplevel info exists imagedir] } {
1189 uplevel set test_image \$casename
1191 # create subdirectory in temp named after group and grid with timestamp
1192 set rootlogdir [_get_temp_dir]
1194 set imagedir "${group}-${gridname}-${::casename}-[clock format [clock seconds] -format {%Y-%m-%dT%Hh%Mm%Ss}]"
1195 set imagedir [file normalize ${rootlogdir}/$imagedir]
1197 if { [catch {file mkdir $imagedir}] || ! [file writable $imagedir] ||
1198 ! [catch {glob -directory $imagedir *}] } {
1199 # puts "Warning: Cannot create directory \"$imagedir\", or it is not empty; \"${rootlogdir}\" is used"
1200 set imagedir $rootlogdir
1203 uplevel set imagedir \"$imagedir\"
1207 # execute test scripts
1208 if { [file exists $scriptsdir/$group/begin] } {
1209 puts "Executing $scriptsdir/$group/begin..."; flush stdout
1210 uplevel source $scriptsdir/$group/begin
1212 if { [file exists $scriptsdir/$group/$gridname/begin] } {
1213 puts "Executing $scriptsdir/$group/$gridname/begin..."; flush stdout
1214 uplevel source $scriptsdir/$group/$gridname/begin
1217 puts "Executing $casefile..."; flush stdout
1218 uplevel source $casefile
1220 if { [file exists $scriptsdir/$group/$gridname/end] } {
1221 puts "Executing $scriptsdir/$group/$gridname/end..."; flush stdout
1222 uplevel source $scriptsdir/$group/$gridname/end
1224 if { [file exists $scriptsdir/$group/end] } {
1225 puts "Executing $scriptsdir/$group/end..."; flush stdout
1226 uplevel source $scriptsdir/$group/end
1229 puts "Tcl Exception: $res"
1233 if { $dlog_exists } {
1238 rename puts-saved puts
1243 # stop cpulimit killer if armed by the test
1246 # add memory and timing info
1248 if { ! [catch {uplevel meminfo h} memuse] } {
1249 append stats "MEMORY DELTA: [expr ($memuse - $membase) / 1024] KiB\n"
1251 uplevel dchrono _timer stop
1252 set time [uplevel dchrono _timer show]
1253 if { [regexp -nocase {CPU user time:[ \t]*([0-9.e-]+)} $time res cpu_usr] } {
1254 append stats "TOTAL CPU TIME: $cpu_usr sec\n"
1256 if { $dlog_exists && ! $echo } {
1263 uplevel unset casename groupname gridname dirname
1264 if { $tmp_imagedir } { uplevel unset imagedir test_image }
1267 # Internal procedure to check log of test execution and decide if it passed or failed
1268 proc _check_log {dir group gridname casename errors log {_summary {}} {_html_log {}}} {
1270 if { $_summary != "" } { upvar $_summary summary }
1271 if { $_html_log != "" } { upvar $_html_log html_log }
1278 # load definition of 'bad words' indicating test failure
1279 # note that rules are loaded in the order of decreasing priority (grid - group - common),
1280 # thus grid rules will override group ones
1282 foreach rulesfile [list $dir/$group/$gridname/parse.rules $dir/$group/parse.rules $dir/parse.rules] {
1283 if [catch {set fd [open $rulesfile r]}] { continue }
1284 while { [gets $fd line] >= 0 } {
1285 # skip comments and empty lines
1286 if { [regexp "\[ \t\]*\#.*" $line] } { continue }
1287 if { [string trim $line] == "" } { continue }
1289 if { ! [regexp {^([^/]*)/([^/]*)/(.*)$} $line res status rexp comment] } {
1290 puts "Warning: cannot recognize parsing rule \"$line\" in file $rulesfile"
1293 set status [string trim $status]
1294 if { $comment != "" } { append status " ([string trim $comment])" }
1295 set rexp [regsub -all {\\b} $rexp {\\y}] ;# convert regexp from Perl to Tcl style
1296 lappend badwords [list $status $rexp]
1300 if { [llength $badwords] <= 0 } {
1301 puts "Warning: no definition of error indicators found (check files parse.rules)"
1304 # analyse log line-by-line
1305 set todos {} ;# TODO statements
1306 set requs {} ;# REQUIRED statements
1307 set todo_incomplete -1
1309 foreach line [split $log "\n"] {
1310 # check if line defines specific treatment of some messages
1311 if [regexp -nocase {^[ \s]*TODO ([^:]*):(.*)$} $line res platforms pattern] {
1312 if { ! [regexp -nocase {\mAll\M} $platforms] &&
1313 ! [regexp -nocase "\\m[checkplatform]\\M" $platforms] } {
1314 lappend html_log [_html_highlight IGNORE $line]
1315 continue ;# TODO statement is for another platform
1318 # record TODOs that mark unstable cases
1319 if { [regexp {[\?]} $platforms] } {
1320 set todos_unstable([llength $todos]) 1
1323 # convert legacy regexps from Perl to Tcl style
1324 set pattern [regsub -all {\\b} [string trim $pattern] {\\y}]
1326 # special case: TODO TEST INCOMPLETE
1327 if { [string trim $pattern] == "TEST INCOMPLETE" } {
1328 set todo_incomplete [llength $todos]
1331 lappend todos [list $pattern [llength $html_log] $line]
1332 lappend html_log [_html_highlight BAD $line]
1335 if [regexp -nocase {^[ \s]*REQUIRED ([^:]*):[ \s]*(.*)$} $line res platforms pattern] {
1336 if { ! [regexp -nocase {\mAll\M} $platforms] &&
1337 ! [regexp -nocase "\\m[checkplatform]\\M" $platforms] } {
1338 lappend html_log [_html_highlight IGNORE $line]
1339 continue ;# REQUIRED statement is for another platform
1341 lappend requs [list $pattern [llength $html_log] $line]
1342 lappend html_log [_html_highlight OK $line]
1346 # check for presence of required messages
1348 for {set i 0} {$i < [llength $requs]} {incr i} {
1349 set pattern [lindex $requs $i 0]
1350 if { [regexp $pattern $line] } {
1351 incr required_count($i)
1352 lappend html_log [_html_highlight OK $line]
1361 # check for presence of messages indicating test result
1362 foreach bw $badwords {
1363 if { [regexp [lindex $bw 1] $line] } {
1364 # check if this is known bad case
1366 for {set i 0} {$i < [llength $todos]} {incr i} {
1367 set pattern [lindex $todos $i 0]
1368 if { [regexp $pattern $line] } {
1371 lappend html_log [_html_highlight BAD $line]
1376 # if it is not in todo, define status
1377 if { ! $is_known } {
1378 set stat [lindex $bw 0 0]
1380 lappend errors_log $line
1382 lappend html_log [_html_highlight $stat $line]
1383 if { $status == "" && $stat != "OK" && ! [regexp -nocase {^IGNOR} $stat] } {
1384 set status [lindex $bw 0]
1391 if { ! $ismarked } {
1392 lappend html_log $line
1396 # check for presence of TEST COMPLETED statement
1397 if { $status == "" && ! [regexp {TEST COMPLETED} $log] } {
1398 # check whether absence of TEST COMPLETED is known problem
1399 if { $todo_incomplete >= 0 } {
1400 incr todo_count($todo_incomplete)
1402 set status "FAILED (no final message is found)"
1406 # report test as failed if it doesn't contain required pattern
1407 if { $status == "" } {
1408 for {set i 0} {$i < [llength $requs]} {incr i} {
1409 if { ! [info exists required_count($i)] } {
1410 set linenum [lindex $requs $i 1]
1411 set html_log [lreplace $html_log $linenum $linenum [_html_highlight FAILED [lindex $requs $i 2]]]
1412 set status "FAILED (REQUIRED statement no. [expr $i + 1] is not found)"
1417 # check declared bad cases and diagnose possible improvement
1418 # (bad case declared but not detected).
1419 # Note that absence of the problem marked by TODO with question mark
1420 # (unstable) is not reported as improvement.
1421 if { $status == "" } {
1422 for {set i 0} {$i < [llength $todos]} {incr i} {
1423 if { ! [info exists todos_unstable($i)] &&
1424 (! [info exists todo_count($i)] || $todo_count($i) <= 0) } {
1425 set linenum [lindex $todos $i 1]
1426 set html_log [lreplace $html_log $linenum $linenum [_html_highlight IMPROVEMENT [lindex $todos $i 2]]]
1427 set status "IMPROVEMENT (expected problem TODO no. [expr $i + 1] is not detected)"
1433 # report test as known bad if at least one of expected problems is found
1434 if { $status == "" && [llength [array names todo_count]] > 0 } {
1435 set status "BAD (known problem)"
1439 if { $status == "" } {set status "OK" }
1442 set status "FAILED ($res)"
1446 _log_and_puts summary "CASE $group $gridname $casename: $status"
1447 set summary [join $summary "\n"]
1449 foreach error $errors_log {
1450 _log_and_puts summary " $error"
1453 set html_log "[_html_highlight [lindex $status 0] $summary]\n[join $html_log \n]"
1456 # Auxiliary procedure putting message to both cout and log variable (list)
1457 proc _log_and_puts {logvar message} {
1458 if { $logvar != "" } {
1460 lappend log $message
1465 # Auxiliary procedure to log result on single test case
1466 proc _log_test_case {output logdir dir group grid casename logvar} {
1469 # check result and make HTML log
1470 _check_log $dir $group $grid $casename $show_errors $output summary html_log
1471 lappend log $summary
1474 if { $logdir != "" } {
1475 _log_html $logdir/$group/$grid/$casename.html $html_log "Test $group $grid $casename"
1476 _log_save $logdir/$group/$grid/$casename.log "$output\n$summary" "Test $group $grid $casename"
1480 # Auxiliary procedure to save log to file
1481 proc _log_save {file log {title {}}} {
1482 # create missing directories as needed
1483 catch {file mkdir [file dirname $file]}
1485 # try to open a file
1486 if [catch {set fd [open $file w]} res] {
1487 error "Error saving log file $file: $res"
1490 # dump log and close
1497 # Auxiliary procedure to make a (relative if possible) URL to a file for
1498 # inclusion a reference in HTML log
1499 proc _make_url {htmldir file} {
1500 set htmlpath [file split [file normalize $htmldir]]
1501 set filepath [file split [file normalize $file]]
1502 for {set i 0} {$i < [llength $htmlpath]} {incr i} {
1503 if { "[lindex $htmlpath $i]" != "[lindex $filepath $i]" } {
1504 if { $i == 0 } { break }
1505 return "[string repeat "../" [expr [llength $htmlpath] - $i - 1]][eval file join [lrange $filepath $i end]]"
1509 # if relative path could not be made, return full file URL
1510 return "file://[file normalize $file]"
1513 # Auxiliary procedure to save log to file
1514 proc _log_html {file log {title {}}} {
1515 # create missing directories as needed
1516 catch {file mkdir [file dirname $file]}
1518 # try to open a file
1519 if [catch {set fd [open $file w]} res] {
1520 error "Error saving log file $file: $res"
1524 puts $fd "<html><head><title>$title</title></head><body><h1>$title</h1>"
1526 # add images if present; these should have either PNG, GIF, or JPG extension,
1527 # and start with name of the test script, with optional suffix separated
1528 # by underscore or dash
1529 set imgbasename [file rootname [file tail $file]]
1530 foreach img [lsort [glob -nocomplain -directory [file dirname $file] -tails \
1531 ${imgbasename}.gif ${imgbasename}.png ${imgbasename}.jpg \
1532 ${imgbasename}_*.gif ${imgbasename}_*.png ${imgbasename}_*.jpg \
1533 ${imgbasename}-*.gif ${imgbasename}-*.png ${imgbasename}-*.jpg]] {
1534 puts $fd "<p>[file tail $img]<br><img src=\"$img\"/><p>"
1537 # print log body, trying to add HTML links to script files on lines like
1538 # "Executing <filename>..."
1540 foreach line [split $log "\n"] {
1541 if { [regexp {Executing[ \t]+([a-zA-Z0-9._/:-]+[^.])} $line res script] &&
1542 [file exists $script] } {
1543 set line [regsub $script $line "<a href=\"[_make_url $file $script]\">$script</a>"]
1547 puts $fd "</pre></body></html>"
1553 # Auxiliary method to make text with HTML highlighting according to status
1554 proc _html_color {status} {
1555 # choose a color for the cell according to result
1556 if { $status == "OK" } {
1558 } elseif { [regexp -nocase {^FAIL} $status] } {
1560 } elseif { [regexp -nocase {^BAD} $status] } {
1562 } elseif { [regexp -nocase {^IMP} $status] } {
1564 } elseif { [regexp -nocase {^SKIP} $status] } {
1566 } elseif { [regexp -nocase {^IGNOR} $status] } {
1569 puts "Warning: no color defined for status $status, using red as if FAILED"
1574 # Format text line in HTML to be colored according to the status
1575 proc _html_highlight {status line} {
1576 return "<table><tr><td bgcolor=\"[_html_color $status]\">$line</td></tr></table>"
1579 # Internal procedure to generate HTML page presenting log of the tests
1580 # execution in tabular form, with links to reports on individual cases
1581 proc _log_html_summary {logdir log totals regressions improvements skipped total_time} {
1582 global _test_case_regexp
1584 # create missing directories as needed
1587 # try to open a file and start HTML
1588 if [catch {set fd [open $logdir/summary.html w]} res] {
1589 error "Error creating log file: $res"
1592 # write HRML header, including command to refresh log if still in progress
1593 puts $fd "<html><head>"
1594 puts $fd "<title>Tests summary</title>"
1595 if { $total_time == "" } {
1596 puts $fd "<meta http-equiv=\"refresh\" content=\"10\">"
1598 puts $fd "<meta http-equiv=\"pragma\" content=\"NO-CACHE\">"
1599 puts $fd "</head><body>"
1602 set legend(OK) "Test passed OK"
1603 set legend(FAILED) "Test failed (regression)"
1604 set legend(BAD) "Known problem"
1605 set legend(IMPROVEMENT) "Possible improvement (expected problem not detected)"
1606 set legend(SKIPPED) "Test skipped due to lack of data file"
1607 puts $fd "<h1>Summary</h1><table>"
1608 foreach nbstat $totals {
1609 set status [lindex $nbstat 1]
1610 if { [info exists legend($status)] } {
1611 set comment $legend($status)
1613 set comment "User-defined status"
1615 puts $fd "<tr><td align=\"right\">[lindex $nbstat 0]</td><td bgcolor=\"[_html_color $status]\">$status</td><td>$comment</td></tr>"
1619 # time stamp and elapsed time info
1620 if { $total_time != "" } {
1621 puts $fd "<p>Generated on [clock format [clock seconds] -format {%Y-%m-%d %H:%M:%S}] on [info hostname]\n<p>"
1622 puts $fd [join [split $total_time "\n"] "<p>"]
1624 puts $fd "<p>NOTE: This is intermediate summary; the tests are still running! This page will refresh automatically until tests are finished."
1627 # print regressions and improvements
1628 foreach featured [list $regressions $improvements $skipped] {
1629 if { [llength $featured] <= 1 } { continue }
1630 set status [string trim [lindex $featured 0] { :}]
1631 puts $fd "<h2>$status</h2>"
1634 foreach test [lrange $featured 1 end] {
1635 if { ! [regexp {^(.*)\s+([\w.]+)$} $test res gg name] } {
1637 set name "Error building short list; check details"
1639 if { $gg != $groupgrid } {
1640 if { $groupgrid != "" } { puts $fd "</tr>" }
1642 puts $fd "<tr><td>$gg</td>"
1644 puts $fd "<td bgcolor=\"[_html_color $status]\"><a href=\"[regsub -all { } $gg /]/${name}.html\">$name</a></td>"
1646 if { $groupgrid != "" } { puts $fd "</tr>" }
1650 # put detailed log with TOC
1651 puts $fd "<hr><h1>Details</h1>"
1652 puts $fd "<div style=\"float:right; padding: 10px; border-style: solid; border-color: blue; border-width: 2px;\">"
1654 # process log line-by-line
1658 foreach line [lsort -dictionary $log] {
1659 # check that the line is case report in the form "CASE group grid name: result (explanation)"
1660 if { ! [regexp $_test_case_regexp $line res grp grd casename result message] } {
1665 if { $grp != $group } {
1666 if { $letter != "" } { lappend body "</tr></table>" }
1670 puts $fd "<a href=\"#$group\">$group</a><br>"
1671 lappend body "<h2><a name=\"$group\">Group $group</a></h2>"
1675 if { $grd != $grid } {
1676 if { $letter != "" } { lappend body "</tr></table>" }
1679 puts $fd " <a href=\"#$group-$grid\">$grid</a><br>"
1680 lappend body "<h2><a name=\"$group-$grid\">Grid $group $grid</a></h2>"
1683 # check if test case name is <letter><digit>;
1684 # if not, set alnum to period "." to recognize non-standard test name
1685 if { ! [regexp {\A([A-Za-z]{1,2})([0-9]{1,2})\Z} $casename res alnum number] &&
1686 ! [regexp {\A([A-Za-z0-9]+)_([0-9]+)\Z} $casename res alnum number] } {
1690 # start new row when letter changes or for non-standard names
1691 if { $alnum != $letter || $alnum == "." } {
1692 if { $letter != "" } {
1693 lappend body "</tr><tr>"
1695 lappend body "<table><tr>"
1700 lappend body "<td bgcolor=\"[_html_color $result]\"><a href=\"$group/$grid/${casename}.html\">$casename</a></td>"
1702 puts $fd "</div>\n[join $body "\n"]</tr></table>"
1704 # add remaining lines of log as plain text
1705 puts $fd "<h2>Plain text messages</h2>\n<pre>"
1707 if { ! [regexp $_test_case_regexp $line] } {
1713 # close file and exit
1719 # Procedure to dump summary logs of tests
1720 proc _log_summarize {logdir log {total_time {}}} {
1722 # sort log records alphabetically to have the same behavior on Linux and Windows
1723 # (also needed if tests are run in parallel)
1724 set loglist [lsort -dictionary $log]
1726 # classify test cases by status
1727 foreach line $loglist {
1728 if { [regexp {^CASE ([^:]*): ([[:alnum:]]+).*$} $line res caseid status] } {
1729 lappend stat($status) $caseid
1733 set improvements {Improvements:}
1734 set regressions {Failed:}
1735 set skipped {Skipped:}
1736 if { [info exists stat] } {
1737 foreach status [lsort [array names stat]] {
1738 lappend totals [list [llength $stat($status)] $status]
1740 # separately count improvements (status starting with IMP), skipped (status starting with SKIP) and regressions (all except IMP, OK, BAD, and SKIP)
1741 if { [regexp -nocase {^IMP} $status] } {
1742 eval lappend improvements $stat($status)
1743 } elseif { [regexp -nocase {^SKIP} $status] } {
1744 eval lappend skipped $stat($status)
1745 } elseif { $status != "OK" && ! [regexp -nocase {^BAD} $status] && ! [regexp -nocase {^SKIP} $status] } {
1746 eval lappend regressions $stat($status)
1751 # if time is specified, add totals
1752 if { $total_time != "" } {
1753 if { [llength $improvements] > 1 } {
1754 _log_and_puts log [join $improvements "\n "]
1756 if { [llength $regressions] > 1 } {
1757 _log_and_puts log [join $regressions "\n "]
1759 if { [llength $skipped] > 1 } {
1760 _log_and_puts log [join $skipped "\n "]
1762 if { [llength $improvements] == 1 && [llength $regressions] == 1 } {
1763 _log_and_puts log "No regressions"
1765 _log_and_puts log "Total cases: [join $totals {, }]"
1766 _log_and_puts log $total_time
1770 if { $logdir != "" } {
1771 _log_html_summary $logdir $log $totals $regressions $improvements $skipped $total_time
1772 _log_save $logdir/tests.log [join $log "\n"] "Tests summary"
1778 # Internal procedure to generate XML log in JUnit style, for further
1779 # consumption by Jenkins or similar systems.
1781 # The output is intended to conform to XML schema supported by Jenkins found at
1782 # https://svn.jenkins-ci.org/trunk/hudson/dtkit/dtkit-format/dtkit-junit-model/src/main/resources/com/thalesgroup/dtkit/junit/model/xsd/junit-4.xsd
1784 # The mapping of the fields is inspired by annotated schema of Apache Ant JUnit XML format found at
1785 # http://windyroad.org/dl/Open%20Source/JUnit.xsd
1786 proc _log_xml_summary {logdir filename log include_cout} {
1787 global _test_case_regexp
1789 catch {file mkdir [file dirname $filename]}
1791 # try to open a file and start XML
1792 if [catch {set fd [open $filename w]} res] {
1793 error "Error creating XML summary file $filename: $res"
1795 puts $fd "<?xml version='1.0' encoding='utf-8'?>"
1796 puts $fd "<testsuites>"
1798 # prototype for command to generate test suite tag
1799 set time_and_host "timestamp=\"[clock format [clock seconds] -format {%Y-%m-%dT%H:%M:%S}]\" hostname=\"[info hostname]\""
1800 set cmd_testsuite {puts $fd "<testsuite name=\"$group $grid\" tests=\"$nbtests\" failures=\"$nbfail\" errors=\"$nberr\" time=\"$time\" skipped=\"$nbskip\" $time_and_host>\n$testcases\n</testsuite>\n"}
1802 # sort log and process it line-by-line
1804 foreach line [lsort -dictionary $log] {
1805 # check that the line is case report in the form "CASE group grid name: result (explanation)"
1806 if { ! [regexp $_test_case_regexp $line res grp grd casename result message] } {
1809 set message [string trim $message " \t\r\n()"]
1811 # start new testsuite for each grid
1812 if { $grp != $group || $grd != $grid } {
1814 # write previous test suite
1815 if [info exists testcases] { eval $cmd_testsuite }
1830 # parse test log and get its CPU time
1833 if { [catch {set fdlog [open $logdir/$group/$grid/${casename}.log r]} ret] } {
1834 puts "Error: cannot open $logdir/$group/$grid/${casename}.log: $ret"
1836 while { [gets $fdlog logline] >= 0 } {
1837 if { $include_cout } {
1838 append testout "$logline\n"
1840 if [regexp -nocase {TOTAL CPU TIME:\s*([\d.]+)\s*sec} $logline res cpu] {
1841 set add_cpu " time=\"$cpu\""
1842 set time [expr $time + $cpu]
1847 if { ! $include_cout } {
1848 set testout "$line\n"
1851 # record test case with its output and status
1852 # Mapping is: SKIPPED, BAD, and OK to OK, all other to failure
1853 append testcases "\n <testcase name=\"$casename\"$add_cpu status=\"$result\">\n"
1854 append testcases "\n <system-out>\n$testout </system-out>"
1855 if { $result != "OK" } {
1856 if { [regexp -nocase {^SKIP} $result] } {
1858 append testcases "\n <error name=\"$result\" message=\"$message\"/>"
1859 } elseif { [regexp -nocase {^BAD} $result] } {
1861 append testcases "\n <skipped>$message</skipped>"
1864 append testcases "\n <failure name=\"$result\" message=\"$message\"/>"
1867 append testcases "\n </testcase>"
1870 # write last test suite
1871 if [info exists testcases] { eval $cmd_testsuite }
1874 puts $fd "</testsuites>"
1879 # Auxiliary procedure to split path specification (usually defined by
1880 # environment variable) into list of directories or files
1881 proc _split_path {pathspec} {
1884 # first replace all \ (which might occur on Windows) by /
1885 regsub -all "\\\\" $pathspec "/" pathspec
1887 # split path by platform-specific separator
1888 return [split $pathspec [_path_separator]]
1891 # Auxiliary procedure to define platform-specific separator for directories in
1892 # path specification
1893 proc _path_separator {} {
1896 # split path by platform-specific separator
1897 if { $tcl_platform(platform) == "windows" } {
1904 # Procedure to make a diff and common of two lists
1905 proc _list_diff {list1 list2 _in1 _in2 _common} {
1908 upvar $_common common
1913 foreach item $list1 {
1914 if { [lsearch -exact $list2 $item] >= 0 } {
1915 lappend common $item
1920 foreach item $list2 {
1921 if { [lsearch -exact $common $item] < 0 } {
1928 # procedure to load a file to Tcl string
1929 proc _read_file {filename} {
1930 set fd [open $filename r]
1931 set result [read -nonewline $fd]
1936 # procedure to construct name for the mage diff file
1937 proc _diff_img_name {dir1 dir2 casepath imgfile} {
1938 return [file join $dir1 $casepath "diff-[file tail $dir2]-$imgfile"]
1941 # auxiliary procedure to produce string comparing two values
1942 proc _diff_show_ratio {value1 value2} {
1943 if {[expr double ($value2)] == 0.} {
1944 return "$value1 / $value2"
1946 return "$value1 / $value2 \[[format "%+5.2f%%" [expr 100 * ($value1 - $value2) / double($value2)]]\]"
1950 # procedure to check cpu user time
1951 proc _check_time {regexp_msg} {
1955 upvar log_cpu log_cpu
1957 upvar basename basename
1958 upvar casename casename
1959 set time1_list [dict create]
1960 set time2_list [dict create]
1961 set cpu_find UNDEFINED
1963 foreach line1 [split $log1 "\n"] {
1964 if { [regexp "${regexp_msg}" $line1 dump chronometer_name cpu_find] } {
1965 dict set time1_list "${chronometer_name}" "${cpu_find}"
1969 foreach line2 [split $log2 "\n"] {
1970 if { [regexp "${regexp_msg}" $line2 dump chronometer_name cpu_find] } {
1971 dict set time2_list "${chronometer_name}" "${cpu_find}"
1975 if { [llength [dict keys $time1_list]] != [llength [dict keys $time2_list]] } {
1976 puts "Error: number of dchrono/chrono COUNTER are different in the same test cases"
1978 foreach key [dict keys $time1_list] {
1979 set time1 [dict get $time1_list $key]
1980 set time2 [dict get $time2_list $key]
1982 # compare CPU user time with 10% precision (but not less 0.5 sec)
1983 if { [expr abs ($time1 - $time2) > 0.5 + 0.05 * abs ($time1 + $time2)] } {
1984 if {$cpu != false} {
1985 _log_and_puts log_cpu "COUNTER $key: [split $basename /] $casename: [_diff_show_ratio $time1 $time2]"
1987 _log_and_puts log "COUNTER $key: [split $basename /] $casename: [_diff_show_ratio $time1 $time2]"
1994 # Procedure to compare results of two runs of test cases
1995 proc _test_diff {dir1 dir2 basename image cpu memory status verbose _logvar _logimage _logcpu _logmemory {_statvar ""}} {
1997 upvar $_logimage log_image
1998 upvar $_logcpu log_cpu
1999 upvar $_logmemory log_memory
2001 # make sure to load diffimage command
2002 uplevel pload VISUALIZATION
2004 # prepare variable (array) for collecting statistics
2005 if { "$_statvar" != "" } {
2006 upvar $_statvar stat
2018 # first check subdirectories
2019 set path1 [file join $dir1 $basename]
2020 set path2 [file join $dir2 $basename]
2021 set list1 [glob -directory $path1 -types d -tails -nocomplain *]
2022 set list2 [glob -directory $path2 -types d -tails -nocomplain *]
2023 if { [llength $list1] >0 || [llength $list2] > 0 } {
2024 _list_diff $list1 $list2 in1 in2 common
2025 if { "$verbose" > 1 } {
2026 if { [llength $in1] > 0 } { _log_and_puts log "Only in $path1: $in1" }
2027 if { [llength $in2] > 0 } { _log_and_puts log "Only in $path2: $in2" }
2029 foreach subdir $common {
2030 if { "$verbose" > 2 } {
2031 _log_and_puts log "Checking [file join $basename $subdir]"
2033 _test_diff $dir1 $dir2 [file join $basename $subdir] $image $cpu $memory $status $verbose log log_image log_cpu log_memory stat
2036 # check log files (only if directory has no subdirs)
2037 set list1 [glob -directory $path1 -types f -tails -nocomplain *.log]
2038 set list2 [glob -directory $path2 -types f -tails -nocomplain *.log]
2039 _list_diff $list1 $list2 in1 in2 common
2040 if { "$verbose" > 1 } {
2041 if { [llength $in1] > 0 } { _log_and_puts log "Only in $path1: $in1" }
2042 if { [llength $in2] > 0 } { _log_and_puts log "Only in $path2: $in2" }
2048 foreach logfile $common {
2050 set log1 [_read_file [file join $dir1 $basename $logfile]]
2051 set log2 [_read_file [file join $dir2 $basename $logfile]]
2052 set casename [file rootname $logfile]
2054 # check execution statuses
2055 if {$image == false && $cpu == false && $memory == false} {
2056 set status1 UNDEFINED
2057 set status2 UNDEFINED
2058 if { ! [regexp {CASE [^:]*:\s*([\w]+)} $log1 res1 status1] ||
2059 ! [regexp {CASE [^:]*:\s*([\w]+)} $log2 res2 status2] ||
2060 "$status1" != "$status2" } {
2061 _log_and_puts log "STATUS [split $basename /] $casename: $status1 / $status2"
2062 # if test statuses are different, further comparison makes
2063 # no sense unless explicitly requested
2064 if { "$status" != "all" } {
2068 if { "$status" == "ok" && "$status1" != "OK" } {
2074 # check CPU user time in test cases
2075 set checkCPURegexp "COUNTER (.+): (\[-0-9.+eE\]+)"
2076 if { [regexp "${checkCPURegexp}" $log1] &&
2077 [regexp "${checkCPURegexp}" $log2] } {
2078 _check_time "${checkCPURegexp}"
2083 if {$cpu != false || ($image == false && $cpu == false && $memory == false)} {
2086 if { [regexp {TOTAL CPU TIME:\s*([\d.]+)} $log1 res1 cpu1] &&
2087 [regexp {TOTAL CPU TIME:\s*([\d.]+)} $log2 res1 cpu2] } {
2088 set stat(cpu1) [expr $stat(cpu1) + $cpu1]
2089 set stat(cpu2) [expr $stat(cpu2) + $cpu2]
2090 set gcpu1 [expr $gcpu1 + $cpu1]
2091 set gcpu2 [expr $gcpu2 + $cpu2]
2093 # compare CPU times with 10% precision (but not less 0.5 sec)
2094 if { [expr abs ($cpu1 - $cpu2) > 0.5 + 0.05 * abs ($cpu1 + $cpu2)] } {
2095 if {$cpu != false} {
2096 _log_and_puts log_cpu "CPU [split $basename /] $casename: [_diff_show_ratio $cpu1 $cpu2]"
2098 _log_and_puts log "CPU [split $basename /] $casename: [_diff_show_ratio $cpu1 $cpu2]"
2104 # check memory delta
2105 if {$memory != false || ($image == false && $cpu == false && $memory == false)} {
2108 if { [regexp {MEMORY DELTA:\s*([\d.]+)} $log1 res1 mem1] &&
2109 [regexp {MEMORY DELTA:\s*([\d.]+)} $log2 res1 mem2] } {
2110 set stat(mem1) [expr $stat(mem1) + $mem1]
2111 set stat(mem2) [expr $stat(mem2) + $mem2]
2112 set gmem1 [expr $gmem1 + $mem1]
2113 set gmem2 [expr $gmem2 + $mem2]
2115 # compare memory usage with 10% precision (but not less 16 KiB)
2116 if { [expr abs ($mem1 - $mem2) > 16 + 0.05 * abs ($mem1 + $mem2)] } {
2117 if {$memory != false} {
2118 _log_and_puts log_memory "MEMORY [split $basename /] $casename: [_diff_show_ratio $mem1 $mem2]"
2120 _log_and_puts log "MEMORY [split $basename /] $casename: [_diff_show_ratio $mem1 $mem2]"
2127 if {$image != false || ($image == false && $cpu == false && $memory == false)} {
2128 set imglist1 [glob -directory $path1 -types f -tails -nocomplain ${casename}.{png,gif} ${casename}-*.{png,gif} ${casename}_*.{png,gif}]
2129 set imglist2 [glob -directory $path2 -types f -tails -nocomplain ${casename}.{png,gif} ${casename}-*.{png,gif} ${casename}_*.{png,gif}]
2130 _list_diff $imglist1 $imglist2 imgin1 imgin2 imgcommon
2131 if { "$verbose" > 1 } {
2132 if { [llength $imgin1] > 0 } {
2133 if {$image != false} {
2134 _log_and_puts log_image "Only in $path1: $imgin1"
2136 _log_and_puts log "Only in $path1: $imgin1"
2139 if { [llength $imgin2] > 0 } {
2140 if {$image != false} {
2141 _log_and_puts log_image "Only in $path2: $imgin2"
2143 _log_and_puts log "Only in $path2: $imgin2"
2147 foreach imgfile $imgcommon {
2148 # if { $verbose > 1 } { _log_and_puts log "Checking [split basename /] $casename: $imgfile" }
2149 set diffile [_diff_img_name $dir1 $dir2 $basename $imgfile]
2150 if { [catch {diffimage [file join $dir1 $basename $imgfile] \
2151 [file join $dir2 $basename $imgfile] \
2152 0 0 0 $diffile} diff] } {
2153 if {$image != false} {
2154 _log_and_puts log_image "IMAGE [split $basename /] $casename: $imgfile cannot be compared"
2156 _log_and_puts log "IMAGE [split $basename /] $casename: $imgfile cannot be compared"
2158 file delete -force $diffile ;# clean possible previous result of diffimage
2159 } elseif { $diff != 0 } {
2160 if {$image != false} {
2161 _log_and_puts log_image "IMAGE [split $basename /] $casename: $imgfile differs"
2163 _log_and_puts log "IMAGE [split $basename /] $casename: $imgfile differs"
2166 file delete -force $diffile ;# clean useless artifact of diffimage
2172 # report CPU and memory difference in group if it is greater than 10%
2173 if {$cpu != false || ($image == false && $cpu == false && $memory == false)} {
2174 if { [expr abs ($gcpu1 - $gcpu2) > 0.5 + 0.005 * abs ($gcpu1 + $gcpu2)] } {
2175 if {$cpu != false} {
2176 _log_and_puts log_cpu "CPU [split $basename /]: [_diff_show_ratio $gcpu1 $gcpu2]"
2178 _log_and_puts log "CPU [split $basename /]: [_diff_show_ratio $gcpu1 $gcpu2]"
2182 if {$memory != false || ($image == false && $cpu == false && $memory == false)} {
2183 if { [expr abs ($gmem1 - $gmem2) > 16 + 0.005 * abs ($gmem1 + $gmem2)] } {
2184 if {$memory != false} {
2185 _log_and_puts log_memory "MEMORY [split $basename /]: [_diff_show_ratio $gmem1 $gmem2]"
2187 _log_and_puts log "MEMORY [split $basename /]: [_diff_show_ratio $gmem1 $gmem2]"
2193 if { "$_statvar" == "" } {
2194 if {$memory != false || ($image == false && $cpu == false && $memory == false)} {
2195 if {$memory != false} {
2196 _log_and_puts log_memory "Total MEMORY difference: [_diff_show_ratio $stat(mem1) $stat(mem2)]"
2198 _log_and_puts log "Total MEMORY difference: [_diff_show_ratio $stat(mem1) $stat(mem2)]"
2201 if {$cpu != false || ($image == false && $cpu == false && $memory == false)} {
2202 if {$cpu != false} {
2203 _log_and_puts log_cpu "Total CPU difference: [_diff_show_ratio $stat(cpu1) $stat(cpu2)]"
2205 _log_and_puts log "Total CPU difference: [_diff_show_ratio $stat(cpu1) $stat(cpu2)]"
2211 # Auxiliary procedure to save log of results comparison to file
2212 proc _log_html_diff {file log dir1 dir2 highlight_percent} {
2213 # create missing directories as needed
2214 catch {file mkdir [file dirname $file]}
2216 # try to open a file
2217 if [catch {set fd [open $file w]} res] {
2218 error "Error saving log file $file: $res"
2222 puts $fd "<html><head><title>Diff $dir1 vs. $dir2</title></head><body>"
2223 puts $fd "<h1>Comparison of test results:</h1>"
2224 puts $fd "<h2>Version A - $dir1</h2>"
2225 puts $fd "<h2>Version B - $dir2</h2>"
2227 # add script for switching between images on click
2229 puts $fd "<script type=\"text/javascript\">"
2230 puts $fd " function diffimage_toggle(img,url1,url2)"
2232 puts $fd " if (img.show2nd) { img.src = url1; img.show2nd = false; }"
2233 puts $fd " else { img.src = url2; img.show2nd = true; }"
2235 puts $fd " function diffimage_reset(img,url) { img.src = url; img.show2nd = true; }"
2236 puts $fd "</script>"
2241 set logpath [file split [file normalize $file]]
2243 # put a line; highlight considerable (> ${highlight_percent}%) deviations of CPU and memory
2244 if { [regexp "\[\\\[](\[0-9.e+-]+)%\[\]]" $line res value] &&
2245 [expr abs($value)] > ${highlight_percent} } {
2246 puts $fd "<table><tr><td bgcolor=\"[expr $value > 0 ? \"ff8080\" : \"lightgreen\"]\">$line</td></tr></table>"
2252 if { [regexp {IMAGE[ \t]+([^:]+):[ \t]+([A-Za-z0-9_.-]+)} $line res case img] } {
2253 if { [catch {eval file join "" [lrange $case 0 end-1]} gridpath] } {
2254 # note: special handler for the case if test grid directoried are compared directly
2257 set aCaseName [lindex $case end]
2258 set img1url [_make_url $file [file join $dir1 $gridpath $img]]
2259 set img2url [_make_url $file [file join $dir2 $gridpath $img]]
2260 set img1 "<a href=\"[_make_url $file [file join $dir1 $gridpath $aCaseName.html]]\"><img src=\"$img1url\"></a>"
2261 set img2 "<a href=\"[_make_url $file [file join $dir2 $gridpath $aCaseName.html]]\"><img src=\"$img2url\"></a>"
2263 set difffile [_diff_img_name $dir1 $dir2 $gridpath $img]
2264 set imgdurl [_make_url $file $difffile]
2265 if { [file exists $difffile] } {
2266 set imgd "<img src=\"$imgdurl\" onmouseout=diffimage_reset(this,\"$imgdurl\") onclick=diffimage_toggle(this,\"$img1url\",\"$img2url\")>"
2271 puts $fd "<table><tr><th><abbr title=\"$dir1\">Version A</abbr></th><th><abbr title=\"$dir2\">Version B</abbr></th><th>Diff (click to toggle)</th></tr>"
2272 puts $fd "<tr><td>$img1</td><td>$img2</td><td>$imgd</td></tr></table>"
2275 puts $fd "</pre></body></html>"
2281 # get number of CPUs on the system
2282 proc _get_nb_cpus {} {
2283 global tcl_platform env
2285 if { "$tcl_platform(platform)" == "windows" } {
2286 # on Windows, take the value of the environment variable
2287 if { [info exists env(NUMBER_OF_PROCESSORS)] &&
2288 ! [catch {expr $env(NUMBER_OF_PROCESSORS) > 0} res] && $res >= 0 } {
2289 return $env(NUMBER_OF_PROCESSORS)
2291 } elseif { "$tcl_platform(os)" == "Linux" } {
2292 # on Linux, take number of logical processors listed in /proc/cpuinfo
2293 if { [catch {open "/proc/cpuinfo" r} fd] } {
2294 return 0 ;# should never happen, but...
2297 while { [gets $fd line] >= 0 } {
2298 if { [regexp {^processor[ \t]*:} $line] } {
2304 } elseif { "$tcl_platform(os)" == "Darwin" } {
2305 # on MacOS X, call sysctl command
2306 if { ! [catch {exec sysctl hw.ncpu} ret] &&
2307 [regexp {^hw[.]ncpu[ \t]*:[ \t]*([0-9]+)} $ret res nb] } {
2312 # if cannot get good value, return 0 as default
2316 # check two files for difference
2317 proc _diff_files {file1 file2} {
2318 set fd1 [open $file1 "r"]
2319 set fd2 [open $file2 "r"]
2323 set nb1 [gets $fd1 line1]
2324 set nb2 [gets $fd2 line2]
2325 if { $nb1 != $nb2 } { set differ t; break }
2326 if { $nb1 < 0 } { break }
2327 if { [string compare $line1 $line2] } {
2338 # Check if file is in DOS encoding.
2339 # This check is done by presence of \r\n combination at the end of the first
2340 # line (i.e. prior to any other \n symbol).
2341 # Note that presence of non-ascii symbols typically used for recognition
2342 # of binary files is not suitable since some IGES and STEP files contain
2343 # non-ascii symbols.
2344 # Special check is added for PNG files which contain \r\n in the beginning.
2345 proc _check_dos_encoding {file} {
2346 set fd [open $file rb]
2348 if { [gets $fd line] && [regexp {.*\r$} $line] &&
2349 ! [regexp {^.PNG} $line] } {
2356 # procedure to recognize format of a data file by its first symbols (for OCCT
2357 # BREP and geometry DRAW formats, IGES, and STEP) and extension (all others)
2358 proc _check_file_format {file} {
2359 set fd [open $file rb]
2360 set line [read $fd 1024]
2364 set ext [file extension $file]
2366 if { [regexp {^DBRep_DrawableShape} $line] } {
2368 if { "$ext" != ".brep" && "$ext" != ".rle" &&
2369 "$ext" != ".draw" && "$ext" != "" } {
2372 } elseif { [regexp {^DrawTrSurf_} $line] } {
2374 if { "$ext" != ".rle" &&
2375 "$ext" != ".draw" && "$ext" != "" } {
2378 } elseif { [regexp {^[ \t]*ISO-10303-21} $line] } {
2380 if { "$ext" != ".step" && "$ext" != ".stp" } {
2383 } elseif { [regexp {^.\{72\}S[0 ]\{6\}1} $line] } {
2385 if { "$ext" != ".iges" && "$ext" != ".igs" } {
2388 } elseif { "$ext" == ".igs" } {
2390 } elseif { "$ext" == ".stp" } {
2393 set format [string toupper [string range $ext 1 end]]
2397 puts "$file: Warning: extension ($ext) does not match format ($format)"
2403 # procedure to load file knowing its format
2404 proc load_data_file {file format shape} {
2406 BREP { uplevel restore $file $shape }
2407 DRAW { uplevel restore $file $shape }
2408 IGES { pload XSDRAW; uplevel igesbrep $file $shape * }
2409 STEP { pload XSDRAW; uplevel stepread $file __a *; uplevel renamevar __a_1 $shape }
2410 STL { pload XSDRAW; uplevel readstl $shape $file triangulation }
2411 default { error "Cannot read $format file $file" }
2415 # procedure to get name of temporary directory,
2416 # ensuring it is existing and writeable
2417 proc _get_temp_dir {} {
2418 global env tcl_platform
2420 # check typical environment variables
2421 foreach var {TempDir Temp Tmp} {
2422 # check different case
2423 foreach name [list [string toupper $var] $var [string tolower $var]] {
2424 if { [info exists env($name)] && [file isdirectory $env($name)] &&
2425 [file writable $env($name)] } {
2426 return [regsub -all {\\} $env($name) /]
2431 # check platform-specific locations
2433 if { "$tcl_platform(platform)" == "windows" } {
2434 set paths "c:/TEMP c:/TMP /TEMP /TMP"
2435 if { [info exists env(HOMEDRIVE)] && [info exists env(HOMEPATH)] } {
2436 set fallback [regsub -all {\\} "$env(HOMEDRIVE)$env(HOMEPATH)/tmp" /]
2439 set paths "/tmp /var/tmp /usr/tmp"
2440 if { [info exists env(HOME)] } {
2441 set fallback "$env(HOME)/tmp"
2444 foreach dir $paths {
2445 if { [file isdirectory $dir] && [file writable $dir] } {
2450 # fallback case: use subdir /tmp of home or current dir
2451 file mkdir $fallback
2455 # extract of code from testgrid command used to process jobs running in
2456 # parallel until number of jobs in the queue becomes equal or less than
2458 proc _testgrid_process_jobs {worker {nb_ok 0}} {
2459 # bind local vars to variables of the caller procedure
2462 upvar job_def job_def
2463 upvar nbpooled nbpooled
2464 upvar userbreak userbreak
2465 upvar refresh refresh
2466 upvar refresh_timer refresh_timer
2468 catch {tpool::resume $worker}
2469 while { ! $userbreak && $nbpooled > $nb_ok } {
2470 foreach job [tpool::wait $worker [array names job_def]] {
2471 eval _log_test_case \[tpool::get $worker $job\] $job_def($job) log
2476 # check for user break
2477 if { "[info commands dbreak]" == "dbreak" && [catch dbreak] } {
2481 # update summary log with requested period
2482 if { $logdir != "" && $refresh > 0 && [clock seconds] > $refresh_timer + $refresh } {
2483 _log_summarize $logdir $log
2484 set refresh_timer [clock seconds]
2487 catch {tpool::suspend $worker}
2492 Use: checkcolor x y red green blue
2493 x y - pixel coordinates
2494 red green blue - expected pixel color (values from 0 to 1)
2495 Function check color with tolerance (5x5 area)
2497 # Procedure to check color using command vreadpixel with tolerance
2498 proc checkcolor { coord_x coord_y rd_get gr_get bl_get } {
2499 puts "Coordinate x = $coord_x"
2500 puts "Coordinate y = $coord_y"
2501 puts "RED color of RGB is $rd_get"
2502 puts "GREEN color of RGB is $gr_get"
2503 puts "BLUE color of RGB is $bl_get"
2505 if { $coord_x <= 1 || $coord_y <= 1 } {
2506 puts "Error : minimal coordinate is x = 2, y = 2. But we have x = $coord_x y = $coord_y"
2511 catch { [set color "[vreadpixel ${coord_x} ${coord_y} rgb]"] }
2512 if {"$color" == ""} {
2513 puts "Error : Pixel coordinates (${position_x}; ${position_y}) are out of view"
2515 set rd [lindex $color 0]
2516 set gr [lindex $color 1]
2517 set bl [lindex $color 2]
2518 set rd_int [expr int($rd * 1.e+05)]
2519 set gr_int [expr int($gr * 1.e+05)]
2520 set bl_int [expr int($bl * 1.e+05)]
2521 set rd_ch [expr int($rd_get * 1.e+05)]
2522 set gr_ch [expr int($gr_get * 1.e+05)]
2523 set bl_ch [expr int($bl_get * 1.e+05)]
2525 if { $rd_ch != 0 } {
2526 set tol_rd [expr abs($rd_ch - $rd_int)/$rd_ch]
2530 if { $gr_ch != 0 } {
2531 set tol_gr [expr abs($gr_ch - $gr_int)/$gr_ch]
2535 if { $bl_ch != 0 } {
2536 set tol_bl [expr abs($bl_ch - $bl_int)/$bl_ch]
2542 if { $tol_rd > 0.2 } {
2543 puts "Warning : RED light of additive color model RGB is invalid"
2546 if { $tol_gr > 0.2 } {
2547 puts "Warning : GREEN light of additive color model RGB is invalid"
2550 if { $tol_bl > 0.2 } {
2551 puts "Warning : BLUE light of additive color model RGB is invalid"
2555 if { $status != 0 } {
2556 puts "Warning : Colors of default coordinate are not equal"
2560 if { $tol_rd > 0.2 || $tol_gr > 0.2 || $tol_bl > 0.2 } {
2561 set info [_checkpoint $coord_x $coord_y $rd_ch $gr_ch $bl_ch]
2562 set stat [lindex $info end]
2563 if { ${stat} != 1 } {
2564 puts "Error : Colors are not equal in default coordinate and in the near coordinates too"
2567 puts "Point with valid color was found"
2575 # Procedure to check color in the point near default coordinate
2576 proc _checkpoint {coord_x coord_y rd_ch gr_ch bl_ch} {
2577 set x_start [expr ${coord_x} - 2]
2578 set y_start [expr ${coord_y} - 2]
2581 while { $mistake != 1 && $i <= 5 } {
2583 while { $mistake != 1 && $j <= 5 } {
2584 set position_x [expr ${x_start} + $j]
2585 set position_y [expr ${y_start} + $i]
2590 catch { [set color "[vreadpixel ${position_x} ${position_y} rgb]"] }
2591 if {"$color" == ""} {
2592 puts "Warning : Pixel coordinates (${position_x}; ${position_y}) are out of view"
2596 set rd [lindex $color 0]
2597 set gr [lindex $color 1]
2598 set bl [lindex $color 2]
2599 set rd_int [expr int($rd * 1.e+05)]
2600 set gr_int [expr int($gr * 1.e+05)]
2601 set bl_int [expr int($bl * 1.e+05)]
2603 if { $rd_ch != 0 } {
2604 set tol_rd [expr abs($rd_ch - $rd_int)/$rd_ch]
2608 if { $gr_ch != 0 } {
2609 set tol_gr [expr abs($gr_ch - $gr_int)/$gr_ch]
2613 if { $bl_ch != 0 } {
2614 set tol_bl [expr abs($bl_ch - $bl_int)/$bl_ch]
2619 if { $tol_rd > 0.2 || $tol_gr > 0.2 || $tol_bl > 0.2 } {
2620 puts "Warning : Point with true color was not found near default coordinates"
2632 # Procedure to check if sequence of values in listval follows linear trend
2633 # adding the same delta on each step.
2635 # The function does statistical estimation of the mean variation of the
2636 # values of the sequence, and dispersion, and returns true only if both
2637 # dispersion and deviation of the mean from expected delta are within
2638 # specified tolerance.
2640 # If mean variation differs from expected delta on more than two dispersions,
2641 # the check fails and procedure raises error with specified message.
2643 # Otherwise the procedure returns false meaning that more iterations are needed.
2644 # Note that false is returned in any case if length of listval is less than 3.
2646 # See example of use to check memory leaks in bugs/caf/bug23489
2648 proc checktrend {listval delta tolerance message} {
2649 set nbval [llength $listval]
2654 # calculate mean value
2656 set prev [lindex $listval 0]
2657 foreach val [lrange $listval 1 end] {
2658 set mean [expr $mean + ($val - $prev)]
2661 set mean [expr $mean / ($nbval - 1)]
2663 # calculate dispersion
2665 set prev [lindex $listval 0]
2666 foreach val [lrange $listval 1 end] {
2667 set d [expr ($val - $prev) - $mean]
2668 set sigma [expr $sigma + $d * $d]
2671 set sigma [expr sqrt ($sigma / ($nbval - 2))]
2673 puts "Checking trend: nb = $nbval, mean delta = $mean, sigma = $sigma"
2675 # check if deviation is definitely too big
2676 if { abs ($mean - $delta) > $tolerance + 2. * $sigma } {
2677 puts "Checking trend failed: mean delta per step = $mean, sigma = $sigma, expected delta = $delta"
2681 # check if deviation is clearly within a range
2682 return [expr abs ($mean - $delta) <= $sigma && $sigma <= $tolerance]