toolchain/gcc-linaro-6.3.1-2017.02-x8.../share/dejagnu/framework.exp

1031 lines
26 KiB
Plaintext

# Copyright (C) 1992, 1993, 1994, 1995, 1996, 1997, 1998, 1999, 2000,
# 2001 2002, 2003, 2005, 2006, 2007, 2008, 2009, 2010, 2012, 2016 Free
# Software Foundation, Inc.
#
# This file is part of DejaGnu.
#
# DejaGnu is free software; you can redistribute it and/or modify it
# under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 3 of the License, or
# (at your option) any later version.
#
# DejaGnu is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with DejaGnu; if not, write to the Free Software Foundation,
# Inc., 51 Franklin Street - Fifth Floor, Boston, MA 02110-1301, USA.
# This file was written by Rob Savoye <rob@welcomehome.org>.
# These variables are local to this file.
# This or more warnings and a test fails.
set warning_threshold 3
# This or more errors and a test fails.
set perror_threshold 1
proc mail_file { file to subject } {
if {[file readable $file]} {
catch "exec mail -s \"$subject\" $to < $file"
}
}
# Insert DTD for xml format checking.
#
proc insertdtd { } {
xml_output "<!DOCTYPE testsuite \[
<!-- testsuite.dtd -->
<!ELEMENT testsuite (test | summary)+>
<!ELEMENT test (input, output, result, name, prms_id )>
<!ELEMENT input (#PCDATA)>
<!ELEMENT output (#PCDATA)>
<!ELEMENT result (#PCDATA)>
<!ELEMENT name (#PCDATA)>
<!ELEMENT prms_id (#PCDATA)>
<!ELEMENT summary (result, description, total)>
<!ELEMENT description (#PCDATA)>
<!ELEMENT total (#PCDATA)>
\]>"
}
# Open the output logs.
#
proc open_logs { } {
global outdir
global tool
global sum_file
global xml_file
global xml_file_name
global xml
if { ${tool} == "" } {
set tool testrun
}
catch "file delete -force -- $outdir/$tool.sum"
set sum_file [open [file join $outdir $tool.sum] w]
if { $xml } {
catch "file delete -force -- $outdir/$tool.xml"
if { ![string compare $xml_file_name ""] } {
set xml_file_name $tool.xml
}
set xml_file [open [file join $outdir $xml_file_name] w]
xml_output "<?xml version=\"1.0\"?>"
insertdtd
xml_output "<testsuite>"
}
catch "file delete -force -- $outdir/$tool.log"
log_file -a "$outdir/$tool.log"
verbose "Opening log files in $outdir"
if { ${tool} == "testrun" } {
set tool ""
}
fconfigure $sum_file -buffering line
}
# Close the output logs.
#
proc close_logs { } {
global sum_file
global xml
global xml_file
if { $xml } {
xml_output "</testsuite>"
catch "close $xml_file"
}
catch "close $sum_file"
}
# Check build host triplet for PATTERN.
# With no arguments it returns the triplet string.
#
proc isbuild { pattern } {
global build_triplet
global host_triplet
if {![info exists build_triplet]} {
set build_triplet ${host_triplet}
}
if {[string match "" $pattern]} {
return $build_triplet
}
verbose "Checking pattern \"$pattern\" with $build_triplet" 2
if {[string match "$pattern" $build_triplet]} {
return 1
} else {
return 0
}
}
# Is $board remote? Return a non-zero value if so.
#
proc is_remote { board } {
global host_board
global target_list
verbose "calling is_remote $board" 3
# Remove any target variant specifications from the name.
set board [lindex [split $board "/"] 0]
# Map the host or build back into their short form.
if { [board_info build name] == $board } {
set board "build"
} elseif { [board_info host name] == $board } {
set board "host"
}
# We're on the "build". The check for the empty string is just for
# paranoia's sake--we shouldn't ever get one. "unix" is a magic
# string that should really go away someday.
if { $board == "build" || $board == "unix" || $board == "" } {
verbose "board is $board, not remote" 3
return 0
}
if { $board == "host" } {
if { [info exists host_board] && $host_board != "" } {
verbose "board is $board, is remote" 3
return 1
} else {
verbose "board is $board, host is local" 3
return 0
}
}
if { $board == "target" } {
global current_target_name
if {[info exists current_target_name]} {
# This shouldn't happen, but we'll be paranoid anyway.
if { $current_target_name != "target" } {
return [is_remote $current_target_name]
}
}
return 0
}
if {[board_info $board exists isremote]} {
verbose "board is $board, isremote is [board_info $board isremote]" 3
return [board_info $board isremote]
}
return 1
}
# If this is a Canadian (3 way) cross. This means the tools are
# being built with a cross compiler for another host.
#
proc is3way {} {
global host_triplet
global build_triplet
if {![info exists build_triplet]} {
set build_triplet ${host_triplet}
}
verbose "Checking $host_triplet against $build_triplet" 2
if { "$build_triplet" == "$host_triplet" } {
return 0
}
return 1
}
# Check host triplet for PATTERN.
# With no arguments it returns the triplet string.
#
proc ishost { pattern } {
global host_triplet
if {[string match "" $pattern]} {
return $host_triplet
}
verbose "Checking pattern \"$pattern\" with $host_triplet" 2
if {[string match "$pattern" $host_triplet]} {
return 1
} else {
return 0
}
}
# Check target triplet for pattern.
#
# With no arguments it returns the triplet string.
# Returns 1 if the target looked for, or 0 if not.
#
proc istarget { args } {
global target_triplet
# if no arg, return the config string
if {[string match "" $args]} {
if {[info exists target_triplet]} {
return $target_triplet
} else {
perror "No target configuration names found."
}
}
set triplet [lindex $args 0]
# now check against the cannonical name
if {[info exists target_triplet]} {
verbose "Checking \"$triplet\" against \"$target_triplet\"" 2
if {[string match $triplet $target_triplet]} {
return 1
}
}
# nope, no match
return 0
}
# Check to see if we're running the tests in a native environment
# Returns 1 if running native, 0 if on a target.
#
proc isnative { } {
global target_triplet
global build_triplet
if {[string match $build_triplet $target_triplet]} {
return 1
}
return 0
}
# unknown -- called by expect if a proc is called that doesn't exist
#
# Rename unknown to tcl_unknown so that we can wrap tcl_unknown.
# This allows Tcl package autoloading to work in the modern age.
rename ::unknown ::tcl_unknown
proc unknown args {
if {[catch {uplevel 1 ::tcl_unknown $args} msg]} {
global errorCode
global errorInfo
global exit_status
clone_output "ERROR: (DejaGnu) proc \"$args\" does not exist."
if {[info exists errorCode]} {
send_error "The error code is $errorCode\n"
}
if {[info exists errorInfo]} {
send_error "The info on the error is:\n$errorInfo\n"
}
set exit_status 2
log_and_exit
}
}
# Print output to stdout (or stderr) and to log file
#
# If the --all flag (-a) option was used then all messages go the the screen.
# Without this, all messages that start with a keyword are written only to the
# detail log file. All messages that go to the screen will also appear in the
# detail log. This should only be used by the framework itself using pass,
# fail, xpass, xfail, kpass, kfail, warning, perror, note, untested, unresolved,
# or unsupported procedures.
#
proc clone_output { message } {
global sum_file
global all_flag
if { $sum_file != "" } {
puts $sum_file "$message"
}
regsub "^\[ \t\]*(\[^ \t\]+).*$" "$message" "\\1" firstword
switch -glob -- "$firstword" {
"PASS:" -
"XFAIL:" -
"KFAIL:" -
"UNRESOLVED:" -
"UNSUPPORTED:" -
"UNTESTED:" {
if {$all_flag} {
send_user -- "$message\n"
return "$message"
} else {
send_log -- "$message\n"
}
}
{"ERROR:" "WARNING:" "NOTE:"} {
send_error -- "$message\n"
return "$message"
}
default {
send_user -- "$message\n"
return "$message"
}
}
}
# Reset a few counters.
#
proc reset_vars {} {
global test_names test_counts
global warncnt errcnt
# other miscellaneous variables
global prms_id
global bug_id
# reset them all
set prms_id 0
set bug_id 0
set warncnt 0
set errcnt 0
foreach x $test_names {
set test_counts($x,count) 0
}
# Variables local to this file.
global warning_threshold perror_threshold
set warning_threshold 3
set perror_threshold 1
}
proc log_and_exit {} {
global exit_status
global tool mail_logs outdir mailing_list
log_summary total
# extract version number
if {[info procs ${tool}_version] != ""} {
if {[catch "${tool}_version" output]} {
warning "${tool}_version failed:\n$output"
}
}
close_logs
verbose -log "runtest completed at [timestamp -format %c]"
if {$mail_logs} {
if { ${tool} == "" } {
set tool testrun
}
mail_file $outdir/$tool.sum $mailing_list "Dejagnu Summary Log"
}
remote_close host
remote_close target
exit $exit_status
}
proc xml_output { message } {
global xml_file
if { $xml_file != "" } {
puts $xml_file "$message"
}
}
# Print summary of all pass/fail counts.
#
proc log_summary { args } {
global tool
global sum_file
global xml_file
global xml
global exit_status
global mail_logs
global outdir
global mailing_list
global current_target_name
global test_counts
global testcnt
if { [llength $args] == 0 } {
set which "count"
} else {
set which [lindex $args 0]
}
if { [llength $args] == 0 } {
clone_output "\n\t\t=== $tool Summary for $current_target_name ===\n"
} else {
clone_output "\n\t\t=== $tool Summary ===\n"
}
# If the tool set `testcnt', it wants us to do a sanity check on the
# total count, so compare the reported number of testcases with the
# expected number. Maintaining an accurate count in `testcnt' isn't easy
# so it's not clear how often this will be used.
if {[info exists testcnt]} {
if { $testcnt > 0 } {
set totlcnt 0
# total all the testcases reported
foreach x { FAIL PASS XFAIL KFAIL XPASS KPASS UNTESTED UNRESOLVED UNSUPPORTED } {
incr totlcnt test_counts($x,$which)
}
set testcnt test_counts(total,$which)
if { $testcnt>$totlcnt || $testcnt<$totlcnt } {
if { $testcnt > $totlcnt } {
set mismatch "unreported [expr {$testcnt - $totlcnt}]"
}
if { $testcnt < $totlcnt } {
set mismatch "misreported [expr {$totlcnt - $testcnt}]"
}
} else {
verbose "# of testcases run $testcnt"
}
if {[info exists mismatch]} {
clone_output "### ERROR: totals do not equal number of testcases run"
clone_output "### ERROR: # of testcases expected $testcnt"
clone_output "### ERROR: # of testcases reported $totlcnt"
clone_output "### ERROR: # of testcases $mismatch\n"
}
}
}
foreach x { PASS FAIL XPASS XFAIL KPASS KFAIL UNRESOLVED UNTESTED UNSUPPORTED } {
set val $test_counts($x,$which)
if { $val > 0 } {
set mess "# of $test_counts($x,name)"
if { $xml } {
xml_output " <summary>"
xml_output " <result>$x</result>"
xml_output " <description>$mess</description>"
xml_output " <total>$val</total>"
xml_output " </summary>"
}
if { [string length $mess] < 24 } {
append mess "\t"
}
clone_output "$mess\t$val"
}
}
}
# Setup a flag to control whether a failure is expected or not
#
# Multiple target triplet patterns can be specified for targets
# for which the test fails. A bug report ID can be specified,
# which is a string without '-'.
#
proc setup_xfail { args } {
global xfail_flag
global xfail_prms
set xfail_prms 0
set argc [ llength $args ]
for { set i 0 } { $i < $argc } { incr i } {
set sub_arg [ lindex $args $i ]
# is a prms number. we assume this is a string with no '-' characters
if {[regexp "^\[^\-\]+$" $sub_arg]} {
set xfail_prms $sub_arg
continue
}
if {[istarget $sub_arg]} {
set xfail_flag 1
continue
}
}
}
# Setup a flag to control whether it is a known failure.
#
# A bug report ID _MUST_ be specified, and is the first argument.
# It still must be a string without '-' so we can be sure someone
# did not just forget it and we end-up using a target triple as
# bug id.
#
# Multiple target triplet patterns can be specified for targets
# for which the test is known to fail.
#
proc setup_kfail { args } {
global kfail_flag
global kfail_prms
set kfail_prms 0
set argc [ llength $args ]
for { set i 0 } { $i < $argc } { incr i } {
set sub_arg [ lindex $args $i ]
# is a prms number. we assume this is a string with no '-' characters
if {[regexp "^\[^\-\]+$" $sub_arg]} {
set kfail_prms $sub_arg
continue
}
if {[istarget $sub_arg]} {
set kfail_flag 1
continue
}
}
if {$kfail_prms == 0} {
perror "Attempt to set a kfail without specifying bug tracking id"
}
}
# Check to see if a conditional xfail is triggered.
# message {targets} {include} {exclude}
#
proc check_conditional_xfail { args } {
global compiler_flags
set all_args [lindex $args 0]
set message [lindex $all_args 0]
set target_list [lindex $all_args 1]
verbose "Limited to targets: $target_list" 3
# get the list of flags to look for
set includes [lindex $all_args 2]
verbose "Will search for options $includes" 3
# get the list of flags to exclude
if { [llength $all_args] > 3 } {
set excludes [lindex $all_args 3]
verbose "Will exclude for options $excludes" 3
} else {
set excludes ""
}
# loop through all the targets, checking the options for each one
verbose "Compiler flags are: $compiler_flags" 2
set incl_hit 0
set excl_hit 0
foreach targ $target_list {
if {[istarget $targ]} {
# look through the compiler options for flags we want to see
# this is really messy cause each set of options to look for
# may also be a list. We also want to find each element of the
# list, regardless of order to make sure they're found.
# So we look for lists in side of lists, and make sure all
# the elements match before we decide this is legit.
# Se we 'incl_hit' to 1 before the loop so that if the 'includes'
# list is empty, this test will report a hit. (This can be
# useful if a target will always fail unless certain flags,
# specified in the 'excludes' list, are used.)
set incl_hit 1
for { set i 0 } { $i < [llength $includes] } { incr i } {
set incl_hit 0
set opt [lindex $includes $i]
verbose "Looking for $opt to include in the compiler flags" 2
foreach j "$opt" {
if {[string match "* $j *" $compiler_flags]} {
verbose "Found $j to include in the compiler flags" 2
incr incl_hit
}
}
# if the number of hits we get is the same as the number of
# specified options, then we got a match
if {$incl_hit == [llength $opt]} {
break
} else {
set incl_hit 0
}
}
# look through the compiler options for flags we don't
# want to see
for { set i 0 } { $i < [llength $excludes] } { incr i } {
set excl_hit 0
set opt [lindex $excludes $i]
verbose "Looking for $opt to exclude in the compiler flags" 2
foreach j "$opt" {
if {[string match "* $j *" $compiler_flags]} {
verbose "Found $j to exclude in the compiler flags" 2
incr excl_hit
}
}
# if the number of hits we get is the same as the number of
# specified options, then we got a match
if {$excl_hit == [llength $opt]} {
break
} else {
set excl_hit 0
}
}
# if we got a match for what to include, but didn't find any reasons
# to exclude this, then we got a match! So return one to turn this into
# an expected failure.
if {$incl_hit && ! $excl_hit } {
verbose "This is a conditional match" 2
return 1
} else {
verbose "This is not a conditional match" 2
return 0
}
}
}
return 0
}
# Clear the xfail flag for a particular target.
#
proc clear_xfail { args } {
global xfail_flag
global xfail_prms
set argc [ llength $args ]
for { set i 0 } { $i < $argc } { incr i } {
set sub_arg [ lindex $args $i ]
switch -glob -- $sub_arg {
"*-*-*" { # is a configuration triplet
if {[istarget $sub_arg]} {
set xfail_flag 0
set xfail_prms 0
}
continue
}
}
}
}
# Clear the kfail flag for a particular target.
#
proc clear_kfail { args } {
global kfail_flag
global kfail_prms
set argc [ llength $args ]
for { set i 0 } { $i < $argc } { incr i } {
set sub_arg [ lindex $args $i ]
switch -glob -- $sub_arg {
"*-*-*" { # is a configuration triplet
if {[istarget $sub_arg]} {
set kfail_flag 0
set kfail_prms 0
}
continue
}
}
}
}
# Record that a test has passed or failed (perhaps unexpectedly).
# This is an internal procedure, only used in this file.
#
proc record_test { type message args } {
global exit_status
global xml
global prms_id bug_id
global xfail_flag xfail_prms
global kfail_flag kfail_prms
global errcnt warncnt
global warning_threshold perror_threshold
global pf_prefix
if { [llength $args] > 0 } {
set count [lindex $args 0]
} else {
set count 1
}
if {[info exists pf_prefix]} {
set message [concat $pf_prefix " " $message]
}
# If we have too many warnings or errors,
# the output of the test can't be considered correct.
if { $warning_threshold > 0 && $warncnt >= $warning_threshold
|| $perror_threshold > 0 && $errcnt >= $perror_threshold } {
verbose "Error/Warning threshold exceeded: \
$errcnt $warncnt (max. $perror_threshold $warning_threshold)"
set type UNRESOLVED
}
incr_count $type
if { $xml } {
global errorInfo
set error ""
if {[info exists errorInfo]} {
set error $errorInfo
}
global expect_out
set rio { "" "" }
if { [catch { set rio [split $expect_out(buffer) "\n"] } result]} {
#do nothing - leave as { "" "" }
}
set output ""
set output "expect_out(buffer)"
xml_output " <test>"
xml_output " <input>[string trimright [lindex $rio 0]]</input>"
xml_output " <output>[string trimright [lindex $rio 1]]</output>"
xml_output " <result>$type</result>"
xml_output " <name>$message</name>"
xml_output " <prms_id>$prms_id</prms_id>"
xml_output " </test>"
}
switch -- $type {
PASS {
if {$prms_id} {
set message [concat $message "\t(PRMS $prms_id)"]
}
}
FAIL {
set exit_status 1
if {$prms_id} {
set message [concat $message "\t(PRMS $prms_id)"]
}
}
XPASS {
set exit_status 1
if { $xfail_prms != 0 } {
set message [concat $message "\t(PRMS $xfail_prms)"]
}
}
XFAIL {
if { $xfail_prms != 0 } {
set message [concat $message "\t(PRMS $xfail_prms)"]
}
}
KPASS {
set exit_status 1
if { $kfail_prms != 0 } {
set message [concat $message "\t(PRMS $kfail_prms)"]
}
}
KFAIL {
if { $kfail_prms != 0 } {
set message [concat $message "\t(PRMS: $kfail_prms)"]
}
}
UNTESTED {
# The only reason we look at the xfail/kfail stuff is to pick up
# `xfail_prms'.
if { $kfail_flag && $kfail_prms != 0 } {
set message [concat $message "\t(PRMS $kfail_prms)"]
} elseif { $xfail_flag && $xfail_prms != 0 } {
set message [concat $message "\t(PRMS $xfail_prms)"]
} elseif { $prms_id } {
set message [concat $message "\t(PRMS $prms_id)"]
}
}
UNRESOLVED {
set exit_status 1
# The only reason we look at the xfail/kfail stuff is to pick up
# `xfail_prms'.
if { $kfail_flag && $kfail_prms != 0 } {
set message [concat $message "\t(PRMS $kfail_prms)"]
} elseif { $xfail_flag && $xfail_prms != 0 } {
set message [concat $message "\t(PRMS $xfail_prms)"]
} elseif { $prms_id } {
set message [concat $message "\t(PRMS $prms_id)"]
}
}
UNSUPPORTED {
# The only reason we look at the xfail/kfail stuff is to pick up
# `xfail_prms'.
if { $kfail_flag && $kfail_prms != 0 } {
set message [concat $message "\t(PRMS $kfail_prms)"]
} elseif { $xfail_flag && $xfail_prms != 0 } {
set message [concat $message "\t(PRMS $xfail_prms)"]
} elseif { $prms_id } {
set message [concat $message "\t(PRMS $prms_id)"]
}
}
default {
perror "record_test called with bad type `$type'"
set errcnt 0
return
}
}
if { $bug_id } {
set message [concat $message "\t(BUG $bug_id)"]
}
global multipass_name
if { $multipass_name != "" } {
set message [format "$type: %s: $message" "$multipass_name"]
} else {
set message "$type: $message"
}
clone_output "$message"
# If a command name exists in the $local_record_procs associative
# array for this type of result, then invoke it.
set lowcase_type [string tolower $type]
global local_record_procs
if {[info exists local_record_procs($lowcase_type)]} {
$local_record_procs($lowcase_type) "$message"
}
# Reset these so they're ready for the next test case. We don't reset
# prms_id or bug_id here. There may be multiple tests for them. Instead
# they are reset in the main loop after each test. It is also the
# testsuite driver's responsibility to reset them after each testcase.
set warncnt 0
set errcnt 0
set xfail_flag 0
set kfail_flag 0
set xfail_prms 0
set kfail_prms 0
}
# Record that a test has passed.
#
proc pass { message } {
global xfail_flag kfail_flag compiler_conditional_xfail_data
# if we have a conditional xfail setup, then see if our compiler flags match
if {[ info exists compiler_conditional_xfail_data ]} {
if {[check_conditional_xfail $compiler_conditional_xfail_data]} {
set xfail_flag 1
}
unset compiler_conditional_xfail_data
}
if { $kfail_flag } {
record_test KPASS $message
} elseif { $xfail_flag } {
record_test XPASS $message
} else {
record_test PASS $message
}
}
# Record that a test has failed.
#
proc fail { message } {
global xfail_flag kfail_flag compiler_conditional_xfail_data
# if we have a conditional xfail setup, then see if our compiler flags match
if {[ info exists compiler_conditional_xfail_data ]} {
if {[check_conditional_xfail $compiler_conditional_xfail_data]} {
set xfail_flag 1
}
unset compiler_conditional_xfail_data
}
if { $kfail_flag } {
record_test KFAIL $message
} elseif { $xfail_flag } {
record_test XFAIL $message
} else {
record_test FAIL $message
}
}
# Record that a test that was expected to fail has passed unexpectedly.
#
proc xpass { message } {
record_test XPASS $message
}
# Record that a test that was expected to fail did indeed fail.
#
proc xfail { message } {
record_test XFAIL $message
}
# Record that a test for a known bug has passed unexpectedly.
#
proc kpass { bugid message } {
global kfail_flag kfail_prms
set kfail_flag 1
set kfail_prms $bugid
record_test KPASS $message
}
# Record that a test has failed due to a known bug.
#
proc kfail { bugid message } {
global kfail_flag kfail_prms
set kfail_flag 1
set kfail_prms $bugid
record_test KFAIL $message
}
# Set warning threshold.
#
proc set_warning_threshold { threshold } {
global warning_threshold
set warning_threshold $threshold
}
# Get warning threshold.
#
proc get_warning_threshold { } {
global warning_threshold
return $warning_threshold
}
# Prints warning messages.
# These are warnings from the framework, not from the tools being
# tested. It takes a string, and an optional number and returns
# nothing.
#
proc warning { args } {
global warncnt
if { [llength $args] > 1 } {
set warncnt [lindex $args 1]
} else {
incr warncnt
}
set message [lindex $args 0]
clone_output "WARNING: $message"
global errorInfo
if {[info exists errorInfo]} {
unset errorInfo
}
}
# Prints error messages.
# These are errors from the framework, not from the tools being
# tested. It takes a string, and an optional number and returns
# nothing.
#
proc perror { args } {
global errcnt
if { [llength $args] > 1 } {
set errcnt [lindex $args 1]
} else {
incr errcnt
}
set message [lindex $args 0]
clone_output "ERROR: $message"
global errorInfo
if {[info exists errorInfo]} {
unset errorInfo
}
}
# Prints informational messages.
#
# These are messages from the framework, not from the tools being
# tested. This means that it is currently illegal to call this proc
# outside of dejagnu proper.
#
proc note { message } {
clone_output "NOTE: $message"
}
# untested -- mark the test case as untested.
#
proc untested { message } {
record_test UNTESTED $message
}
# Mark the test case as unresolved.
#
proc unresolved { message } {
record_test UNRESOLVED $message
}
# Mark the test case as unsupported.
# Usually this is used for a test that is missing OS support.
#
proc unsupported { message } {
record_test UNSUPPORTED $message
}
# Set up the values in the test_counts array (name and initial
# totals).
#
proc init_testcounts { } {
global test_counts test_names
set test_counts(TOTAL,name) "testcases run"
set test_counts(PASS,name) "expected passes"
set test_counts(FAIL,name) "unexpected failures"
set test_counts(XFAIL,name) "expected failures"
set test_counts(XPASS,name) "unexpected successes"
set test_counts(KFAIL,name) "known failures"
set test_counts(KPASS,name) "unknown successes"
set test_counts(WARNING,name) "warnings"
set test_counts(ERROR,name) "errors"
set test_counts(UNSUPPORTED,name) "unsupported tests"
set test_counts(UNRESOLVED,name) "unresolved testcases"
set test_counts(UNTESTED,name) "untested testcases"
set j ""
foreach i [lsort [array names test_counts]] {
regsub ",.*$" "$i" "" i
if { $i == $j } {
continue
}
set test_counts($i,total) 0
lappend test_names $i
set j $i
}
}
# Increment NAME in the test_counts array; the amount to increment can
# be is optional (defaults to 1).
#
proc incr_count { name args } {
global test_counts
if { [llength $args] == 0 } {
set count 1
} else {
set count [lindex $args 0]
}
if {[info exists test_counts($name,count)]} {
incr test_counts($name,count) $count
incr test_counts($name,total) $count
} else {
perror "$name doesn't exist in incr_count"
}
}