| 1 |
schorsch |
1.1 |
''' Text comparison functions for Radiance unit testing. |
| 2 |
|
|
|
| 3 |
|
|
This allows differences in whitespace, which is why the text |
| 4 |
|
|
corpora are split into tokens first. |
| 5 |
|
|
Tokens are then converted into an appropriate data type, so |
| 6 |
|
|
that floating point items will still be considered correct |
| 7 |
|
|
even if they are slightly different, eg. as a consequence of |
| 8 |
|
|
binary rounding errors. |
| 9 |
|
|
''' |
| 10 |
|
|
|
| 11 |
|
|
import string |
| 12 |
|
|
import types |
| 13 |
|
|
|
| 14 |
|
|
class error(Exception): pass |
| 15 |
|
|
|
| 16 |
|
|
# internal functions |
| 17 |
|
|
def _icompare(itest, iref): |
| 18 |
|
|
'''compare ints (not public)''' |
| 19 |
|
|
if type(itest) == str: |
| 20 |
|
|
iftest = int(itest) |
| 21 |
|
|
else: iftest = itest |
| 22 |
|
|
if iftest == iref: return 1 |
| 23 |
|
|
return 0 |
| 24 |
|
|
|
| 25 |
|
|
def _fcompare(ftest, fref): |
| 26 |
|
|
'''compare floats (not public)''' |
| 27 |
|
|
FUZZ = 0.0000001 # XXX heuristically determined |
| 28 |
|
|
if type(ftest) == str: |
| 29 |
|
|
fftest = float(ftest) |
| 30 |
|
|
else: fftest = ftest |
| 31 |
|
|
if (fftest < (fref + FUZZ)) and (fftest > (fref - FUZZ)): |
| 32 |
|
|
return 1 |
| 33 |
|
|
return 0 |
| 34 |
|
|
|
| 35 |
|
|
def _typify_token(t): |
| 36 |
|
|
'''return the token as int resp. float if possible (not public)''' |
| 37 |
|
|
try: return int(t) |
| 38 |
|
|
except ValueError: pass |
| 39 |
|
|
try: return float(t) |
| 40 |
|
|
except ValueError: pass |
| 41 |
|
|
return t |
| 42 |
|
|
|
| 43 |
|
|
|
| 44 |
|
|
# public comparison functions |
| 45 |
|
|
|
| 46 |
|
|
def lcompare(ltest, lref): |
| 47 |
|
|
'''compare a list of tokens |
| 48 |
|
|
raise an error if there are intolerable differences |
| 49 |
|
|
the reference tokens in lref should already be of the correct type. |
| 50 |
|
|
''' |
| 51 |
|
|
if len(ltest) != len(lref): |
| 52 |
|
|
raise error, ('List comparision failed: Different number of tokens' |
| 53 |
|
|
' (%d, %d)' % (len(ltest), len(lref))) |
| 54 |
|
|
for i in range(len(lref)): |
| 55 |
|
|
tref = lref[i] |
| 56 |
|
|
ttest = ltest[i] |
| 57 |
|
|
if type(tref) == str and tref != ttest: |
| 58 |
|
|
raise error, 'Token comparison failed: "%s" != "%s"' % (ttest, tref) |
| 59 |
|
|
elif type(tref) == int and not _icompare(ttest, tref): |
| 60 |
|
|
raise error, 'Token comparison failed: %s != %s' % (ttest, tref) |
| 61 |
|
|
elif type(tref) == float and not _fcompare(ttest, tref): |
| 62 |
|
|
raise error, 'Token comparison failed: %s != %s' % (ttest, tref) |
| 63 |
|
|
|
| 64 |
|
|
def llcompare(lltest, llref, ignore_empty=0, recurse=[]): |
| 65 |
|
|
'''compare a list of lists of tokens recursively |
| 66 |
|
|
raise an error if there are intolerable differences |
| 67 |
|
|
the reference tokens in lref should already be of the correct type. |
| 68 |
|
|
if ignore_empty is true, empty lines are not included in the comparison |
| 69 |
|
|
the recurse argument is only used internally |
| 70 |
|
|
''' |
| 71 |
|
|
if ignore_empty: |
| 72 |
|
|
lltest = filter(None, lltest) |
| 73 |
|
|
llref = filter(None, llref) |
| 74 |
|
|
if len(lltest) != len(llref): |
| 75 |
|
|
raise error, 'Comparision failed: Different number of lines (%d,%d)' %( |
| 76 |
|
|
len(lltest), len(llref)) |
| 77 |
|
|
for i in range(len(llref)): |
| 78 |
|
|
if llref[i]: |
| 79 |
|
|
rtype = type(llref[i][0]) |
| 80 |
|
|
if rtype == list or rtype == tuple: |
| 81 |
|
|
return llcompare(lltest[i], llref[i], |
| 82 |
|
|
recurse=recurse.append(i), ignore_empty=ignore_empty) |
| 83 |
|
|
try: lcompare(lltest[i], llref[i]) |
| 84 |
|
|
except error, e: |
| 85 |
|
|
if recurse: |
| 86 |
|
|
raise error, '%s (line %s)' % (str(e), recurse.append(i + 1)) |
| 87 |
|
|
else: raise error, '%s (line %d)' % (str(e), i + 1) |
| 88 |
|
|
|
| 89 |
|
|
def split_headers(s): |
| 90 |
|
|
'''split Radiance file headers |
| 91 |
|
|
return a list of lists of tokens suitable for llcompare() |
| 92 |
|
|
this is useful to check the output of getinfo''' |
| 93 |
|
|
ll = map(string.strip,string.split(s, '\n')) |
| 94 |
|
|
nll = [] |
| 95 |
|
|
for l in ll: |
| 96 |
|
|
parts = string.split(l, '=', 1) |
| 97 |
|
|
if len(parts) == 2: |
| 98 |
|
|
left = map(_typify_token, string.split(parts[0])) |
| 99 |
|
|
right = map(_typify_token, string.split(parts[1])) |
| 100 |
|
|
nll.append(left + ['='] + right) |
| 101 |
|
|
else: nll.append(map(_typify_token, string.split(l))) |
| 102 |
|
|
return nll |
| 103 |
|
|
|
| 104 |
|
|
def split_rad(s): |
| 105 |
|
|
'''Split the contents of a scene description string |
| 106 |
|
|
return a list of list of tokens suitable for llcompare()''' |
| 107 |
|
|
ll = map(string.strip,string.split(s, '\n')) |
| 108 |
|
|
nll = [] |
| 109 |
|
|
for l in ll: |
| 110 |
|
|
nll.append(map(_typify_token, string.split(l))) |
| 111 |
|
|
return nll |
| 112 |
|
|
|
| 113 |
|
|
def split_radfile(fn): |
| 114 |
|
|
'''Split the contents of a file object containing a scene description |
| 115 |
|
|
return a list of list of tokens suitable for llcompare()''' |
| 116 |
|
|
f = open(fn, 'r') |
| 117 |
|
|
ll = split_rad(f.read()) |
| 118 |
|
|
f.close() |
| 119 |
|
|
return ll |
| 120 |
|
|
|