echo "PASSED"; \
done
+ ## other tests
+
@gcov -o coverage scanner.c libscanner.c > /dev/null 2> /dev/null
@mv *.gcov coverage; [ -f .gcov ] && mv .gcov coverage || true
@awk '/NOTEST/ { next } /^ *[1-9]/ {ran+=1} /^ *###/ {skip+=1} \
END {printf "coverage: %6.2f%%\n", ran * 100 / (ran + skip); \
- if (ran < (ran + skip) *0.90) exit(1) }' \
+ if (ran < (ran + skip) *0.95) exit(1) }' \
coverage/scanner.mdc.gcov
@rm -f .tmp*
15:0 newline()
15:0 eof()
+## Nested tests.
+
+We need to test various aspects of tokenizing code that is stored
+in multiple nodes. For example, comments and multi-line strings mustn't
+cross a node boundary.
+
+For this we tell `scanner` to extract sections directly from this file.
+As the file changes, line numbers might change as well, so we need to factor
+that out when testing. A simple awk script can normalise the first line number
+to one.
+
+###### other tests
+ @for T in $(scanner_section_tests); do \
+ echo -n "Test $$T ... "; \
+ i="$IFS"; IFS=,; set $$T; IFS="$$i"; section="$$1"; shift; \
+ ./md2c scanner-tests.mdc "output: $$T" | grep -v '^#' > .tmp.want; \
+ ./coverage_scanner --file scanner-tests.mdc --section "test: $$section" \
+ $${1+"$$@"} | awk -F: ' BEGIN {OFS=":"} $$1 ~ /^[0-9]/ {if (!first) first = $$1 - 1; \
+ $$1 = $$1 - first} { print } '> .tmp.have; \
+ if ! cmp -s .tmp.want .tmp.have; then \
+ echo "FAILED"; diff -u .tmp.want .tmp.have; exit 1; fi ; \
+ echo "PASSED"; \
+ done
+
+###### test list
+ scanner_section_tests += section1 section_string section_comment
+
+###### test: section1
+
+ foreach s in sections:
+ ## section2
+ print done
+
+###### section2
+
+ This is another
+ section
+
+###### output: section1
+ Tokenizing: test: section1
+ 1:8 ident(foreach)
+ 1:16 ident(s)
+ 1:18 ident(in)
+ 1:21 ident(sections)
+ 1:29 mark(:)
+ 7:16 in()
+ 7:16 ident(This)
+ 7:21 ident(is)
+ 7:24 ident(another)
+ 8:8 newline()
+ 8:8 out()
+ 8:8 in()
+ 8:8 ident(section)
+ 3:16 newline()
+ 3:16 ident(print)
+ 3:22 ident(done)
+ 4:0 newline()
+ 4:0 out()
+ 4:0 newline()
+ 4:0 eof()
+
+###### test: section_string
+ a = '''
+ A sting mustn't cross
+ ## string B
+ skip
+
+###### string B
+ to a new node
+ '''
+
+###### output: section_string
+ Tokenizing: test: section_string
+ 1:8 ident(a)
+ 1:10 mark(=)
+ 1:12 ERROR('''\x0a\x09 A sting mus..)
+ 7:8 in()
+ 7:8 ident(to)
+ 7:11 ident(a)
+ 7:13 ident(new)
+ 7:17 ident(node)
+ 8:8 newline()
+ 8:8 ERROR(''')
+ 4:8 newline()
+ 4:8 out()
+ 4:8 newline()
+ 4:8 ident(skip)
+ 5:0 newline()
+ 5:0 eof()
+
+###### test: section_comment
+ /* Mult-line comment must stay within
+ ## comment B
+ */
+
+###### comment B
+ a single node, they cannot cross nodes.
+
+###### output: section_comment
+ Tokenizing: test: section_comment
+ 1:8 ERROR(/* Mult-line comme..)
+ 6:8 ident(a)
+ 6:10 ident(single)
+ 6:17 ident(node)
+ 6:21 mark(,)
+ 6:23 ident(they)
+ 6:28 ident(cannot)
+ 6:35 ident(cross)
+ 6:41 ident(nodes)
+ 6:46 mark(.)
+ 3:8 newline()
+ 3:8 mark(*/)
+ 4:0 newline()
+ 4:0 eof()
+
## Ad-hoc test
These tests test bugs that were found in practice, and so prevent them recuring.