diff options
author | The Android Open Source Project <initial-contribution@android.com> | 2009-03-03 18:28:41 -0800 |
---|---|---|
committer | The Android Open Source Project <initial-contribution@android.com> | 2009-03-03 18:28:41 -0800 |
commit | 648161bb0edfc3d43db63caed5cc5213bc6cb78f (patch) | |
tree | 4b825dc642cb6eb9a060e54bf8d69288fbee4904 /WebKitTools/CodeCoverage/cov.py | |
parent | a65af38181ac7d34544586bdb5cd004de93897ad (diff) | |
download | external_webkit-648161bb0edfc3d43db63caed5cc5213bc6cb78f.zip external_webkit-648161bb0edfc3d43db63caed5cc5213bc6cb78f.tar.gz external_webkit-648161bb0edfc3d43db63caed5cc5213bc6cb78f.tar.bz2 |
auto import from //depot/cupcake/@135843
Diffstat (limited to 'WebKitTools/CodeCoverage/cov.py')
-rw-r--r-- | WebKitTools/CodeCoverage/cov.py | 201 |
1 files changed, 0 insertions, 201 deletions
diff --git a/WebKitTools/CodeCoverage/cov.py b/WebKitTools/CodeCoverage/cov.py deleted file mode 100644 index 443e601..0000000 --- a/WebKitTools/CodeCoverage/cov.py +++ /dev/null @@ -1,201 +0,0 @@ -# Copyright (C) 2004, 2005, 2006 Nathaniel Smith -# Copyright (C) 2006, 2007 Holger Hans Peter Freyther -# -# Redistribution and use in source and binary forms, with or without -# modification, are permitted provided that the following conditions -# are met: -# -# 1. Redistributions of source code must retain the above copyright -# notice, this list of conditions and the following disclaimer. -# 2. Redistributions in binary form must reproduce the above copyright -# notice, this list of conditions and the following disclaimer in the -# documentation and/or other materials provided with the distribution. -# 3. Neither the name of Apple Computer, Inc. ("Apple") nor the names of -# its contributors may be used to endorse or promote products derived -# from this software without specific prior written permission. -# -# THIS SOFTWARE IS PROVIDED BY APPLE AND ITS CONTRIBUTORS "AS IS" AND ANY -# EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED -# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE -# DISCLAIMED. IN NO EVENT SHALL APPLE OR ITS CONTRIBUTORS BE LIABLE FOR ANY -# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES -# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; -# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND -# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF -# THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -import csv -import time -import os.path -import shutil - -def analyze_coverage(possible_gcov_files, source_files, runid, data_dir, base): - - if not os.path.exists(data_dir): - os.makedirs(data_dir) - - output = open(os.path.join(data_dir, runid + ".csv"), "w") - w = csv.writer(output) - # First row: id and time - w.writerow([runid, time.time()]) - - results = scan_gcov_files(possible_gcov_files, source_files) - annotated_dir = os.path.join(data_dir, runid + ".annotated") - if os.path.exists(annotated_dir): - shutil.rmtree(annotated_dir) - - keys = results.keys() - keys.sort() - for path in keys: - (total, covered, annotated_data) = results[path] - path = path[path.find(base)+len(base):] - # Rest of the rows: filename, total_lines, covered_lines - w.writerow([path, total, covered]) - - if path[:1] == "/": - path = path[1:] - annotated_path = os.path.join(annotated_dir, path) - try: - os.makedirs(os.path.dirname(annotated_path)) - except OSError: - pass - a = open(annotated_path, "w") - a.write(annotated_data) - a.close() - - -# zecke's rewrite -STATE_NOT_CODE = -1 -STATE_NOT_SEEN = -2 -STATE_TEST_CODE = -3 - -def find_gcov(f, possible_gcovs): - """ - Find .gcov files that could be of interest for us - """ - try: - return possible_gcovs[f] - except: - return [] - - -def parse_source_file(file): - """ - Parse one source file and return a list of lines - """ - f_source_list = [] - init_state = STATE_NOT_SEEN - in_test_code = False - nesting = 0 - - for line in open(file, "r"): - code = line.split(":", 2)[-1] - if not in_test_code and code.startswith("#ifdef BUILD_UNIT_TESTS"): - in_test_code = 1 - if in_test_code and code.startswith("#if"): - nesting += 1 - if in_test_code and code.startswith("#endif"): - nesting -= 1 - if not nesting: - in_test_code = True - if in_test_code: - init_state = STATE_TEST_CODE - else: - init_state = STATE_NOT_SEEN - f_source_list.append([init_state, line.split(":", 1)[1]]) - - return f_source_list - -# Runner-up, 3rd annual "write Python that looks like Perl" competition, -# Well, not really. It doesn't even use regexps. -# He is right so I'm cleaning it up (zecke) -def scan_gcov_files(possible_gcov_files, source_files): - """Takes a list of gcov filenames and a list of source filenames. - - The gcov files should have names of the form foo.o##foo.cc.gcov, as - created by 'gcov -l'. - - Returns a dict mapping source filenames to tuples - (total_lines, tested_lines, gcov_annotated_source) - which are a number, a number, and a very long string, respectively. - - The fun bit is that we merge .gcov output generated by different object - files; this way we can provide accurate information for header files and - for monotone's current unit test system.""" - results = {} - for f in source_files: - possible_gcovs = find_gcov(f, possible_gcov_files) - base_name = os.path.splitext(os.path.basename(f))[0] - if len(possible_gcovs) == 0: - print "No gcov files found for: '%s' but it was compiled" % f - continue - - (garbage,extension) = os.path.splitext(f) - if extension in [".cc", ".c", ".moc", ".cpp", ".cxx", ".m", ".mm"]: - lines = open(f, "r").readlines() - results[f] = (len(lines), 0, "".join(lines)) - continue - elif len(possible_gcovs) > 1: - print "More than one gcov file for %s %d" % (f,len(possible_gcovs)) - base_gcov_lines = parse_source_file(possible_gcovs[0]) - - # Now we will try hard to merge the results with others - # Our requirement is that we have the same amount of lines as - # as the original file - for cov_file in possible_gcovs: - lines = open(cov_file, "r").readlines() - - # e.g. with phonon we have visualisation.h and we can not know - # which header file (foldername) it is refering to. This is a gcov - # limitation and i have no workaround yet. We just hope we will pick - # the right header file... - if len(lines) != len(base_gcov_lines): - print "Error Base: %s and Target: %s have different amount of lines" % (possible_gcovs[0],cov_file) - continue - - # now do the merging of the file. If it has the same basename - # and the same number of lines things might work out - # In the future take a look at the header of the file - i = 0 - for line in lines: - accumulator = base_gcov_lines[i] - if accumulator[0] != STATE_TEST_CODE: - info = line.split(":", 1)[0] - if info.endswith("-"): - if accumulator[0] == STATE_NOT_SEEN: - accumulator[0] = STATE_NOT_CODE - else: - if info.endswith("#"): - num = 0 - else: - num = int(info) - if accumulator[0] in (STATE_NOT_SEEN, STATE_NOT_CODE): - accumulator[0] = 0 - accumulator[0] += num - i += 1 - - # post processing of ths file - (total_lines, total_covered) = (0, 0) - annotated_lines = [] - for state, line in base_gcov_lines: - if state == STATE_NOT_SEEN: - desc = "?????" - elif state == STATE_TEST_CODE: - desc = "+" - elif state == STATE_NOT_CODE: - desc = "-" - elif state == 0: - desc = "#####" - total_lines += 1 - else: - desc = str(state) - total_lines += 1 - total_covered += 1 - annotated_lines.append(":".join([desc.rjust(9), line])) - results[f] = (total_lines, total_covered, "".join(annotated_lines)) - return results - - - - return results |