summaryrefslogtreecommitdiffstats
path: root/WebKitTools/CodeCoverage/cov.py
diff options
context:
space:
mode:
authorThe Android Open Source Project <initial-contribution@android.com>2009-03-03 18:28:41 -0800
committerThe Android Open Source Project <initial-contribution@android.com>2009-03-03 18:28:41 -0800
commit648161bb0edfc3d43db63caed5cc5213bc6cb78f (patch)
tree4b825dc642cb6eb9a060e54bf8d69288fbee4904 /WebKitTools/CodeCoverage/cov.py
parenta65af38181ac7d34544586bdb5cd004de93897ad (diff)
downloadexternal_webkit-648161bb0edfc3d43db63caed5cc5213bc6cb78f.zip
external_webkit-648161bb0edfc3d43db63caed5cc5213bc6cb78f.tar.gz
external_webkit-648161bb0edfc3d43db63caed5cc5213bc6cb78f.tar.bz2
auto import from //depot/cupcake/@135843
Diffstat (limited to 'WebKitTools/CodeCoverage/cov.py')
-rw-r--r--WebKitTools/CodeCoverage/cov.py201
1 files changed, 0 insertions, 201 deletions
diff --git a/WebKitTools/CodeCoverage/cov.py b/WebKitTools/CodeCoverage/cov.py
deleted file mode 100644
index 443e601..0000000
--- a/WebKitTools/CodeCoverage/cov.py
+++ /dev/null
@@ -1,201 +0,0 @@
-# Copyright (C) 2004, 2005, 2006 Nathaniel Smith
-# Copyright (C) 2006, 2007 Holger Hans Peter Freyther
-#
-# Redistribution and use in source and binary forms, with or without
-# modification, are permitted provided that the following conditions
-# are met:
-#
-# 1. Redistributions of source code must retain the above copyright
-# notice, this list of conditions and the following disclaimer.
-# 2. Redistributions in binary form must reproduce the above copyright
-# notice, this list of conditions and the following disclaimer in the
-# documentation and/or other materials provided with the distribution.
-# 3. Neither the name of Apple Computer, Inc. ("Apple") nor the names of
-# its contributors may be used to endorse or promote products derived
-# from this software without specific prior written permission.
-#
-# THIS SOFTWARE IS PROVIDED BY APPLE AND ITS CONTRIBUTORS "AS IS" AND ANY
-# EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
-# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
-# DISCLAIMED. IN NO EVENT SHALL APPLE OR ITS CONTRIBUTORS BE LIABLE FOR ANY
-# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
-# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
-# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
-# THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-import csv
-import time
-import os.path
-import shutil
-
-def analyze_coverage(possible_gcov_files, source_files, runid, data_dir, base):
-
- if not os.path.exists(data_dir):
- os.makedirs(data_dir)
-
- output = open(os.path.join(data_dir, runid + ".csv"), "w")
- w = csv.writer(output)
- # First row: id and time
- w.writerow([runid, time.time()])
-
- results = scan_gcov_files(possible_gcov_files, source_files)
- annotated_dir = os.path.join(data_dir, runid + ".annotated")
- if os.path.exists(annotated_dir):
- shutil.rmtree(annotated_dir)
-
- keys = results.keys()
- keys.sort()
- for path in keys:
- (total, covered, annotated_data) = results[path]
- path = path[path.find(base)+len(base):]
- # Rest of the rows: filename, total_lines, covered_lines
- w.writerow([path, total, covered])
-
- if path[:1] == "/":
- path = path[1:]
- annotated_path = os.path.join(annotated_dir, path)
- try:
- os.makedirs(os.path.dirname(annotated_path))
- except OSError:
- pass
- a = open(annotated_path, "w")
- a.write(annotated_data)
- a.close()
-
-
-# zecke's rewrite
-STATE_NOT_CODE = -1
-STATE_NOT_SEEN = -2
-STATE_TEST_CODE = -3
-
-def find_gcov(f, possible_gcovs):
- """
- Find .gcov files that could be of interest for us
- """
- try:
- return possible_gcovs[f]
- except:
- return []
-
-
-def parse_source_file(file):
- """
- Parse one source file and return a list of lines
- """
- f_source_list = []
- init_state = STATE_NOT_SEEN
- in_test_code = False
- nesting = 0
-
- for line in open(file, "r"):
- code = line.split(":", 2)[-1]
- if not in_test_code and code.startswith("#ifdef BUILD_UNIT_TESTS"):
- in_test_code = 1
- if in_test_code and code.startswith("#if"):
- nesting += 1
- if in_test_code and code.startswith("#endif"):
- nesting -= 1
- if not nesting:
- in_test_code = True
- if in_test_code:
- init_state = STATE_TEST_CODE
- else:
- init_state = STATE_NOT_SEEN
- f_source_list.append([init_state, line.split(":", 1)[1]])
-
- return f_source_list
-
-# Runner-up, 3rd annual "write Python that looks like Perl" competition,
-# Well, not really. It doesn't even use regexps.
-# He is right so I'm cleaning it up (zecke)
-def scan_gcov_files(possible_gcov_files, source_files):
- """Takes a list of gcov filenames and a list of source filenames.
-
- The gcov files should have names of the form foo.o##foo.cc.gcov, as
- created by 'gcov -l'.
-
- Returns a dict mapping source filenames to tuples
- (total_lines, tested_lines, gcov_annotated_source)
- which are a number, a number, and a very long string, respectively.
-
- The fun bit is that we merge .gcov output generated by different object
- files; this way we can provide accurate information for header files and
- for monotone's current unit test system."""
- results = {}
- for f in source_files:
- possible_gcovs = find_gcov(f, possible_gcov_files)
- base_name = os.path.splitext(os.path.basename(f))[0]
- if len(possible_gcovs) == 0:
- print "No gcov files found for: '%s' but it was compiled" % f
- continue
-
- (garbage,extension) = os.path.splitext(f)
- if extension in [".cc", ".c", ".moc", ".cpp", ".cxx", ".m", ".mm"]:
- lines = open(f, "r").readlines()
- results[f] = (len(lines), 0, "".join(lines))
- continue
- elif len(possible_gcovs) > 1:
- print "More than one gcov file for %s %d" % (f,len(possible_gcovs))
- base_gcov_lines = parse_source_file(possible_gcovs[0])
-
- # Now we will try hard to merge the results with others
- # Our requirement is that we have the same amount of lines as
- # as the original file
- for cov_file in possible_gcovs:
- lines = open(cov_file, "r").readlines()
-
- # e.g. with phonon we have visualisation.h and we can not know
- # which header file (foldername) it is refering to. This is a gcov
- # limitation and i have no workaround yet. We just hope we will pick
- # the right header file...
- if len(lines) != len(base_gcov_lines):
- print "Error Base: %s and Target: %s have different amount of lines" % (possible_gcovs[0],cov_file)
- continue
-
- # now do the merging of the file. If it has the same basename
- # and the same number of lines things might work out
- # In the future take a look at the header of the file
- i = 0
- for line in lines:
- accumulator = base_gcov_lines[i]
- if accumulator[0] != STATE_TEST_CODE:
- info = line.split(":", 1)[0]
- if info.endswith("-"):
- if accumulator[0] == STATE_NOT_SEEN:
- accumulator[0] = STATE_NOT_CODE
- else:
- if info.endswith("#"):
- num = 0
- else:
- num = int(info)
- if accumulator[0] in (STATE_NOT_SEEN, STATE_NOT_CODE):
- accumulator[0] = 0
- accumulator[0] += num
- i += 1
-
- # post processing of ths file
- (total_lines, total_covered) = (0, 0)
- annotated_lines = []
- for state, line in base_gcov_lines:
- if state == STATE_NOT_SEEN:
- desc = "?????"
- elif state == STATE_TEST_CODE:
- desc = "+"
- elif state == STATE_NOT_CODE:
- desc = "-"
- elif state == 0:
- desc = "#####"
- total_lines += 1
- else:
- desc = str(state)
- total_lines += 1
- total_covered += 1
- annotated_lines.append(":".join([desc.rjust(9), line]))
- results[f] = (total_lines, total_covered, "".join(annotated_lines))
- return results
-
-
-
- return results