webengine/osswebengine/WebKitTools/CodeCoverage/cov.py
changeset 0 dd21522fd290
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/webengine/osswebengine/WebKitTools/CodeCoverage/cov.py	Mon Mar 30 12:54:55 2009 +0300
@@ -0,0 +1,201 @@
+# Copyright (C) 2004, 2005, 2006 Nathaniel Smith
+# Copyright (C) 2006, 2007 Holger Hans Peter Freyther
+#
+# Redistribution and use in source and binary forms, with or without
+# modification, are permitted provided that the following conditions
+# are met:
+#
+# 1.  Redistributions of source code must retain the above copyright
+#     notice, this list of conditions and the following disclaimer. 
+# 2.  Redistributions in binary form must reproduce the above copyright
+#     notice, this list of conditions and the following disclaimer in the
+#     documentation and/or other materials provided with the distribution. 
+# 3.  Neither the name of Apple Computer, Inc. ("Apple") nor the names of
+#     its contributors may be used to endorse or promote products derived
+#     from this software without specific prior written permission. 
+#
+# THIS SOFTWARE IS PROVIDED BY APPLE AND ITS CONTRIBUTORS "AS IS" AND ANY
+# EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
+# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
+# DISCLAIMED. IN NO EVENT SHALL APPLE OR ITS CONTRIBUTORS BE LIABLE FOR ANY
+# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
+# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
+# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
+# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
+# THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+import csv
+import time
+import os.path
+import shutil
+
+def analyze_coverage(possible_gcov_files, source_files, runid, data_dir, base):
+
+    if not os.path.exists(data_dir):
+        os.makedirs(data_dir)
+
+    output = open(os.path.join(data_dir, runid + ".csv"), "w")
+    w = csv.writer(output)
+    # First row: id and time
+    w.writerow([runid, time.time()])
+
+    results = scan_gcov_files(possible_gcov_files, source_files)
+    annotated_dir = os.path.join(data_dir, runid + ".annotated")
+    if os.path.exists(annotated_dir):
+        shutil.rmtree(annotated_dir)
+
+    keys = results.keys()
+    keys.sort()
+    for path in keys:
+        (total, covered, annotated_data) = results[path]
+        path = path[path.find(base)+len(base):]
+        # Rest of the rows: filename, total_lines, covered_lines
+        w.writerow([path, total, covered])
+
+        if path[:1] == "/":
+            path = path[1:]
+        annotated_path = os.path.join(annotated_dir, path)
+        try:
+            os.makedirs(os.path.dirname(annotated_path))
+        except OSError:
+            pass
+        a = open(annotated_path, "w")
+        a.write(annotated_data)
+        a.close()
+
+
+# zecke's rewrite
+STATE_NOT_CODE = -1
+STATE_NOT_SEEN = -2
+STATE_TEST_CODE = -3
+
+def find_gcov(f, possible_gcovs):
+    """
+    Find .gcov files that could be of interest for us
+    """
+    try:
+        return possible_gcovs[f]
+    except:
+        return []
+
+
+def parse_source_file(file):
+    """
+    Parse one source file and return a list of lines
+    """
+    f_source_list = []
+    init_state = STATE_NOT_SEEN
+    in_test_code = False
+    nesting = 0
+
+    for line in open(file, "r"):
+        code = line.split(":", 2)[-1]
+        if not in_test_code and code.startswith("#ifdef BUILD_UNIT_TESTS"):
+            in_test_code = 1
+        if in_test_code and code.startswith("#if"):
+            nesting += 1
+        if in_test_code and code.startswith("#endif"):
+            nesting -= 1
+            if not nesting:
+                in_test_code = True
+        if in_test_code:
+            init_state = STATE_TEST_CODE
+        else:
+            init_state = STATE_NOT_SEEN
+        f_source_list.append([init_state, line.split(":", 1)[1]])
+
+    return f_source_list
+
+# Runner-up, 3rd annual "write Python that looks like Perl" competition,
+# Well, not really.  It doesn't even use regexps.
+# He is right so I'm cleaning it up (zecke)
+def scan_gcov_files(possible_gcov_files, source_files):
+    """Takes a list of gcov filenames and a list of source filenames.
+
+    The gcov files should have names of the form foo.o##foo.cc.gcov, as
+    created by 'gcov -l'.
+
+    Returns a dict mapping source filenames to tuples
+      (total_lines, tested_lines, gcov_annotated_source)
+    which are a number, a number, and a very long string, respectively.
+
+    The fun bit is that we merge .gcov output generated by different object
+    files; this way we can provide accurate information for header files and
+    for monotone's current unit test system."""
+    results = {}
+    for f in source_files:
+        possible_gcovs = find_gcov(f, possible_gcov_files)
+        base_name = os.path.splitext(os.path.basename(f))[0]
+        if len(possible_gcovs) == 0:
+            print "No gcov files found for: '%s' but it was compiled" % f
+            continue
+
+            (garbage,extension) = os.path.splitext(f)
+            if extension  in [".cc", ".c", ".moc", ".cpp", ".cxx", ".m", ".mm"]:
+                lines = open(f, "r").readlines()
+                results[f] = (len(lines), 0, "".join(lines))
+            continue
+        elif len(possible_gcovs) > 1:
+            print "More than one gcov file for %s %d" % (f,len(possible_gcovs))
+        base_gcov_lines = parse_source_file(possible_gcovs[0])
+
+        # Now we will try hard to merge the results with others
+        # Our requirement is that we have the same amount of lines as
+        # as the original file
+        for cov_file in possible_gcovs:
+            lines = open(cov_file, "r").readlines()
+
+            # e.g. with phonon we have visualisation.h and we can not know
+            # which header file (foldername) it is refering to. This is a gcov
+            # limitation and i have no workaround yet. We just hope we will pick
+            # the right header file...
+            if len(lines) != len(base_gcov_lines):
+                print "Error Base: %s and Target: %s have different amount of lines" % (possible_gcovs[0],cov_file)
+                continue
+
+            # now do the merging of the file. If it has the same basename
+            # and the same number of lines things might work out
+            # In the future take a look at the header of the file
+            i = 0
+            for line in lines:
+                accumulator = base_gcov_lines[i]
+                if accumulator[0] != STATE_TEST_CODE:
+                    info = line.split(":", 1)[0]
+                    if info.endswith("-"):
+                        if accumulator[0] == STATE_NOT_SEEN:
+                            accumulator[0] = STATE_NOT_CODE
+                    else:
+                        if info.endswith("#"):
+                            num = 0
+                        else:
+                            num = int(info)
+                        if accumulator[0] in (STATE_NOT_SEEN, STATE_NOT_CODE):
+                            accumulator[0] = 0
+                        accumulator[0] += num
+                i += 1
+
+        # post processing of ths file
+        (total_lines, total_covered) = (0, 0)
+        annotated_lines = []
+        for state, line in base_gcov_lines:
+            if state == STATE_NOT_SEEN:
+                desc = "?????"
+            elif state == STATE_TEST_CODE:
+                desc = "+"
+            elif state == STATE_NOT_CODE:
+                desc = "-"
+            elif state == 0:
+                desc = "#####"
+                total_lines += 1
+            else:
+                desc = str(state)
+                total_lines += 1
+                total_covered += 1
+            annotated_lines.append(":".join([desc.rjust(9), line]))
+        results[f] = (total_lines, total_covered, "".join(annotated_lines))
+    return results
+
+
+
+    return results