Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(22)

Unified Diff: third_party/lcov-1.9/contrib/galaxy/genflat.pl

Issue 23189008: Upgrades lcov to 1.10, removes lcov-1.9 (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: Re-adds UNKNOWN suppression Created 7 years, 4 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: third_party/lcov-1.9/contrib/galaxy/genflat.pl
diff --git a/third_party/lcov-1.9/contrib/galaxy/genflat.pl b/third_party/lcov-1.9/contrib/galaxy/genflat.pl
deleted file mode 100755
index b8b8ff47b1a666db978b207436f88672a420344c..0000000000000000000000000000000000000000
--- a/third_party/lcov-1.9/contrib/galaxy/genflat.pl
+++ /dev/null
@@ -1,1238 +0,0 @@
-#!/usr/bin/perl -w
-#
-# Copyright (c) International Business Machines Corp., 2002
-#
-# This program is free software; you can redistribute it and/or modify
-# it under the terms of the GNU General Public License as published by
-# the Free Software Foundation; either version 2 of the License, or (at
-# your option) any later version.
-#
-# This program is distributed in the hope that it will be useful, but
-# WITHOUT ANY WARRANTY; without even the implied warranty of
-# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
-# General Public License for more details.
-#
-# You should have received a copy of the GNU General Public License
-# along with this program; if not, write to the Free Software
-# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
-#
-#
-# genflat
-#
-# This script generates std output from .info files as created by the
-# geninfo script. Call it with --help to get information on usage and
-# available options. This code is based on the lcov genhtml script
-# by Peter Oberparleiter <Peter.Oberparleiter@de.ibm.com>
-#
-#
-# History:
-# 2003-08-19 ripped up Peter's script James M Kenefick Jr. <jkenefic@us.ibm.com>
-#
-
-use strict;
-use File::Basename;
-use Getopt::Long;
-# Constants
-our $lcov_version = "";
-our $lcov_url = "";
-
-# Specify coverage rate limits (in %) for classifying file entries
-# HI: $hi_limit <= rate <= 100 graph color: green
-# MED: $med_limit <= rate < $hi_limit graph color: orange
-# LO: 0 <= rate < $med_limit graph color: red
-our $hi_limit = 50;
-our $med_limit = 15;
-
-# Data related prototypes
-sub print_usage(*);
-sub gen_html();
-sub process_dir($);
-sub process_file($$$);
-sub info(@);
-sub read_info_file($);
-sub get_info_entry($);
-sub set_info_entry($$$$;$$);
-sub get_prefix(@);
-sub shorten_prefix($);
-sub get_dir_list(@);
-sub get_relative_base_path($);
-sub get_date_string();
-sub split_filename($);
-sub subtract_counts($$);
-sub add_counts($$);
-sub apply_baseline($$);
-sub combine_info_files($$);
-sub combine_info_entries($$);
-sub apply_prefix($$);
-sub escape_regexp($);
-
-
-# HTML related prototypes
-
-
-sub write_file_table(*$$$$);
-
-
-# Global variables & initialization
-our %info_data; # Hash containing all data from .info file
-our $dir_prefix; # Prefix to remove from all sub directories
-our %test_description; # Hash containing test descriptions if available
-our $date = get_date_string();
-
-our @info_filenames; # List of .info files to use as data source
-our $test_title; # Title for output as written to each page header
-our $output_directory; # Name of directory in which to store output
-our $base_filename; # Optional name of file containing baseline data
-our $desc_filename; # Name of file containing test descriptions
-our $css_filename; # Optional name of external stylesheet file to use
-our $quiet; # If set, suppress information messages
-our $help; # Help option flag
-our $version; # Version option flag
-our $show_details; # If set, generate detailed directory view
-our $no_prefix; # If set, do not remove filename prefix
-our $frames; # If set, use frames for source code view
-our $keep_descriptions; # If set, do not remove unused test case descriptions
-our $no_sourceview; # If set, do not create a source code view for each file
-our $tab_size = 8; # Number of spaces to use in place of tab
-
-our $cwd = `pwd`; # Current working directory
-chomp($cwd);
-our $tool_dir = dirname($0); # Directory where genhtml tool is installed
-
-
-#
-# Code entry point
-#
-
-# Add current working directory if $tool_dir is not already an absolute path
-if (! ($tool_dir =~ /^\/(.*)$/))
-{
- $tool_dir = "$cwd/$tool_dir";
-}
-
-# Parse command line options
-if (!GetOptions("output-directory=s" => \$output_directory,
- "css-file=s" => \$css_filename,
- "baseline-file=s" => \$base_filename,
- "prefix=s" => \$dir_prefix,
- "num-spaces=i" => \$tab_size,
- "no-prefix" => \$no_prefix,
- "quiet" => \$quiet,
- "help" => \$help,
- "version" => \$version
- ))
-{
- print_usage(*STDERR);
- exit(1);
-}
-
-@info_filenames = @ARGV;
-
-# Check for help option
-if ($help)
-{
- print_usage(*STDOUT);
- exit(0);
-}
-
-# Check for version option
-if ($version)
-{
- print($lcov_version."\n");
- exit(0);
-}
-
-# Check for info filename
-if (!@info_filenames)
-{
- print(STDERR "No filename specified\n");
- print_usage(*STDERR);
- exit(1);
-}
-
-# Generate a title if none is specified
-if (!$test_title)
-{
- if (scalar(@info_filenames) == 1)
- {
- # Only one filename specified, use it as title
- $test_title = basename($info_filenames[0]);
- }
- else
- {
- # More than one filename specified, used default title
- $test_title = "unnamed";
- }
-}
-
-# Make sure tab_size is within valid range
-if ($tab_size < 1)
-{
- print(STDERR "ERROR: invalid number of spaces specified: ".
- "$tab_size!\n");
- exit(1);
-}
-
-# Do something
-gen_html();
-
-exit(0);
-
-
-
-#
-# print_usage(handle)
-#
-# Print usage information.
-#
-
-sub print_usage(*)
-{
- local *HANDLE = $_[0];
- my $executable_name = basename($0);
-
- print(HANDLE <<END_OF_USAGE);
-Usage: $executable_name [OPTIONS] INFOFILE(S)
-
-Create HTML output for coverage data found in INFOFILE. Note that INFOFILE
-may also be a list of filenames.
-
- -h, --help Print this help, then exit
- -v, --version Print version number, then exit
- -q, --quiet Do not print progress messages
- -b, --baseline-file BASEFILE Use BASEFILE as baseline file
- -p, --prefix PREFIX Remove PREFIX from all directory names
- --no-prefix Do not remove prefix from directory names
- --no-source Do not create source code view
- --num-spaces NUM Replace tabs with NUM spaces in source view
-
-See $lcov_url for more information about this tool.
-END_OF_USAGE
- ;
-}
-
-
-#
-# gen_html()
-#
-# Generate a set of HTML pages from contents of .info file INFO_FILENAME.
-# Files will be written to the current directory. If provided, test case
-# descriptions will be read from .tests file TEST_FILENAME and included
-# in ouput.
-#
-# Die on error.
-#
-
-sub gen_html()
-{
- local *HTML_HANDLE;
- my %overview;
- my %base_data;
- my $lines_found;
- my $lines_hit;
- my $overall_found = 0;
- my $overall_hit = 0;
- my $dir_name;
- my $link_name;
- my @dir_list;
- my %new_info;
-
- # Read in all specified .info files
- foreach (@info_filenames)
- {
- info("Reading data file $_\n");
- %new_info = %{read_info_file($_)};
-
- # Combine %new_info with %info_data
- %info_data = %{combine_info_files(\%info_data, \%new_info)};
- }
-
- info("Found %d entries.\n", scalar(keys(%info_data)));
-
- # Read and apply baseline data if specified
- if ($base_filename)
- {
- # Read baseline file
- info("Reading baseline file $base_filename\n");
- %base_data = %{read_info_file($base_filename)};
- info("Found %d entries.\n", scalar(keys(%base_data)));
-
- # Apply baseline
- info("Subtracting baseline data.\n");
- %info_data = %{apply_baseline(\%info_data, \%base_data)};
- }
-
- @dir_list = get_dir_list(keys(%info_data));
-
- if ($no_prefix)
- {
- # User requested that we leave filenames alone
- info("User asked not to remove filename prefix\n");
- }
- elsif (!defined($dir_prefix))
- {
- # Get prefix common to most directories in list
- $dir_prefix = get_prefix(@dir_list);
-
- if ($dir_prefix)
- {
- info("Found common filename prefix \"$dir_prefix\"\n");
- }
- else
- {
- info("No common filename prefix found!\n");
- $no_prefix=1;
- }
- }
- else
- {
- info("Using user-specified filename prefix \"".
- "$dir_prefix\"\n");
- }
-
- # Process each subdirectory and collect overview information
- foreach $dir_name (@dir_list)
- {
- ($lines_found, $lines_hit) = process_dir($dir_name);
-
- $overview{$dir_name} = "$lines_found,$lines_hit, ";
- $overall_found += $lines_found;
- $overall_hit += $lines_hit;
- }
-
-
- if ($overall_found == 0)
- {
- info("Warning: No lines found!\n");
- }
- else
- {
- info("Overall coverage rate: %d of %d lines (%.1f%%)\n",
- $overall_hit, $overall_found,
- $overall_hit*100/$overall_found);
- }
-}
-
-
-#
-# process_dir(dir_name)
-#
-
-sub process_dir($)
-{
- my $abs_dir = $_[0];
- my $trunc_dir;
- my $rel_dir = $abs_dir;
- my $base_dir;
- my $filename;
- my %overview;
- my $lines_found;
- my $lines_hit;
- my $overall_found=0;
- my $overall_hit=0;
- my $base_name;
- my $extension;
- my $testdata;
- my %testhash;
- local *HTML_HANDLE;
-
- # Remove prefix if applicable
- if (!$no_prefix)
- {
- # Match directory name beginning with $dir_prefix
- $rel_dir = apply_prefix($rel_dir, $dir_prefix);
- }
-
- $trunc_dir = $rel_dir;
-
- # Remove leading /
- if ($rel_dir =~ /^\/(.*)$/)
- {
- $rel_dir = substr($rel_dir, 1);
- }
-
- $base_dir = get_relative_base_path($rel_dir);
-
- $abs_dir = escape_regexp($abs_dir);
-
- # Match filenames which specify files in this directory, not including
- # sub-directories
- foreach $filename (grep(/^$abs_dir\/[^\/]*$/,keys(%info_data)))
- {
- ($lines_found, $lines_hit, $testdata) =
- process_file($trunc_dir, $rel_dir, $filename);
-
- $base_name = basename($filename);
-
- $overview{$base_name} = "$lines_found,$lines_hit";
-
- $testhash{$base_name} = $testdata;
-
- $overall_found += $lines_found;
- $overall_hit += $lines_hit;
- }
- write_file_table($abs_dir, "./linux/", \%overview, \%testhash, 4);
-
-
- # Calculate resulting line counts
- return ($overall_found, $overall_hit);
-}
-
-
-#
-# process_file(trunc_dir, rel_dir, filename)
-#
-
-sub process_file($$$)
-{
- info("Processing file ".apply_prefix($_[2], $dir_prefix)."\n");
- my $trunc_dir = $_[0];
- my $rel_dir = $_[1];
- my $filename = $_[2];
- my $base_name = basename($filename);
- my $base_dir = get_relative_base_path($rel_dir);
- my $testdata;
- my $testcount;
- my $sumcount;
- my $funcdata;
- my $lines_found;
- my $lines_hit;
- my @source;
- my $pagetitle;
-
- ($testdata, $sumcount, $funcdata, $lines_found, $lines_hit) =
- get_info_entry($info_data{$filename});
- return ($lines_found, $lines_hit, $testdata);
-}
-
-
-#
-# read_info_file(info_filename)
-#
-# Read in the contents of the .info file specified by INFO_FILENAME. Data will
-# be returned as a reference to a hash containing the following mappings:
-#
-# %result: for each filename found in file -> \%data
-#
-# %data: "test" -> \%testdata
-# "sum" -> \%sumcount
-# "func" -> \%funcdata
-# "found" -> $lines_found (number of instrumented lines found in file)
-# "hit" -> $lines_hit (number of executed lines in file)
-#
-# %testdata: name of test affecting this file -> \%testcount
-#
-# %testcount: line number -> execution count for a single test
-# %sumcount : line number -> execution count for all tests
-# %funcdata : line number -> name of function beginning at that line
-#
-# Note that .info file sections referring to the same file and test name
-# will automatically be combined by adding all execution counts.
-#
-# Note that if INFO_FILENAME ends with ".gz", it is assumed that the file
-# is compressed using GZIP. If available, GUNZIP will be used to decompress
-# this file.
-#
-# Die on error
-#
-
-sub read_info_file($)
-{
- my $tracefile = $_[0]; # Name of tracefile
- my %result; # Resulting hash: file -> data
- my $data; # Data handle for current entry
- my $testdata; # " "
- my $testcount; # " "
- my $sumcount; # " "
- my $funcdata; # " "
- my $line; # Current line read from .info file
- my $testname; # Current test name
- my $filename; # Current filename
- my $hitcount; # Count for lines hit
- my $count; # Execution count of current line
- my $negative; # If set, warn about negative counts
- local *INFO_HANDLE; # Filehandle for .info file
-
- # Check if file exists and is readable
- stat($_[0]);
- if (!(-r _))
- {
- die("ERROR: cannot read file $_[0]!\n");
- }
-
- # Check if this is really a plain file
- if (!(-f _))
- {
- die("ERROR: not a plain file: $_[0]!\n");
- }
-
- # Check for .gz extension
- if ($_[0] =~ /^(.*)\.gz$/)
- {
- # Check for availability of GZIP tool
- system("gunzip -h >/dev/null 2>/dev/null")
- and die("ERROR: gunzip command not available!\n");
-
- # Check integrity of compressed file
- system("gunzip -t $_[0] >/dev/null 2>/dev/null")
- and die("ERROR: integrity check failed for ".
- "compressed file $_[0]!\n");
-
- # Open compressed file
- open(INFO_HANDLE, "gunzip -c $_[0]|")
- or die("ERROR: cannot start gunzip to uncompress ".
- "file $_[0]!\n");
- }
- else
- {
- # Open uncompressed file
- open(INFO_HANDLE, $_[0])
- or die("ERROR: cannot read file $_[0]!\n");
- }
-
- $testname = "";
- while (<INFO_HANDLE>)
- {
- chomp($_);
- $line = $_;
-
- # Switch statement
- foreach ($line)
- {
- /^TN:(\w+)/ && do
- {
- # Test name information found
- $testname = $1;
- last;
- };
-
- /^[SK]F:(.*)/ && do
- {
- # Filename information found
- # Retrieve data for new entry
- $filename = $1;
-
- $data = $result{$filename};
- ($testdata, $sumcount, $funcdata) =
- get_info_entry($data);
-
- if (defined($testname))
- {
- $testcount = $testdata->{$testname};
- }
- else
- {
- my %new_hash;
- $testcount = \%new_hash;
- }
- last;
- };
-
- /^DA:(\d+),(-?\d+)/ && do
- {
- # Fix negative counts
- $count = $2 < 0 ? 0 : $2;
- if ($2 < 0)
- {
- $negative = 1;
- }
- # Execution count found, add to structure
- # Add summary counts
- $sumcount->{$1} += $count;
-
- # Add test-specific counts
- if (defined($testname))
- {
- $testcount->{$1} += $count;
- }
- last;
- };
-
- /^FN:(\d+),([^,]+)/ && do
- {
- # Function data found, add to structure
- $funcdata->{$1} = $2;
- last;
- };
-
- /^end_of_record/ && do
- {
- # Found end of section marker
- if ($filename)
- {
- # Store current section data
- if (defined($testname))
- {
- $testdata->{$testname} =
- $testcount;
- }
- set_info_entry($data, $testdata,
- $sumcount, $funcdata);
- $result{$filename} = $data;
- }
-
- };
-
- # default
- last;
- }
- }
- close(INFO_HANDLE);
-
- # Calculate lines_found and lines_hit for each file
- foreach $filename (keys(%result))
- {
- $data = $result{$filename};
-
- ($testdata, $sumcount, $funcdata) = get_info_entry($data);
-
- $data->{"found"} = scalar(keys(%{$sumcount}));
- $hitcount = 0;
-
- foreach (keys(%{$sumcount}))
- {
- if ($sumcount->{$_} >0) { $hitcount++; }
- }
-
- $data->{"hit"} = $hitcount;
-
- $result{$filename} = $data;
- }
-
- if (scalar(keys(%result)) == 0)
- {
- die("ERROR: No valid records found in tracefile $tracefile\n");
- }
- if ($negative)
- {
- warn("WARNING: Negative counts found in tracefile ".
- "$tracefile\n");
- }
-
- return(\%result);
-}
-
-
-#
-# get_info_entry(hash_ref)
-#
-# Retrieve data from an entry of the structure generated by read_info_file().
-# Return a list of references to hashes:
-# (test data hash ref, sum count hash ref, funcdata hash ref, lines found,
-# lines hit)
-#
-
-sub get_info_entry($)
-{
- my $testdata_ref = $_[0]->{"test"};
- my $sumcount_ref = $_[0]->{"sum"};
- my $funcdata_ref = $_[0]->{"func"};
- my $lines_found = $_[0]->{"found"};
- my $lines_hit = $_[0]->{"hit"};
-
- return ($testdata_ref, $sumcount_ref, $funcdata_ref, $lines_found,
- $lines_hit);
-}
-
-
-#
-# set_info_entry(hash_ref, testdata_ref, sumcount_ref, funcdata_ref[,
-# lines_found, lines_hit])
-#
-# Update the hash referenced by HASH_REF with the provided data references.
-#
-
-sub set_info_entry($$$$;$$)
-{
- my $data_ref = $_[0];
-
- $data_ref->{"test"} = $_[1];
- $data_ref->{"sum"} = $_[2];
- $data_ref->{"func"} = $_[3];
-
- if (defined($_[4])) { $data_ref->{"found"} = $_[4]; }
- if (defined($_[5])) { $data_ref->{"hit"} = $_[5]; }
-}
-
-
-#
-# get_prefix(filename_list)
-#
-# Search FILENAME_LIST for a directory prefix which is common to as many
-# list entries as possible, so that removing this prefix will minimize the
-# sum of the lengths of all resulting shortened filenames.
-#
-
-sub get_prefix(@)
-{
- my @filename_list = @_; # provided list of filenames
- my %prefix; # mapping: prefix -> sum of lengths
- my $current; # Temporary iteration variable
-
- # Find list of prefixes
- foreach (@filename_list)
- {
- # Need explicit assignment to get a copy of $_ so that
- # shortening the contained prefix does not affect the list
- $current = shorten_prefix($_);
- while ($current = shorten_prefix($current))
- {
- # Skip rest if the remaining prefix has already been
- # added to hash
- if ($prefix{$current}) { last; }
-
- # Initialize with 0
- $prefix{$current}="0";
- }
-
- }
-
- # Calculate sum of lengths for all prefixes
- foreach $current (keys(%prefix))
- {
- foreach (@filename_list)
- {
- # Add original length
- $prefix{$current} += length($_);
-
- # Check whether prefix matches
- if (substr($_, 0, length($current)) eq $current)
- {
- # Subtract prefix length for this filename
- $prefix{$current} -= length($current);
- }
- }
- }
-
- # Find and return prefix with minimal sum
- $current = (keys(%prefix))[0];
-
- foreach (keys(%prefix))
- {
- if ($prefix{$_} < $prefix{$current})
- {
- $current = $_;
- }
- }
-
- return($current);
-}
-
-
-#
-# shorten_prefix(prefix)
-#
-# Return PREFIX shortened by last directory component.
-#
-
-sub shorten_prefix($)
-{
- my @list = split("/", $_[0]);
-
- pop(@list);
- return join("/", @list);
-}
-
-
-
-#
-# get_dir_list(filename_list)
-#
-# Return sorted list of directories for each entry in given FILENAME_LIST.
-#
-
-sub get_dir_list(@)
-{
- my %result;
-
- foreach (@_)
- {
- $result{shorten_prefix($_)} = "";
- }
-
- return(sort(keys(%result)));
-}
-
-
-#
-# get_relative_base_path(subdirectory)
-#
-# Return a relative path string which references the base path when applied
-# in SUBDIRECTORY.
-#
-# Example: get_relative_base_path("fs/mm") -> "../../"
-#
-
-sub get_relative_base_path($)
-{
- my $result = "";
- my $index;
-
- # Make an empty directory path a special case
- if (!$_[0]) { return(""); }
-
- # Count number of /s in path
- $index = ($_[0] =~ s/\//\//g);
-
- # Add a ../ to $result for each / in the directory path + 1
- for (; $index>=0; $index--)
- {
- $result .= "../";
- }
-
- return $result;
-}
-
-
-#
-# get_date_string()
-#
-# Return the current date in the form: yyyy-mm-dd
-#
-
-sub get_date_string()
-{
- my $year;
- my $month;
- my $day;
-
- ($year, $month, $day) = (localtime())[5, 4, 3];
-
- return sprintf("%d-%02d-%02d", $year+1900, $month+1, $day);
-}
-
-
-#
-# split_filename(filename)
-#
-# Return (path, filename, extension) for a given FILENAME.
-#
-
-sub split_filename($)
-{
- if (!$_[0]) { return(); }
- my @path_components = split('/', $_[0]);
- my @file_components = split('\.', pop(@path_components));
- my $extension = pop(@file_components);
-
- return (join("/",@path_components), join(".",@file_components),
- $extension);
-}
-
-
-#
-# write_file_table(filehandle, base_dir, overview, testhash, fileview)
-#
-# Write a complete file table. OVERVIEW is a reference to a hash containing
-# the following mapping:
-#
-# filename -> "lines_found,lines_hit,page_link"
-#
-# TESTHASH is a reference to the following hash:
-#
-# filename -> \%testdata
-# %testdata: name of test affecting this file -> \%testcount
-# %testcount: line number -> execution count for a single test
-#
-# Heading of first column is "Filename" if FILEVIEW is true, "Directory name"
-# otherwise.
-#
-
-sub write_file_table(*$$$$)
-{
- my $dir = $_[0];
- my $base_dir = $_[1];
- my %overview = %{$_[2]};
- my %testhash = %{$_[3]};
- my $fileview = $_[4];
- my $filename;
- my $hit;
- my $found;
- my $classification;
- my $rate_string;
- my $rate;
- my $junk;
-
-
- foreach $filename (sort(keys(%overview)))
- {
- ($found, $hit, $junk) = split(",", $overview{$filename});
-#James I think this is right
- $rate = $hit * 100 / $found;
- $rate_string = sprintf("%.1f", $rate);
-
- if ($rate < 0.001) { $classification = "None"; }
- elsif ($rate < $med_limit) { $classification = "Lo"; }
- elsif ($rate < $hi_limit) { $classification = "Med"; }
- else { $classification = "Hi"; }
-
- print "$dir/$filename\t$classification\t$rate_string\n";
-
- }
-}
-
-
-#
-# info(printf_parameter)
-#
-# Use printf to write PRINTF_PARAMETER to stdout only when the $quiet flag
-# is not set.
-#
-
-sub info(@)
-{
- if (!$quiet)
- {
- # Print info string
- printf(STDERR @_);
- }
-}
-
-
-#
-# subtract_counts(data_ref, base_ref)
-#
-
-sub subtract_counts($$)
-{
- my %data = %{$_[0]};
- my %base = %{$_[1]};
- my $line;
- my $data_count;
- my $base_count;
- my $hit = 0;
- my $found = 0;
-
- foreach $line (keys(%data))
- {
- $found++;
- $data_count = $data{$line};
- $base_count = $base{$line};
-
- if (defined($base_count))
- {
- $data_count -= $base_count;
-
- # Make sure we don't get negative numbers
- if ($data_count<0) { $data_count = 0; }
- }
-
- $data{$line} = $data_count;
- if ($data_count > 0) { $hit++; }
- }
-
- return (\%data, $found, $hit);
-}
-
-
-#
-# add_counts(data1_ref, data2_ref)
-#
-# DATA1_REF and DATA2_REF are references to hashes containing a mapping
-#
-# line number -> execution count
-#
-# Return a list (RESULT_REF, LINES_FOUND, LINES_HIT) where RESULT_REF
-# is a reference to a hash containing the combined mapping in which
-# execution counts are added.
-#
-
-sub add_counts($$)
-{
- my %data1 = %{$_[0]}; # Hash 1
- my %data2 = %{$_[1]}; # Hash 2
- my %result; # Resulting hash
- my $line; # Current line iteration scalar
- my $data1_count; # Count of line in hash1
- my $data2_count; # Count of line in hash2
- my $found = 0; # Total number of lines found
- my $hit = 0; # Number of lines with a count > 0
-
- foreach $line (keys(%data1))
- {
- $data1_count = $data1{$line};
- $data2_count = $data2{$line};
-
- # Add counts if present in both hashes
- if (defined($data2_count)) { $data1_count += $data2_count; }
-
- # Store sum in %result
- $result{$line} = $data1_count;
-
- $found++;
- if ($data1_count > 0) { $hit++; }
- }
-
- # Add lines unique to data2
- foreach $line (keys(%data2))
- {
- # Skip lines already in data1
- if (defined($data1{$line})) { next; }
-
- # Copy count from data2
- $result{$line} = $data2{$line};
-
- $found++;
- if ($result{$line} > 0) { $hit++; }
- }
-
- return (\%result, $found, $hit);
-}
-
-
-#
-# apply_baseline(data_ref, baseline_ref)
-#
-# Subtract the execution counts found in the baseline hash referenced by
-# BASELINE_REF from actual data in DATA_REF.
-#
-
-sub apply_baseline($$)
-{
- my %data_hash = %{$_[0]};
- my %base_hash = %{$_[1]};
- my $filename;
- my $testname;
- my $data;
- my $data_testdata;
- my $data_funcdata;
- my $data_count;
- my $base;
- my $base_testdata;
- my $base_count;
- my $sumcount;
- my $found;
- my $hit;
-
- foreach $filename (keys(%data_hash))
- {
- # Get data set for data and baseline
- $data = $data_hash{$filename};
- $base = $base_hash{$filename};
-
- # Get set entries for data and baseline
- ($data_testdata, undef, $data_funcdata) =
- get_info_entry($data);
- ($base_testdata, $base_count) = get_info_entry($base);
-
- # Sumcount has to be calculated anew
- $sumcount = {};
-
- # For each test case, subtract test specific counts
- foreach $testname (keys(%{$data_testdata}))
- {
- # Get counts of both data and baseline
- $data_count = $data_testdata->{$testname};
-
- $hit = 0;
-
- ($data_count, undef, $hit) =
- subtract_counts($data_count, $base_count);
-
- # Check whether this test case did hit any line at all
- if ($hit > 0)
- {
- # Write back resulting hash
- $data_testdata->{$testname} = $data_count;
- }
- else
- {
- # Delete test case which did not impact this
- # file
- delete($data_testdata->{$testname});
- }
-
- # Add counts to sum of counts
- ($sumcount, $found, $hit) =
- add_counts($sumcount, $data_count);
- }
-
- # Write back resulting entry
- set_info_entry($data, $data_testdata, $sumcount,
- $data_funcdata, $found, $hit);
-
- $data_hash{$filename} = $data;
- }
-
- return (\%data_hash);
-}
-
-
-#
-# combine_info_entries(entry_ref1, entry_ref2)
-#
-# Combine .info data entry hashes referenced by ENTRY_REF1 and ENTRY_REF2.
-# Return reference to resulting hash.
-#
-
-sub combine_info_entries($$)
-{
- my $entry1 = $_[0]; # Reference to hash containing first entry
- my $testdata1;
- my $sumcount1;
- my $funcdata1;
-
- my $entry2 = $_[1]; # Reference to hash containing second entry
- my $testdata2;
- my $sumcount2;
- my $funcdata2;
-
- my %result; # Hash containing combined entry
- my %result_testdata;
- my $result_sumcount = {};
- my %result_funcdata;
- my $lines_found;
- my $lines_hit;
-
- my $testname;
-
- # Retrieve data
- ($testdata1, $sumcount1, $funcdata1) = get_info_entry($entry1);
- ($testdata2, $sumcount2, $funcdata2) = get_info_entry($entry2);
-
- # Combine funcdata
- foreach (keys(%{$funcdata1}))
- {
- $result_funcdata{$_} = $funcdata1->{$_};
- }
-
- foreach (keys(%{$funcdata2}))
- {
- $result_funcdata{$_} = $funcdata2->{$_};
- }
-
- # Combine testdata
- foreach $testname (keys(%{$testdata1}))
- {
- if (defined($testdata2->{$testname}))
- {
- # testname is present in both entries, requires
- # combination
- ($result_testdata{$testname}) =
- add_counts($testdata1->{$testname},
- $testdata2->{$testname});
- }
- else
- {
- # testname only present in entry1, add to result
- $result_testdata{$testname} = $testdata1->{$testname};
- }
-
- # update sum count hash
- ($result_sumcount, $lines_found, $lines_hit) =
- add_counts($result_sumcount,
- $result_testdata{$testname});
- }
-
- foreach $testname (keys(%{$testdata2}))
- {
- # Skip testnames already covered by previous iteration
- if (defined($testdata1->{$testname})) { next; }
-
- # testname only present in entry2, add to result hash
- $result_testdata{$testname} = $testdata2->{$testname};
-
- # update sum count hash
- ($result_sumcount, $lines_found, $lines_hit) =
- add_counts($result_sumcount,
- $result_testdata{$testname});
- }
-
- # Calculate resulting sumcount
-
- # Store result
- set_info_entry(\%result, \%result_testdata, $result_sumcount,
- \%result_funcdata, $lines_found, $lines_hit);
-
- return(\%result);
-}
-
-
-#
-# combine_info_files(info_ref1, info_ref2)
-#
-# Combine .info data in hashes referenced by INFO_REF1 and INFO_REF2. Return
-# reference to resulting hash.
-#
-
-sub combine_info_files($$)
-{
- my %hash1 = %{$_[0]};
- my %hash2 = %{$_[1]};
- my $filename;
-
- foreach $filename (keys(%hash2))
- {
- if ($hash1{$filename})
- {
- # Entry already exists in hash1, combine them
- $hash1{$filename} =
- combine_info_entries($hash1{$filename},
- $hash2{$filename});
- }
- else
- {
- # Entry is unique in both hashes, simply add to
- # resulting hash
- $hash1{$filename} = $hash2{$filename};
- }
- }
-
- return(\%hash1);
-}
-
-
-#
-# apply_prefix(filename, prefix)
-#
-# If FILENAME begins with PREFIX, remove PREFIX from FILENAME and return
-# resulting string, otherwise return FILENAME.
-#
-
-sub apply_prefix($$)
-{
- my $filename = $_[0];
- my $prefix = $_[1];
- my $clean_prefix = escape_regexp($prefix);
-
- if (defined($prefix) && ($prefix ne ""))
- {
- if ($filename =~ /^$clean_prefix\/(.*)$/)
- {
- return substr($filename, length($prefix) + 1);
- }
- }
-
- return $filename;
-}
-
-
-#
-# escape_regexp(string)
-#
-# Escape special characters in STRING which would be incorrectly interpreted
-# in a PERL regular expression.
-#
-
-sub escape_regexp($)
-{
- my $string = $_[0];
-
- # Escape special characters
- $string =~ s/\\/\\\\/g;
- $string =~ s/\^/\\\^/g;
- $string =~ s/\$/\\\$/g;
- $string =~ s/\./\\\./g;
- $string =~ s/\|/\\\|/g;
- $string =~ s/\(/\\\(/g;
- $string =~ s/\)/\\\)/g;
- $string =~ s/\[/\\\[/g;
- $string =~ s/\]/\\\]/g;
- $string =~ s/\*/\\\*/g;
- $string =~ s/\?/\\\?/g;
- $string =~ s/\{/\\\{/g;
- $string =~ s/\}/\\\}/g;
- $string =~ s/\+/\\\+/g;
-
- return $string;
-}
« no previous file with comments | « third_party/lcov-1.9/contrib/galaxy/gen_makefile.sh ('k') | third_party/lcov-1.9/contrib/galaxy/posterize.pl » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698