wsdiff.py revision 598cc7dfe526c1fa47f03934c5c6fa6508961296
#!/usr/bin/python2.6
#
# CDDL HEADER START
#
# The contents of this file are subject to the terms of the
# Common Development and Distribution License (the "License").
# You may not use this file except in compliance with the License.
#
# You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
# or http://www.opensolaris.org/os/licensing.
# See the License for the specific language governing permissions
# and limitations under the License.
#
# When distributing Covered Code, include this CDDL HEADER in each
# file and include the License file at usr/src/OPENSOLARIS.LICENSE.
# If applicable, add the following below this CDDL HEADER, with the
# fields enclosed by brackets "[]" replaced with your own identifying
# information: Portions Copyright [yyyy] [name of copyright owner]
#
# CDDL HEADER END
#
# Copyright (c) 2010, Oracle and/or its affiliates. All rights reserved.
#
#
# wsdiff(1) is a tool that can be used to determine which compiled objects
# have changed as a result of a given source change. Developers backporting
# new features, RFEs and bug fixes need to be able to identify the set of
# patch deliverables necessary for feature/fix realization on a patched system.
#
# The tool works by comparing objects in two trees/proto areas (one build with,
# and without the source changes.
#
# Using wsdiff(1) is fairly simple:
# - Bringover to a fresh workspace
# - Perform a full non-debug build (clobber if workspace isn't fresh)
# - Move the proto area aside, call it proto.old, or something.
# - Integrate your changes to the workspace
# - Perform another full non-debug clobber build.
# - Use wsdiff(1) to see what changed:
# $ wsdiff proto.old proto
#
# By default, wsdiff will print the list of changed objects / deliverables to
# stdout. If a results file is specified via -r, the list of differing objects,
# and details about why wsdiff(1) thinks they are different will be logged to
# the results file.
#
# By invoking nightly(1) with the -w option to NIGHTLY_FLAGS, nightly(1) will use
# wsdiff(1) to report on what objects changed since the last build.
#
# For patch deliverable purposes, it's advised to have nightly do a clobber,
# non-debug build.
#
# Think about the results. Was something flagged that you don't expect? Go look
# at the results file to see details about the differences.
#
# Use the -i option in conjunction with -v and -V to dive deeper and have wsdiff(1)
# report with more verbosity.
#
# Usage: wsdiff [-vVt] [-r results ] [-i filelist ] old new
#
# Where "old" is the path to the proto area build without the changes, and
# "new" is the path to the proto area built with the changes. The following
# options are supported:
#
# -v Do not truncate observed diffs in results
# -V Log *all* ELF sect diffs vs. logging the first diff found
# -t Use onbld tools in $SRC/tools
# -r Log results and observed differences
# -i Tell wsdiff which objects to compare via an input file list
import datetime, fnmatch, getopt, os, profile, commands
import re, resource, select, shutil, signal, string, struct, sys, tempfile
import time, threading
from stat import *
# Human readable diffs truncated by default if longer than this
# Specifying -v on the command line will override
diffs_sz_thresh = 4096
# Lock name Provides exclusive access to
# --------------+------------------------------------------------
# output_lock standard output or temporary file (difference())
# log_lock the results file (log_difference())
# wset_lock changedFiles list (workerThread())
output_lock = threading.Lock()
log_lock = threading.Lock()
wset_lock = threading.Lock()
# Variable for thread control
keep_processing = True
# Default search path for wsdiff
wsdiff_path = [ "/usr/bin",
"/usr/ccs/bin",
"/lib/svc/bin",
"/opt/onbld/bin" ]
# These are objects that wsdiff will notice look different, but will not report.
# Existence of an exceptions list, and adding things here is *dangerous*,
# and therefore the *only* reasons why anything would be listed here is because
# the objects do not build deterministically, yet we *cannot* fix this.
#
# These perl libraries use __DATE__ and therefore always look different.
# Ideally, we would purge use the use of __DATE__ from the source, but because
# this is source we wish to distribute with Solaris "unchanged", we cannot modify.
#
wsdiff_exceptions = [ "usr/perl5/5.8.4/lib/sun4-solaris-64int/CORE/libperl.so.1",
"usr/perl5/5.6.1/lib/sun4-solaris-64int/CORE/libperl.so.1",
"usr/perl5/5.8.4/lib/i86pc-solaris-64int/CORE/libperl.so.1",
"usr/perl5/5.6.1/lib/i86pc-solaris-64int/CORE/libperl.so.1"
]
#####
# Logging routines
#
# Debug message to be printed to the screen, and the log file
def debug(msg) :
# Add prefix to highlight debugging message
msg = "## " + msg
if debugon :
output_lock.acquire()
print >> sys.stdout, msg
sys.stdout.flush()
output_lock.release()
if logging :
log_lock.acquire()
print >> log, msg
log.flush()
log_lock.release()
# Informational message to be printed to the screen, and the log file
def info(msg) :
output_lock.acquire()
print >> sys.stdout, msg
sys.stdout.flush()
output_lock.release()
if logging :
log_lock.acquire()
print >> log, msg
log.flush()
log_lock.release()
# Error message to be printed to the screen, and the log file
def error(msg) :
output_lock.acquire()
print >> sys.stderr, "ERROR:", msg
sys.stderr.flush()
output_lock.release()
if logging :
log_lock.acquire()
print >> log, "ERROR:", msg
log.flush()
log_lock.release()
# Informational message to be printed only to the log, if there is one.
def v_info(msg) :
if logging :
log_lock.acquire()
print >> log, msg
log.flush()
log_lock.release()
#
# Flag a detected file difference
# Display the fileName to stdout, and log the difference
#
def difference(f, dtype, diffs) :
if f in wsdiff_exceptions :
return
output_lock.acquire()
if sorted :
differentFiles.append(f)
else:
print >> sys.stdout, f
sys.stdout.flush()
output_lock.release()
log_difference(f, dtype, diffs)
#
# Do the actual logging of the difference to the results file
#
def log_difference(f, dtype, diffs) :
if logging :
log_lock.acquire()
print >> log, f
print >> log, "NOTE:", dtype, "difference detected."
difflen = len(diffs)
if difflen > 0 :
print >> log
if not vdiffs and difflen > diffs_sz_thresh :
print >> log, diffs[:diffs_sz_thresh]
print >> log, \
"... truncated due to length: " \
"use -v to override ..."
else :
print >> log, diffs
print >> log, "\n"
log.flush()
log_lock.release()
#####
# diff generating routines
#
#
# Return human readable diffs from two temporary files
#
def diffFileData(tmpf1, tmpf2) :
binaries = False
# Filter the data through od(1) if the data is detected
# as being binary
if isBinary(tmpf1) or isBinary(tmpf2) :
binaries = True
tmp_od1 = tmpf1 + ".od"
tmp_od2 = tmpf2 + ".od"
cmd = od_cmd + " -c -t x4" + " " + tmpf1 + " > " + tmp_od1
os.system(cmd)
cmd = od_cmd + " -c -t x4" + " " + tmpf2 + " > " + tmp_od2
os.system(cmd)
tmpf1 = tmp_od1
tmpf2 = tmp_od2
try:
data = commands.getoutput(diff_cmd + " " + tmpf1 + " " + tmpf2)
# Remove the temp files as we no longer need them.
if binaries :
try:
os.unlink(tmp_od1)
except OSError, e:
error("diffFileData: unlink failed %s" % e)
try:
os.unlink(tmp_od2)
except OSError, e:
error("diffFileData: unlink failed %s" % e)
except:
error("failed to get output of command: " + diff_cmd + " " \
+ tmpf1 + " " + tmpf2)
# Send exception for the failed command up
raise
return
return data
#
# Return human readable diffs betweeen two datasets
#
def diffData(base, ptch, d1, d2) :
t = threading.currentThread()
tmpFile1 = tmpDir1 + os.path.basename(base) + t.getName()
tmpFile2 = tmpDir2 + os.path.basename(ptch) + t.getName()
try:
fd1 = open(tmpFile1, "w")
except:
error("failed to open: " + tmpFile1)
cleanup(1)
try:
fd2 = open(tmpFile2, "w")
except:
error("failed to open: " + tmpFile2)
cleanup(1)
fd1.write(d1)
fd2.write(d2)
fd1.close()
fd2.close()
return diffFileData(tmpFile1, tmpFile2)
#####
# Misc utility functions
#
# Prune off the leading prefix from string s
def str_prefix_trunc(s, prefix) :
snipLen = len(prefix)
return s[snipLen:]
#
# Prune off leading proto path goo (if there is one) to yield
# the deliverable's eventual path relative to root
# e.g. proto.base/root_sparc/usr/src/cmd/prstat => usr/src/cmd/prstat
#
def fnFormat(fn) :
root_arch_str = "root_" + arch
pos = fn.find(root_arch_str)
if pos == -1 :
return fn
pos = fn.find("/", pos)
if pos == -1 :
return fn
return fn[pos + 1:]
#####
# Usage / argument processing
#
#
# Display usage message
#
def usage() :
sys.stdout.flush()
print >> sys.stderr, """Usage: wsdiff [-dvVst] [-r results ] [-i filelist ] old new
-d Print debug messages about the progress
-v Do not truncate observed diffs in results
-V Log *all* ELF sect diffs vs. logging the first diff found
-t Use onbld tools in $SRC/tools
-r Log results and observed differences
-s Produce sorted list of differences
-i Tell wsdiff which objects to compare via an input file list"""
sys.exit(1)
#
# Process command line options
#
def args() :
global debugon
global logging
global vdiffs
global reportAllSects
global sorted
validOpts = 'di:r:vVst?'
baseRoot = ""
ptchRoot = ""
fileNamesFile = ""
results = ""
localTools = False
# getopt.getopt() returns:
# an option/value tuple
# a list of remaining non-option arguments
#
# A correct wsdiff invocation will have exactly two non option
# arguments, the paths to the base (old), ptch (new) proto areas
try:
optlist, args = getopt.getopt(sys.argv[1:], validOpts)
except getopt.error, val:
usage()
if len(args) != 2 :
usage();
for opt,val in optlist :
if opt == '-d' :
debugon = True
elif opt == '-i' :
fileNamesFile = val
elif opt == '-r' :
results = val
logging = True
elif opt == '-s' :
sorted = True
elif opt == '-v' :
vdiffs = True
elif opt == '-V' :
reportAllSects = True
elif opt == '-t':
localTools = True
else:
usage()
baseRoot = args[0]
ptchRoot = args[1]
if len(baseRoot) == 0 or len(ptchRoot) == 0 :
usage()
if logging and len(results) == 0 :
usage()
if vdiffs and not logging :
error("The -v option requires a results file (-r)")
sys.exit(1)
if reportAllSects and not logging :
error("The -V option requires a results file (-r)")
sys.exit(1)
# alphabetical order
return baseRoot, fileNamesFile, localTools, ptchRoot, results
#####
# File identification
#
#
# Identify the file type.
# If it's not ELF, use the file extension to identify
# certain file types that require special handling to
# compare. Otherwise just return a basic "ASCII" type.
#
def getTheFileType(f) :
extensions = { 'a' : 'ELF Object Archive',
'jar' : 'Java Archive',
'html' : 'HTML',
'ln' : 'Lint Library',
'db' : 'Sqlite Database' }
try:
if os.stat(f)[ST_SIZE] == 0 :
return 'ASCII'
except:
error("failed to stat " + f)
return 'Error'
if isELF(f) == 1 :
return 'ELF'
fnamelist = f.split('.')
if len(fnamelist) > 1 : # Test the file extension
extension = fnamelist[-1]
if extension in extensions.keys():
return extensions[extension]
return 'ASCII'
#
# Return non-zero if "f" is an ELF file
#
elfmagic = '\177ELF'
def isELF(f) :
try:
fd = open(f)
except:
error("failed to open: " + f)
return 0
magic = fd.read(len(elfmagic))
fd.close()
if magic == elfmagic :
return 1
return 0
#
# Return non-zero is "f" is binary.
# Consider the file to be binary if it contains any null characters
#
def isBinary(f) :
try:
fd = open(f)
except:
error("failed to open: " + f)
return 0
s = fd.read()
fd.close()
if s.find('\0') == -1 :
return 0
else :
return 1
#####
# Directory traversal and file finding
#
#
# Return a sorted list of files found under the specified directory
#
def findFiles(d) :
for path, subdirs, files in os.walk(d) :
files.sort()
for name in files :
yield os.path.join(path, name)
#
# Examine all files in base, ptch
#
# Return a list of files appearing in both proto areas,
# a list of new files (files found only in ptch) and
# a list of deleted files (files found only in base)
#
def protoCatalog(base, ptch) :
compFiles = [] # List of files in both proto areas
ptchList = [] # List of file in patch proto area
newFiles = [] # New files detected
deletedFiles = [] # Deleted files
debug("Getting the list of files in the base area");
baseFilesList = list(findFiles(base))
baseStringLength = len(base)
debug("Found " + str(len(baseFilesList)) + " files")
debug("Getting the list of files in the patch area");
ptchFilesList = list(findFiles(ptch))
ptchStringLength = len(ptch)
debug("Found " + str(len(ptchFilesList)) + " files")
# Inventory files in the base proto area
debug("Determining the list of regular files in the base area");
for fn in baseFilesList :
if os.path.islink(fn) :
continue
fileName = fn[baseStringLength:]
compFiles.append(fileName)
debug("Found " + str(len(compFiles)) + " files")
# Inventory files in the patch proto area
debug("Determining the list of regular files in the patch area");
for fn in ptchFilesList :
if os.path.islink(fn) :
continue
fileName = fn[ptchStringLength:]
ptchList.append(fileName)
debug("Found " + str(len(ptchList)) + " files")
# Deleted files appear in the base area, but not the patch area
debug("Searching for deleted files by comparing the lists")
for fileName in compFiles :
if not fileName in ptchList :
deletedFiles.append(fileName)
debug("Found " + str(len(deletedFiles)) + " deleted files")
# Eliminate "deleted" files from the list of objects appearing
# in both the base and patch proto areas
debug("Eliminating deleted files from the list of objects")
for fileName in deletedFiles :
try:
compFiles.remove(fileName)
except:
error("filelist.remove() failed")
debug("List for comparison reduced to " + str(len(compFiles)) \
+ " files")
# New files appear in the patch area, but not the base
debug("Getting the list of newly added files")
for fileName in ptchList :
if not fileName in compFiles :
newFiles.append(fileName)
debug("Found " + str(len(newFiles)) + " new files")
return compFiles, newFiles, deletedFiles
#
# Examine the files listed in the input file list
#
# Return a list of files appearing in both proto areas,
# a list of new files (files found only in ptch) and
# a list of deleted files (files found only in base)
#
def flistCatalog(base, ptch, flist) :
compFiles = [] # List of files in both proto areas
newFiles = [] # New files detected
deletedFiles = [] # Deleted files
try:
fd = open(flist, "r")
except:
error("could not open: " + flist)
cleanup(1)
files = []
files = fd.readlines()
fd.close()
for f in files :
ptch_present = True
base_present = True
if f == '\n' :
continue
# the fileNames have a trailing '\n'
f = f.rstrip()
# The objects in the file list have paths relative
# to $ROOT or to the base/ptch directory specified on
# the command line.
# If it's relative to $ROOT, we'll need to add back the
# root_`uname -p` goo we stripped off in fnFormat()
if os.path.exists(base + f) :
fn = f;
elif os.path.exists(base + "root_" + arch + "/" + f) :
fn = "root_" + arch + "/" + f
else :
base_present = False
if base_present :
if not os.path.exists(ptch + fn) :
ptch_present = False
else :
if os.path.exists(ptch + f) :
fn = f
elif os.path.exists(ptch + "root_" + arch + "/" + f) :
fn = "root_" + arch + "/" + f
else :
ptch_present = False
if os.path.islink(base + fn) : # ignore links
base_present = False
if os.path.islink(ptch + fn) :
ptch_present = False
if base_present and ptch_present :
compFiles.append(fn)
elif base_present :
deletedFiles.append(fn)
elif ptch_present :
newFiles.append(fn)
else :
if os.path.islink(base + fn) and \
os.path.islink(ptch + fn) :
continue
error(f + " in file list, but not in either tree. " + \
"Skipping...")
return compFiles, newFiles, deletedFiles
#
# Build a fully qualified path to an external tool/utility.
# Consider the default system locations. For onbld tools, if
# the -t option was specified, we'll try to use built tools in $SRC tools,
# and otherwise, we'll fall back on /opt/onbld/
#
def find_tool(tool) :
# First, check what was passed
if os.path.exists(tool) :
return tool
# Next try in wsdiff path
for pdir in wsdiff_path :
location = pdir + "/" + tool
if os.path.exists(location) :
return location + " "
location = pdir + "/" + arch + "/" + tool
if os.path.exists(location) :
return location + " "
error("Could not find path to: " + tool);
sys.exit(1);
#####
# ELF file comparison helper routines
#
#
# Return a dictionary of ELF section types keyed by section name
#
def get_elfheader(f) :
header = {}
hstring = commands.getoutput(elfdump_cmd + " -c " + f)
if len(hstring) == 0 :
error("Failed to dump ELF header for " + f)
raise
return
# elfdump(1) dumps the section headers with the section name
# following "sh_name:", and the section type following "sh_type:"
sections = hstring.split("Section Header")
for sect in sections :
datap = sect.find("sh_name:");
if datap == -1 :
continue
section = sect[datap:].split()[1]
datap = sect.find("sh_type:");
if datap == -1 :
error("Could not get type for sect: " + section + \
" in " + f)
sh_type = sect[datap:].split()[2]
header[section] = sh_type
return header
#
# Extract data in the specified ELF section from the given file
#
def extract_elf_section(f, section) :
data = commands.getoutput(dump_cmd + " -sn " + section + " " + f)
if len(data) == 0 :
error(dump_cmd + "yielded no data on section " + section + \
" of " + f)
raise
return
# dump(1) displays the file name to start...
# get past it to the data itself
dbegin = data.find(":") + 1
data = data[dbegin:];
return (data)
#
# Return a (hopefully meaningful) human readable set of diffs
# for the specified ELF section between f1 and f2
#
# Depending on the section, various means for dumping and diffing
# the data may be employed.
#
text_sections = [ '.text', '.init', '.fini' ]
def diff_elf_section(f1, f2, section, sh_type) :
t = threading.currentThread()
tmpFile1 = tmpDir1 + os.path.basename(f1) + t.getName()
tmpFile2 = tmpDir2 + os.path.basename(f2) + t.getName()
if (sh_type == "SHT_RELA") : # sh_type == SHT_RELA
cmd1 = elfdump_cmd + " -r " + f1 + " > " + tmpFile1
cmd2 = elfdump_cmd + " -r " + f2 + " > " + tmpFile2
elif (section == ".group") :
cmd1 = elfdump_cmd + " -g " + f1 + " > " + tmpFile1
cmd2 = elfdump_cmd + " -g " + f2 + " > " + tmpFile2
elif (section == ".hash") :
cmd1 = elfdump_cmd + " -h " + f1 + " > " + tmpFile1
cmd2 = elfdump_cmd + " -h " + f2 + " > " + tmpFile2
elif (section == ".dynamic") :
cmd1 = elfdump_cmd + " -d " + f1 + " > " + tmpFile1
cmd2 = elfdump_cmd + " -d " + f2 + " > " + tmpFile2
elif (section == ".got") :
cmd1 = elfdump_cmd + " -G " + f1 + " > " + tmpFile1
cmd2 = elfdump_cmd + " -G " + f2 + " > " + tmpFile2
elif (section == ".SUNW_cap") :
cmd1 = elfdump_cmd + " -H " + f1 + " > " + tmpFile1
cmd2 = elfdump_cmd + " -H " + f2 + " > " + tmpFile2
elif (section == ".interp") :
cmd1 = elfdump_cmd + " -i " + f1 + " > " + tmpFile1
cmd2 = elfdump_cmd + " -i " + f2 + " > " + tmpFile2
elif (section == ".symtab" or section == ".dynsym") :
cmd1 = elfdump_cmd + " -s -N " + section + " " + f1 + \
" > " + tmpFile1
cmd2 = elfdump_cmd + " -s -N " + section + " " + f2 + \
" > " + tmpFile2
elif (section in text_sections) :
# dis sometimes complains when it hits something it doesn't
# know how to disassemble. Just ignore it, as the output
# being generated here is human readable, and we've already
# correctly flagged the difference.
cmd1 = dis_cmd + " -t " + section + " " + f1 + \
" 2>/dev/null | grep -v disassembly > " + tmpFile1
cmd2 = dis_cmd + " -t " + section + " " + f2 + \
" 2>/dev/null | grep -v disassembly > " + tmpFile2
else :
cmd1 = elfdump_cmd + " -w " + tmpFile1 + " -N " + \
section + " " + f1
cmd2 = elfdump_cmd + " -w " + tmpFile2 + " -N " + \
section + " " + f2
os.system(cmd1)
os.system(cmd2)
data = diffFileData(tmpFile1, tmpFile2)
# remove temp files as we no longer need them
try:
os.unlink(tmpFile1)
except OSError, e:
error("diff_elf_section: unlink failed %s" % e)
try:
os.unlink(tmpFile2)
except OSError, e:
error("diff_elf_section: unlink failed %s" % e)
return (data)
#
# compare the relevant sections of two ELF binaries
# and report any differences
#
# Returns: 1 if any differenes found
# 0 if no differences found
# -1 on error
#
# Sections deliberately not considered when comparing two ELF
# binaries. Differences observed in these sections are not considered
# significant where patch deliverable identification is concerned.
sections_to_skip = [ ".SUNW_signature",
".comment",
".SUNW_ctf",
".debug",
".plt",
".rela.bss",
".rela.plt",
".line",
".note",
".compcom",
]
sections_preferred = [ ".rodata.str1.8",
".rodata.str1.1",
".rodata",
".data1",
".data",
".text",
]
def compareElfs(base, ptch, quiet) :
global logging
try:
base_header = get_elfheader(base)
except:
return
sections = base_header.keys()
try:
ptch_header = get_elfheader(ptch)
except:
return
e2_only_sections = ptch_header.keys()
e1_only_sections = []
fileName = fnFormat(base)
# Derive the list of ELF sections found only in
# either e1 or e2.
for sect in sections :
if not sect in e2_only_sections :
e1_only_sections.append(sect)
else :
e2_only_sections.remove(sect)
if len(e1_only_sections) > 0 :
if quiet :
return 1
data = ""
if logging :
slist = ""
for sect in e1_only_sections :
slist = slist + sect + "\t"
data = "ELF sections found in " + \
base + " but not in " + ptch + \
"\n\n" + slist
difference(fileName, "ELF", data)
return 1
if len(e2_only_sections) > 0 :
if quiet :
return 1
data = ""
if logging :
slist = ""
for sect in e2_only_sections :
slist = slist + sect + "\t"
data = "ELF sections found in " + \
ptch + " but not in " + base + \
"\n\n" + slist
difference(fileName, "ELF", data)
return 1
# Look for preferred sections, and put those at the
# top of the list of sections to compare
for psect in sections_preferred :
if psect in sections :
sections.remove(psect)
sections.insert(0, psect)
# Compare ELF sections
first_section = True
for sect in sections :
if sect in sections_to_skip :
continue
try:
s1 = extract_elf_section(base, sect);
except:
return
try:
s2 = extract_elf_section(ptch, sect);
except:
return
if len(s1) != len (s2) or s1 != s2:
if not quiet:
sh_type = base_header[sect]
data = diff_elf_section(base, ptch, \
sect, sh_type)
# If all ELF sections are being reported, then
# invoke difference() to flag the file name to
# stdout only once. Any other section differences
# should be logged to the results file directly
if not first_section :
log_difference(fileName, \
"ELF " + sect, data)
else :
difference(fileName, "ELF " + sect, \
data)
if not reportAllSects :
return 1
first_section = False
return 0
#####
# recursively remove 2 directories
#
# Used for removal of temporary directory strucures (ignores any errors).
#
def clearTmpDirs(dir1, dir2) :
if os.path.isdir(dir1) > 0 :
shutil.rmtree(dir1, True)
if os.path.isdir(dir2) > 0 :
shutil.rmtree(dir2, True)
#####
# Archive object comparison
#
# Returns 1 if difference detected
# 0 if no difference detected
# -1 on error
#
def compareArchives(base, ptch, fileType) :
fileName = fnFormat(base)
t = threading.currentThread()
ArchTmpDir1 = tmpDir1 + os.path.basename(base) + t.getName()
ArchTmpDir2 = tmpDir2 + os.path.basename(base) + t.getName()
#
# Be optimistic and first try a straight file compare
# as it will allow us to finish up quickly.
#
if compareBasic(base, ptch, True, fileType) == 0 :
return 0
try:
os.makedirs(ArchTmpDir1)
except OSError, e:
error("compareArchives: makedir failed %s" % e)
return -1
try:
os.makedirs(ArchTmpDir2)
except OSError, e:
error("compareArchives: makedir failed %s" % e)
return -1
# copy over the objects to the temp areas, and
# unpack them
baseCmd = "cp -fp " + base + " " + ArchTmpDir1
status, output = commands.getstatusoutput(baseCmd)
if status != 0 :
error(baseCmd + " failed: " + output)
clearTmpDirs(ArchTmpDir1, ArchTmpDir2)
return -1
ptchCmd = "cp -fp " + ptch + " " + ArchTmpDir2
status, output = commands.getstatusoutput(ptchCmd)
if status != 0 :
error(ptchCmd + " failed: " + output)
clearTmpDirs(ArchTmpDir1, ArchTmpDir2)
return -1
bname = string.split(fileName, '/')[-1]
if fileType == "Java Archive" :
baseCmd = "cd " + ArchTmpDir1 + "; " + "jar xf " + bname + \
"; rm -f " + bname + " META-INF/MANIFEST.MF"
ptchCmd = "cd " + ArchTmpDir2 + "; " + "jar xf " + bname + \
"; rm -f " + bname + " META-INF/MANIFEST.MF"
elif fileType == "ELF Object Archive" :
baseCmd = "cd " + ArchTmpDir1 + "; " + "/usr/ccs/bin/ar x " + \
bname + "; rm -f " + bname
ptchCmd = "cd " + ArchTmpDir2 + "; " + "/usr/ccs/bin/ar x " + \
bname + "; rm -f " + bname
else :
error("unexpected file type: " + fileType)
clearTmpDirs(ArchTmpDir1, ArchTmpDir2)
return -1
os.system(baseCmd)
os.system(ptchCmd)
baseFlist = list(findFiles(ArchTmpDir1))
ptchFlist = list(findFiles(ArchTmpDir2))
# Trim leading path off base/ptch file lists
flist = []
for fn in baseFlist :
flist.append(str_prefix_trunc(fn, ArchTmpDir1))
baseFlist = flist
flist = []
for fn in ptchFlist :
flist.append(str_prefix_trunc(fn, ArchTmpDir2))
ptchFlist = flist
for fn in ptchFlist :
if not fn in baseFlist :
difference(fileName, fileType, \
fn + " added to " + fileName)
clearTmpDirs(ArchTmpDir1, ArchTmpDir2)
return 1
for fn in baseFlist :
if not fn in ptchFlist :
difference(fileName, fileType, \
fn + " removed from " + fileName)
clearTmpDirs(ArchTmpDir1, ArchTmpDir2)
return 1
differs = compareOneFile((ArchTmpDir1 + fn), \
(ArchTmpDir2 + fn), True)
if differs :
difference(fileName, fileType, \
fn + " in " + fileName + " differs")
clearTmpDirs(ArchTmpDir1, ArchTmpDir2)
return 1
clearTmpDirs(ArchTmpDir1, ArchTmpDir2)
return 0
#####
# (Basic) file comparison
#
# There's some special case code here for Javadoc HTML files
#
# Returns 1 if difference detected
# 0 if no difference detected
# -1 on error
#
def compareBasic(base, ptch, quiet, fileType) :
fileName = fnFormat(base);
if quiet and os.stat(base)[ST_SIZE] != os.stat(ptch)[ST_SIZE] :
return 1
try:
baseFile = open(base)
except:
error("could not open " + base)
return -1
try:
ptchFile = open(ptch)
except:
error("could not open " + ptch)
return -1
baseData = baseFile.read()
ptchData = ptchFile.read()
baseFile.close()
ptchFile.close()
needToSnip = False
if fileType == "HTML" :
needToSnip = True
toSnipBeginStr = "<!-- Generated by javadoc"
toSnipEndStr = "-->\n"
if needToSnip :
toSnipBegin = string.find(baseData, toSnipBeginStr)
if toSnipBegin != -1 :
toSnipEnd = string.find(baseData[toSnipBegin:], \
toSnipEndStr) + \
len(toSnipEndStr)
baseData = baseData[:toSnipBegin] + \
baseData[toSnipBegin + toSnipEnd:]
ptchData = ptchData[:toSnipBegin] + \
ptchData[toSnipBegin + toSnipEnd:]
if quiet :
if baseData != ptchData :
return 1
else :
if len(baseData) != len(ptchData) or baseData != ptchData :
diffs = diffData(base, ptch, baseData, ptchData)
difference(fileName, fileType, diffs)
return 1
return 0
#####
# Compare two objects by producing a data dump from
# each object, and then comparing the dump data
#
# Returns: 1 if a difference is detected
# 0 if no difference detected
# -1 upon error
#
def compareByDumping(base, ptch, quiet, fileType) :
fileName = fnFormat(base);
t = threading.currentThread()
tmpFile1 = tmpDir1 + os.path.basename(base) + t.getName()
tmpFile2 = tmpDir2 + os.path.basename(ptch) + t.getName()
if fileType == "Lint Library" :
baseCmd = lintdump_cmd + " -ir " + base + \
" | egrep -v '(LINTOBJ|LINTMOD):'" + \
" | grep -v PASS[1-3]:" + \
" > " + tmpFile1
ptchCmd = lintdump_cmd + " -ir " + ptch + \
" | egrep -v '(LINTOBJ|LINTMOD):'" + \
" | grep -v PASS[1-3]:" + \
" > " + tmpFile2
elif fileType == "Sqlite Database" :
baseCmd = "echo .dump | " + sqlite_cmd + base + " > " + \
tmpFile1
ptchCmd = "echo .dump | " + sqlite_cmd + ptch + " > " + \
tmpFile2
os.system(baseCmd)
os.system(ptchCmd)
try:
baseFile = open(tmpFile1)
except:
error("could not open: " + tmpFile1)
return
try:
ptchFile = open(tmpFile2)
except:
error("could not open: " + tmpFile2)
return
baseData = baseFile.read()
ptchData = ptchFile.read()
baseFile.close()
ptchFile.close()
if len(baseData) != len(ptchData) or baseData != ptchData :
if not quiet :
data = diffFileData(tmpFile1, tmpFile2);
try:
os.unlink(tmpFile1)
except OSError, e:
error("compareByDumping: unlink failed %s" % e)
try:
os.unlink(tmpFile2)
except OSError, e:
error("compareByDumping: unlink failed %s" % e)
difference(fileName, fileType, data)
return 1
# Remove the temporary files now.
try:
os.unlink(tmpFile1)
except OSError, e:
error("compareByDumping: unlink failed %s" % e)
try:
os.unlink(tmpFile2)
except OSError, e:
error("compareByDumping: unlink failed %s" % e)
return 0
#####
#
# SIGINT signal handler. Changes thread control variable to tell the threads
# to finish their current job and exit.
#
def discontinue_processing(signl, frme):
global keep_processing
print >> sys.stderr, "Caught Ctrl-C, stopping the threads"
keep_processing = False
return 0
#####
#
# worker thread for changedFiles processing
#
class workerThread(threading.Thread) :
def run(self):
global wset_lock
global changedFiles
global baseRoot
global ptchRoot
global keep_processing
while (keep_processing) :
# grab the lock to changedFiles and remove one member
# and process it
wset_lock.acquire()
try :
fn = changedFiles.pop()
except IndexError :
# there is nothing more to do
wset_lock.release()
return
wset_lock.release()
base = baseRoot + fn
ptch = ptchRoot + fn
compareOneFile(base, ptch, False)
#####
# Compare two objects. Detect type changes.
# Vector off to the appropriate type specific
# compare routine based on the type.
#
def compareOneFile(base, ptch, quiet) :
# Verify the file types.
# If they are different, indicate this and move on
btype = getTheFileType(base)
ptype = getTheFileType(ptch)
if btype == 'Error' or ptype == 'Error' :
return -1
fileName = fnFormat(base)
if (btype != ptype) :
if not quiet :
difference(fileName, "file type", btype + " to " + ptype)
return 1
else :
fileType = btype
if (fileType == 'ELF') :
return compareElfs(base, ptch, quiet)
elif (fileType == 'Java Archive' or fileType == 'ELF Object Archive') :
return compareArchives(base, ptch, fileType)
elif (fileType == 'HTML') :
return compareBasic(base, ptch, quiet, fileType)
elif ( fileType == 'Lint Library' ) :
return compareByDumping(base, ptch, quiet, fileType)
elif ( fileType == 'Sqlite Database' ) :
return compareByDumping(base, ptch, quiet, fileType)
else :
# it has to be some variety of text file
return compareBasic(base, ptch, quiet, fileType)
# Cleanup and self-terminate
def cleanup(ret) :
debug("Performing cleanup (" + str(ret) + ")")
if os.path.isdir(tmpDir1) > 0 :
shutil.rmtree(tmpDir1)
if os.path.isdir(tmpDir2) > 0 :
shutil.rmtree(tmpDir2)
if logging :
log.close()
sys.exit(ret)
def main() :
# Log file handle
global log
# Globals relating to command line options
global logging, vdiffs, reportAllSects
# Named temporary files / directories
global tmpDir1, tmpDir2
# Command paths
global lintdump_cmd, elfdump_cmd, dump_cmd, dis_cmd, od_cmd, diff_cmd, sqlite_cmd
# Default search path
global wsdiff_path
# Essentially "uname -p"
global arch
# changed files for worker thread processing
global changedFiles
global baseRoot
global ptchRoot
# Sort the list of files from a temporary file
global sorted
global differentFiles
# Debugging indicator
global debugon
# Some globals need to be initialized
debugon = logging = vdiffs = reportAllSects = sorted = False
# Process command line arguments
# Return values are returned from args() in alpha order
# (Yes, python functions can return multiple values (ewww))
# Note that args() also set the globals:
# logging to True if verbose logging (to a file) was enabled
# vdiffs to True if logged differences aren't to be truncated
# reportAllSects to True if all ELF section differences are to be reported
#
baseRoot, fileNamesFile, localTools, ptchRoot, results = args()
#
# Set up the results/log file
#
if logging :
try:
log = open(results, "w")
except:
logging = False
error("failed to open log file: " + log)
sys.exit(1)
dateTimeStr= "# %d/%d/%d at %d:%d:%d" % time.localtime()[:6]
v_info("# This file was produced by wsdiff")
v_info(dateTimeStr)
# Changed files (used only for the sorted case)
if sorted :
differentFiles = []
#
# Build paths to the tools required tools
#
# Try to look for tools in $SRC/tools if the "-t" option
# was specified
#
arch = commands.getoutput("uname -p")
if localTools :
try:
src = os.environ['SRC']
except:
error("-t specified, but $SRC not set. Cannot find $SRC/tools")
src = ""
if len(src) > 0 :
wsdiff_path.insert(0, src + "/tools/proto/opt/onbld/bin")
lintdump_cmd = find_tool("lintdump")
elfdump_cmd = find_tool("elfdump")
dump_cmd = find_tool("dump")
od_cmd = find_tool("od")
dis_cmd = find_tool("dis")
diff_cmd = find_tool("diff")
sqlite_cmd = find_tool("sqlite")
#
# Set resource limit for number of open files as high as possible.
# This might get handy with big number of threads.
#
(nofile_soft, nofile_hard) = resource.getrlimit(resource.RLIMIT_NOFILE)
try:
resource.setrlimit(resource.RLIMIT_NOFILE,
(nofile_hard, nofile_hard))
except:
error("cannot set resource limits for number of open files")
sys.exit(1)
#
# validate the base and patch paths
#
if baseRoot[-1] != '/' :
baseRoot += '/'
if ptchRoot[-1] != '/' :
ptchRoot += '/'
if not os.path.exists(baseRoot) :
error("old proto area: " + baseRoot + " does not exist")
sys.exit(1)
if not os.path.exists(ptchRoot) :
error("new proto area: " + ptchRoot + \
" does not exist")
sys.exit(1)
#
# log some information identifying the run
#
v_info("Old proto area: " + baseRoot)
v_info("New proto area: " + ptchRoot)
v_info("Results file: " + results + "\n")
#
# Set up the temporary directories / files
# Could use python's tmpdir routines, but these should
# be easier to identify / keep around for debugging
pid = os.getpid()
tmpDir1 = "/tmp/wsdiff_tmp1_" + str(pid) + "/"
tmpDir2 = "/tmp/wsdiff_tmp2_" + str(pid) + "/"
try:
os.makedirs(tmpDir1)
except OSError, e:
error("main: makedir failed %s" % e)
try:
os.makedirs(tmpDir2)
except OSError, e:
error("main: makedir failed %s" % e)
# Derive a catalog of new, deleted, and to-be-compared objects
# either from the specified base and patch proto areas, or from
# from an input file list
newOrDeleted = False
if fileNamesFile != "" :
changedFiles, newFiles, deletedFiles = \
flistCatalog(baseRoot, ptchRoot, fileNamesFile)
else :
changedFiles, newFiles, deletedFiles = \
protoCatalog(baseRoot, ptchRoot)
if len(newFiles) > 0 :
newOrDeleted = True
info("\nNew objects found: ")
if sorted :
newFiles.sort()
for fn in newFiles :
info(fnFormat(fn))
if len(deletedFiles) > 0 :
newOrDeleted = True
info("\nObjects removed: ")
if sorted :
deletedFiles.sort()
for fn in deletedFiles :
info(fnFormat(fn))
if newOrDeleted :
info("\nChanged objects: ")
if sorted :
debug("The list will appear after the processing is done")
# Here's where all the heavy lifting happens
# Perform a comparison on each object appearing in
# both proto areas. compareOneFile will examine the
# file types of each object, and will vector off to
# the appropriate comparison routine, where the compare
# will happen, and any differences will be reported / logged
# determine maximum number of worker threads by using
# DMAKE_MAX_JOBS environment variable set by nightly(1)
# or get number of CPUs in the system
try:
max_threads = int(os.environ['DMAKE_MAX_JOBS'])
except:
max_threads = os.sysconf("SC_NPROCESSORS_ONLN")
# If we cannot get number of online CPUs in the system
# run unparallelized otherwise bump the number up 20%
# to achieve best results.
if max_threads == -1 :
max_threads = 1
else :
max_threads += max_threads/5
# Set signal handler to attempt graceful exit
debug("Setting signal handler")
signal.signal( signal.SIGINT, discontinue_processing )
# Create and unleash the threads
# Only at most max_threads must be running at any moment
mythreads = []
debug("Spawning " + str(max_threads) + " threads");
for i in range(max_threads) :
thread = workerThread()
mythreads.append(thread)
mythreads[i].start()
# Wait for the threads to finish and do cleanup if interrupted
debug("Waiting for the threads to finish")
while True:
if not True in [thread.isAlive() for thread in mythreads]:
break
else:
# Some threads are still going
time.sleep(1)
# Interrupted by SIGINT
if keep_processing == False :
cleanup(1)
# If the list of differences was sorted it is stored in an array
if sorted :
differentFiles.sort()
for f in differentFiles :
info(fnFormat(f))
# We're done, cleanup.
cleanup(0)
if __name__ == '__main__' :
try:
main()
except KeyboardInterrupt :
cleanup(1);