cppcheck/tools/daca2.py

229 lines
6.6 KiB
Python

#!/usr/bin/python
#
# 1. Create a folder daca2 in your HOME folder
# 2. Put cppcheck-O2 in daca2. It should be built with all optimisations.
# 3. Optional: Put a file called "suppressions.txt" in the daca2 folder.
# 4. Optional: tweak FTPSERVER and FTPPATH in this script below.
# 5. Run the daca2 script: python daca2.py FOLDER
import argparse
import subprocess
import sys
import shutil
import glob
import os
import datetime
import time
import logging
DEBIAN = ['ftp://ftp.se.debian.org/debian/',
'ftp://ftp.debian.org/debian/']
def wget(filepath):
filename = filepath
if filepath.find('/') >= 0:
filename = filename[filename.rfind('/') + 1:]
for d in DEBIAN:
subprocess.call(
['nice', 'wget', '--tries=10', '--timeout=300', '-O', filename, d + filepath])
if os.path.isfile(filename):
return True
print('Sleep for 10 seconds..')
time.sleep(10)
return False
def getpackages(folder):
if not wget('ls-lR.gz'):
return []
subprocess.call(['nice', 'gunzip', 'ls-lR.gz'])
f = open('ls-lR', 'rt')
lines = f.readlines()
f.close()
subprocess.call(['rm', 'ls-lR'])
path = None
archives = []
filename = None
for line in lines:
line = line.strip()
if len(line) < 4:
if filename:
archives.append(path + '/' + filename)
path = None
filename = None
elif line[:13 + len(folder)] == './pool/main/' + folder + '/':
path = line[2:-1]
elif path and line.find('.orig.tar.') > 0:
filename = line[1 + line.rfind(' '):]
for a in archives:
print(a)
return archives
def handleRemoveReadonly(func, path, exc):
import stat
if not os.access(path, os.W_OK):
# Is the error an access error ?
os.chmod(path, stat.S_IWUSR)
func(path)
else:
raise
def removeAllExceptResults():
filenames = []
for g in glob.glob('[A-Za-z0-9]*'):
filenames.append(g)
for g in glob.glob('.[a-z]*'):
filenames.append(g)
for filename in filenames:
count = 5
while count > 0:
count = count - 1
try:
if os.path.isdir(filename):
shutil.rmtree(filename, onerror=handleRemoveReadonly)
elif filename != RESULTS_FILENAME:
os.remove(filename)
break
except WindowsError as err:
time.sleep(30)
if count == 0:
logging.error('Failed to cleanup {}: {}'.format(filename, err))
except OSError as err:
time.sleep(30)
if count == 0:
logging.error('Failed to cleanup {}: {}'.format(filename, err))
def removeLargeFiles(path):
for g in glob.glob(path + '*'):
if g == '.' or g == '..':
continue
if os.path.islink(g):
continue
if os.path.isdir(g):
# Remove test code
if g.endswith('/testsuite') or g.endswith('/clang/INPUTS'):
shutil.rmtree(g, onerror=handleRemoveReadonly)
else:
removeLargeFiles(g + '/')
elif os.path.isfile(g) and g[-4:] != '.txt':
statinfo = os.stat(g)
if statinfo.st_size > 1000000:
try:
os.remove(g)
except OSError as err:
logging.error('Failed to remove {}: {}'.format(g, err))
def strfCurrTime(fmt):
return datetime.time.strftime(datetime.datetime.now().time(), fmt)
def scanarchive(filepath, jobs, cpulimit):
# remove all files/folders except RESULTS_FILENAME
removeAllExceptResults()
logging.info(DEBIAN[0] + filepath)
if not wget(filepath):
if not wget(filepath):
logging.error('wget failed at {}', filepath)
return
filename = filepath[filepath.rfind('/') + 1:]
if filename[-3:] == '.gz':
subprocess.call(['tar', 'xzvf', filename])
elif filename[-3:] == '.xz':
subprocess.call(['tar', 'xJvf', filename])
elif filename[-4:] == '.bz2':
subprocess.call(['tar', 'xjvf', filename])
removeLargeFiles('')
print(strfCurrTime('[%H:%M] cppcheck ') + filename)
if cpulimit:
cmd = 'cpulimit --limit=' + cpulimit
else:
cmd = 'nice --adjustment=1000'
cmd = cmd + ' ../cppcheck-O2 -D__GCC__ --enable=style --inconclusive --error-exitcode=0 --exception-handling=stderr ' + jobs + ' --template=cppcheck2 .'
cmds = cmd.split()
p = subprocess.Popen(cmds, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
comm = p.communicate()
if p.returncode == 0:
logging.info(comm[1] + strfCurrTime('[%H:%M]'))
elif comm[0].find('cppcheck: error: could not find or open any of the paths given.') < 0:
logging.error(comm[1] + strfCurrTime('[%H:%M]'))
logging.error('Exit code is not zero! Crash?\n')
parser = argparse.ArgumentParser(description='Checks debian source code')
parser.add_argument('folder', metavar='FOLDER')
parser.add_argument('--rev')
parser.add_argument('--workdir', default='~/daca2')
parser.add_argument('-j', '--jobs', default='-j1')
parser.add_argument('--skip', default=[], action='append')
parser.add_argument('--cpulimit')
args = parser.parse_args()
workdir = os.path.expanduser(args.workdir)
if not os.path.isdir(workdir):
print('workdir \'' + workdir + '\' is not a folder')
sys.exit(1)
workdir = os.path.join(workdir, args.folder)
if not os.path.isdir(workdir):
os.makedirs(workdir)
RESULTS_FILENAME = 'results.txt'
RESULTS_FILE = os.path.join(workdir, RESULTS_FILENAME)
logging.basicConfig(
filename=RESULTS_FILE,
level=logging.INFO,
format='%(message)s')
print(workdir)
archives = getpackages(args.folder)
if len(archives) == 0:
logging.critical('failed to load packages')
sys.exit(1)
if not os.path.isdir(workdir):
os.makedirs(workdir)
os.chdir(workdir)
try:
logging.info('STARTDATE ' + str(datetime.date.today()))
logging.info('STARTTIME ' + strfCurrTime('%H:%M:%S'))
if args.rev:
logging.info('GIT-REVISION ' + args.rev + '\n')
logging.info('')
for archive in archives:
if len(args.skip) > 0:
a = archive[:archive.rfind('/')]
a = a[a.rfind('/')+1:]
if a in args.skip:
continue
scanarchive(archive, args.jobs, args.cpulimit)
logging.info('DATE {}'.format(datetime.date.today()))
logging.info('TIME {}'.format(strfCurrTime('%H:%M:%S')))
except EOFError:
pass
# remove all files/folders except RESULTS_FILENAME
removeAllExceptResults()