diff options
author | Noel Grandin <noel.grandin@collabora.co.uk> | 2019-03-14 12:55:25 +0200 |
---|---|---|
committer | Noel Grandin <noel.grandin@collabora.co.uk> | 2019-03-15 12:56:09 +0200 |
commit | 6c46fdd75de8cec4f62f9fed02212a2d1e0b71b5 (patch) | |
tree | 4e1a3c1ebfc9d9b87dbc0810827686df4192c1f8 /compilerplugins/clang/pahole-all-classes.py | |
parent | 22de481db334fcce404cdcf88183cd544cd19271 (diff) |
new pahole-all-classes script, and update a couple of sc/ structs
Create a new script to scan our codebase for holes in our
structures.
Implemented a couple of the things I found
ScSortKeyState 12bytes -> 8bytes
sc::MultiDataCellState 12bytes -> 8bytes
Change-Id: I139dda36aedf02b7f19be121eb312e5552142b87
Diffstat (limited to 'compilerplugins/clang/pahole-all-classes.py')
-rwxr-xr-x | compilerplugins/clang/pahole-all-classes.py | 123 |
1 files changed, 123 insertions, 0 deletions
diff --git a/compilerplugins/clang/pahole-all-classes.py b/compilerplugins/clang/pahole-all-classes.py new file mode 100755 index 000000000000..9fda73c8789f --- /dev/null +++ b/compilerplugins/clang/pahole-all-classes.py @@ -0,0 +1,123 @@ +#!/usr/bin/python3 +# +# Find holes in structures, so that we can pack them and improve our memory density. +# +# In order to make this work, you need to +# (1) be operating in a workspace where you have a debug build of LibreOffice +# (2) first run the unusedfields loplugin to generate a log file +# (3) install the pahole stuff into your gdb, I used this one: https://github.com/PhilArmstrong/pahole-gdb +# (4) ./compilerplugins/clang/pahole-all-classes.py > ./compilerplugins/clang/pahole.results +# Warning: running this script will make GDB soak up about 8G of RAM +# + +import _thread +import io +import os +import subprocess +import time +import re + +# search for all the class names in the file produced by the unusedfields loplugin +#a = subprocess.Popen("grep 'definition:' workdir/loplugin.unusedfields.log | sort -u", stdout=subprocess.PIPE, shell=True) +a = subprocess.Popen("cat n1", stdout=subprocess.PIPE, shell=True) + +classSourceLocDict = dict() +classSet = set() +with a.stdout as txt: + for line in txt: + tokens = line.decode('utf8').strip().split("\t") + className = tokens[2].strip() + srcLoc = tokens[5].strip() + if "anonymous" in className: continue + # for now, just check the stuff in /sc/inc + if not srcLoc.startswith("sc/inc/"): + continue + if className in classSet: continue + classSourceLocDict[srcLoc] = className + classSet.add(className) +a.terminate() + +gdbProc = subprocess.Popen("gdb", stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, shell=True) + +stdin = io.TextIOWrapper(gdbProc.stdin, 'utf-8') + +# make gdb load all the debugging info +stdin.write("set confirm off\n") +for filename in sorted(os.listdir('instdir/program')): + if filename.endswith(".so"): + stdin.write("add-symbol-file instdir/program/" + filename + "\n") +stdin.flush() + + +# Some of the pahole commands are going to fail, and I cannot read the error stream and the input stream +# together because python has no way of (easily) doing a non-blocking read. +# So I have to write the commands out using a background thread, and then read the entire resulting +# stream out below. +def write_pahole_commands(): + for srcLoc in sorted(classSourceLocDict.keys()): + className = classSourceLocDict[srcLoc] + stdin.write("echo " + className + " " + srcLoc + "\n") + stdin.write("pahole " + className + "\n") + stdin.flush() + stdin.write("echo all-done\n") + stdin.flush() + stdin.close() # only way to make it flush the last echo command + +_thread.start_new_thread( write_pahole_commands, () ) + +time.sleep(2) + +# Use generator because lines often end up merged together in gdb's output, and we need +# to split them up, and that creates a mess in the parsing logic. +def read_generator(): + while True: + line = gdbProc.stdout.readline().decode('utf8').strip() + for split in line.split("(gdb)"): + split = split.strip() + if len(split) == 0: continue + if "all-done" in split: return + yield split + +firstLineRegex = re.compile("/\*\s+(\d+)\s+\*/ struct") +fieldLineRegex = re.compile("/\*\s+\d+\s+(\d+)\s+\*/ ") +holeLineRegex = re.compile("/\* XXX (\d+) bit hole, try to pack \*/") +structLines = list() +foundHole = False +cumulativeHoleBits = 0 +structSize = 0 +found8ByteField = False +for line in read_generator(): + structLines.append(line) + firstLineMatch = firstLineRegex.match(line) + if firstLineMatch: + structSize = int(firstLineMatch.group(1)) + holeLineMatch = holeLineRegex.match(line) + if holeLineMatch: + foundHole = True + cumulativeHoleBits += int(holeLineMatch.group(1)) + fieldLineMatch = fieldLineRegex.match(line) + if fieldLineMatch: + fieldSize = int(fieldLineMatch.group(1)) + if fieldSize == 8: + found8ByteField = True + if line == "}": + # Ignore very large structs, packing those is not going to help much, and + # re-organising them can make them much less readable. + if foundHole and len(structLines) < 12 and structSize < 100: + # If we have an 8-byte field, then the whole structure must be 8-byte aligned, otherwise + # it must be 4-byte aligned. (that's my approximation of the rules, the real ones are probably + # more complicated). So check if removing the holes will remove enough space to actually shrink + # this structure. + alignBytes = 4 + if found8ByteField: alignBytes = 8 + if (cumulativeHoleBits / 8) >= alignBytes: + # print("Found one " + str(structSize) + " " + str(cumulativeHoleBits/8) + " " + str(newStructSize%4)) + for line in structLines: + print(line) + structLines.clear() + foundHole = False + cumulativeHoleBits = 0 + structSize = 0 + found8ByteField = False + +gdbProc.terminate()
\ No newline at end of file |