Merge pull request #98992 from AThousandShips/cache_clean_improve_3_x
Some checks are pending
🔗 GHA / 📊 Static (push) Waiting to run
🔗 GHA / 🤖 Android (push) Blocked by required conditions
🔗 GHA / 🍏 iOS (push) Blocked by required conditions
🔗 GHA / 🌐 JavaScript (push) Blocked by required conditions
🔗 GHA / 🐧 Linux (push) Blocked by required conditions
🔗 GHA / 🍎 macOS (push) Blocked by required conditions
🔗 GHA / ☁ Server (push) Blocked by required conditions
🔗 GHA / 🏁 Windows (push) Blocked by required conditions

[3.x] [Buildsystem] Improve cache handling
This commit is contained in:
lawnjelly 2024-11-20 11:54:58 +00:00 committed by GitHub
commit f75abbb3ec
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
2 changed files with 64 additions and 38 deletions

View File

@ -796,12 +796,13 @@ elif selected_platform != "":
# The following only makes sense when the 'env' is defined, and assumes it is. # The following only makes sense when the 'env' is defined, and assumes it is.
if "env" in locals(): if "env" in locals():
# FIXME: This method mixes both cosmetic progress stuff and cache handling...
methods.show_progress(env) methods.show_progress(env)
# TODO: replace this with `env.Dump(format="json")` # TODO: replace this with `env.Dump(format="json")`
# once we start requiring SCons 4.0 as min version. # once we start requiring SCons 4.0 as min version.
methods.dump(env) methods.dump(env)
methods.clean_cache(env)
def print_elapsed_time(): def print_elapsed_time():
elapsed_time_sec = round(time.time() - time_at_start, 3) elapsed_time_sec = round(time.time() - time_at_start, 3)

View File

@ -1126,21 +1126,19 @@ def show_progress(env):
"fname": str(env.Dir("#")) + "/.scons_node_count", "fname": str(env.Dir("#")) + "/.scons_node_count",
} }
import time, math import math
class cache_progress: class cache_progress:
# The default is 1 GB cache and 12 hours half life # The default is 1 GB cache
def __init__(self, path=None, limit=1073741824, half_life=43200): def __init__(self, path=None, limit=pow(1024, 3)):
self.path = path self.path = path
self.limit = limit self.limit = limit
self.exponent_scale = math.log(2) / half_life
if env["verbose"] and path != None: if env["verbose"] and path != None:
screen.write( screen.write(
"Current cache limit is {} (used: {})\n".format( "Current cache limit is {} (used: {})\n".format(
self.convert_size(limit), self.convert_size(self.get_size(path)) self.convert_size(limit), self.convert_size(self.get_size(path))
) )
) )
self.delete(self.file_list())
def __call__(self, node, *args, **kw): def __call__(self, node, *args, **kw):
if show_progress: if show_progress:
@ -1158,12 +1156,65 @@ def show_progress(env):
screen.write("\r[Initial build] ") screen.write("\r[Initial build] ")
screen.flush() screen.flush()
def convert_size(self, size_bytes):
if size_bytes == 0:
return "0 bytes"
size_name = ("bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB")
i = int(math.floor(math.log(size_bytes, 1024)))
p = math.pow(1024, i)
s = round(size_bytes / p, 2)
return "%s %s" % (int(s) if i == 0 else s, size_name[i])
def get_size(self, start_path="."):
total_size = 0
for dirpath, dirnames, filenames in os.walk(start_path):
for f in filenames:
fp = os.path.join(dirpath, f)
total_size += os.path.getsize(fp)
return total_size
def progress_finish(target, source, env):
try:
with open(node_count_data["fname"], "w") as f:
f.write("%d\n" % node_count_data["count"])
except Exception:
pass
try:
with open(node_count_data["fname"]) as f:
node_count_data["max"] = int(f.readline())
except Exception:
pass
cache_directory = os.environ.get("SCONS_CACHE")
# Simple cache pruning, attached to SCons' progress callback. Trim the
# cache directory to a size not larger than cache_limit.
cache_limit = float(os.getenv("SCONS_CACHE_LIMIT", 1024)) * 1024 * 1024
progressor = cache_progress(cache_directory, cache_limit)
Progress(progressor, interval=node_count_data["interval"])
progress_finish_command = Command("progress_finish", [], progress_finish)
AlwaysBuild(progress_finish_command)
def clean_cache(env):
import atexit
import time
class cache_clean:
def __init__(self, path=None, limit=pow(1024, 3)):
self.path = path
self.limit = limit
def clean(self):
self.delete(self.file_list())
def delete(self, files): def delete(self, files):
if len(files) == 0: if len(files) == 0:
return return
if env["verbose"]: if env["verbose"]:
# Utter something # Utter something
screen.write("\rPurging %d %s from cache...\n" % (len(files), len(files) > 1 and "files" or "file")) print("Purging %d %s from cache..." % (len(files), "files" if len(files) > 1 else "file"))
[os.remove(f) for f in files] [os.remove(f) for f in files]
def file_list(self): def file_list(self):
@ -1197,34 +1248,10 @@ def show_progress(env):
else: else:
return [x[0] for x in file_stat[mark:]] return [x[0] for x in file_stat[mark:]]
def convert_size(self, size_bytes): def cache_finally():
if size_bytes == 0: nonlocal cleaner
return "0 bytes"
size_name = ("bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB")
i = int(math.floor(math.log(size_bytes, 1024)))
p = math.pow(1024, i)
s = round(size_bytes / p, 2)
return "%s %s" % (int(s) if i == 0 else s, size_name[i])
def get_size(self, start_path="."):
total_size = 0
for dirpath, dirnames, filenames in os.walk(start_path):
for f in filenames:
fp = os.path.join(dirpath, f)
total_size += os.path.getsize(fp)
return total_size
def progress_finish(target, source, env):
try: try:
with open(node_count_data["fname"], "w") as f: cleaner.clean()
f.write("%d\n" % node_count_data["count"])
progressor.delete(progressor.file_list())
except Exception:
pass
try:
with open(node_count_data["fname"]) as f:
node_count_data["max"] = int(f.readline())
except Exception: except Exception:
pass pass
@ -1232,11 +1259,9 @@ def show_progress(env):
# Simple cache pruning, attached to SCons' progress callback. Trim the # Simple cache pruning, attached to SCons' progress callback. Trim the
# cache directory to a size not larger than cache_limit. # cache directory to a size not larger than cache_limit.
cache_limit = float(os.getenv("SCONS_CACHE_LIMIT", 1024)) * 1024 * 1024 cache_limit = float(os.getenv("SCONS_CACHE_LIMIT", 1024)) * 1024 * 1024
progressor = cache_progress(cache_directory, cache_limit) cleaner = cache_clean(cache_directory, cache_limit)
Progress(progressor, interval=node_count_data["interval"])
progress_finish_command = Command("progress_finish", [], progress_finish) atexit.register(cache_finally)
AlwaysBuild(progress_finish_command)
def dump(env): def dump(env):