12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205 |
- import os
- import sys
- import re
- import glob
- import subprocess
- from collections import OrderedDict
- from collections.abc import Mapping
- from typing import Iterator
- from pathlib import Path
- from os.path import normpath, basename
- # Get the "Godot" folder name ahead of time
- base_folder_path = str(os.path.abspath(Path(__file__).parent)) + "/"
- base_folder_only = os.path.basename(os.path.normpath(base_folder_path))
- # Listing all the folders we have converted
- # for SCU in scu_builders.py
- _scu_folders = set()
- def set_scu_folders(scu_folders):
- global _scu_folders
- _scu_folders = scu_folders
- def add_source_files_orig(self, sources, files, allow_gen=False):
- # Convert string to list of absolute paths (including expanding wildcard)
- if isinstance(files, (str, bytes)):
- # Keep SCons project-absolute path as they are (no wildcard support)
- if files.startswith("#"):
- if "*" in files:
- print("ERROR: Wildcards can't be expanded in SCons project-absolute path: '{}'".format(files))
- return
- files = [files]
- else:
- # Exclude .gen.cpp files from globbing, to avoid including obsolete ones.
- # They should instead be added manually.
- skip_gen_cpp = "*" in files
- dir_path = self.Dir(".").abspath
- files = sorted(glob.glob(dir_path + "/" + files))
- if skip_gen_cpp and not allow_gen:
- files = [f for f in files if not f.endswith(".gen.cpp")]
- # Add each path as compiled Object following environment (self) configuration
- for path in files:
- obj = self.Object(path)
- if obj in sources:
- print('WARNING: Object "{}" already included in environment sources.'.format(obj))
- continue
- sources.append(obj)
- # The section name is used for checking
- # the hash table to see whether the folder
- # is included in the SCU build.
- # It will be something like "core/math".
- def _find_scu_section_name(subdir):
- section_path = os.path.abspath(subdir) + "/"
- folders = []
- folder = ""
- for i in range(8):
- folder = os.path.dirname(section_path)
- folder = os.path.basename(folder)
- if folder == base_folder_only:
- break
- folders += [folder]
- section_path += "../"
- section_path = os.path.abspath(section_path) + "/"
- section_name = ""
- for n in range(len(folders)):
- # section_name += folders[len(folders) - n - 1] + " "
- section_name += folders[len(folders) - n - 1]
- if n != (len(folders) - 1):
- section_name += "/"
- return section_name
- def add_source_files_scu(self, sources, files, allow_gen=False):
- if self["scu_build"] and isinstance(files, str):
- if "*." not in files:
- return False
- # If the files are in a subdirectory, we want to create the scu gen
- # files inside this subdirectory.
- subdir = os.path.dirname(files)
- if subdir != "":
- subdir += "/"
- section_name = _find_scu_section_name(subdir)
- # if the section name is in the hash table?
- # i.e. is it part of the SCU build?
- global _scu_folders
- if section_name not in (_scu_folders):
- return False
- if self["verbose"]:
- print("SCU building " + section_name)
- # Add all the gen.cpp files in the SCU directory
- add_source_files_orig(self, sources, subdir + "scu/scu_*.gen.cpp", True)
- return True
- return False
- # Either builds the folder using the SCU system,
- # or reverts to regular build.
- def add_source_files(self, sources, files, allow_gen=False):
- if not add_source_files_scu(self, sources, files, allow_gen):
- # Wraps the original function when scu build is not active.
- add_source_files_orig(self, sources, files, allow_gen)
- return False
- return True
- def disable_warnings(self):
- # 'self' is the environment
- if self.msvc:
- # We have to remove existing warning level defines before appending /w,
- # otherwise we get: "warning D9025 : overriding '/W3' with '/w'"
- self["CCFLAGS"] = [x for x in self["CCFLAGS"] if not (x.startswith("/W") or x.startswith("/w"))]
- self["CFLAGS"] = [x for x in self["CFLAGS"] if not (x.startswith("/W") or x.startswith("/w"))]
- self["CXXFLAGS"] = [x for x in self["CXXFLAGS"] if not (x.startswith("/W") or x.startswith("/w"))]
- self.AppendUnique(CCFLAGS=["/w"])
- else:
- self.AppendUnique(CCFLAGS=["-w"])
- def force_optimization_on_debug(self):
- # 'self' is the environment
- if self["target"] == "template_release":
- return
- if self.msvc:
- # We have to remove existing optimization level defines before appending /O2,
- # otherwise we get: "warning D9025 : overriding '/0d' with '/02'"
- self["CCFLAGS"] = [x for x in self["CCFLAGS"] if not x.startswith("/O")]
- self["CFLAGS"] = [x for x in self["CFLAGS"] if not x.startswith("/O")]
- self["CXXFLAGS"] = [x for x in self["CXXFLAGS"] if not x.startswith("/O")]
- self.AppendUnique(CCFLAGS=["/O2"])
- else:
- self.AppendUnique(CCFLAGS=["-O3"])
- def add_module_version_string(self, s):
- self.module_version_string += "." + s
- def get_version_info(module_version_string="", silent=False):
- build_name = "custom_build"
- if os.getenv("BUILD_NAME") != None:
- build_name = str(os.getenv("BUILD_NAME"))
- if not silent:
- print(f"Using custom build name: '{build_name}'.")
- import version
- version_info = {
- "short_name": str(version.short_name),
- "name": str(version.name),
- "major": int(version.major),
- "minor": int(version.minor),
- "patch": int(version.patch),
- "status": str(version.status),
- "build": str(build_name),
- "module_config": str(version.module_config) + module_version_string,
- "year": int(version.year),
- "website": str(version.website),
- "docs_branch": str(version.docs),
- }
- # For dev snapshots (alpha, beta, RC, etc.) we do not commit status change to Git,
- # so this define provides a way to override it without having to modify the source.
- if os.getenv("GODOT_VERSION_STATUS") != None:
- version_info["status"] = str(os.getenv("GODOT_VERSION_STATUS"))
- if not silent:
- print(f"Using version status '{version_info['status']}', overriding the original '{version.status}'.")
- # Parse Git hash if we're in a Git repo.
- githash = ""
- gitfolder = ".git"
- if os.path.isfile(".git"):
- module_folder = open(".git", "r").readline().strip()
- if module_folder.startswith("gitdir: "):
- gitfolder = module_folder[8:]
- if os.path.isfile(os.path.join(gitfolder, "HEAD")):
- head = open(os.path.join(gitfolder, "HEAD"), "r", encoding="utf8").readline().strip()
- if head.startswith("ref: "):
- ref = head[5:]
- # If this directory is a Git worktree instead of a root clone.
- parts = gitfolder.split("/")
- if len(parts) > 2 and parts[-2] == "worktrees":
- gitfolder = "/".join(parts[0:-2])
- head = os.path.join(gitfolder, ref)
- packedrefs = os.path.join(gitfolder, "packed-refs")
- if os.path.isfile(head):
- githash = open(head, "r").readline().strip()
- elif os.path.isfile(packedrefs):
- # Git may pack refs into a single file. This code searches .git/packed-refs file for the current ref's hash.
- # https://mirrors.edge.kernel.org/pub/software/scm/git/docs/git-pack-refs.html
- for line in open(packedrefs, "r").read().splitlines():
- if line.startswith("#"):
- continue
- (line_hash, line_ref) = line.split(" ")
- if ref == line_ref:
- githash = line_hash
- break
- else:
- githash = head
- version_info["git_hash"] = githash
- return version_info
- def generate_version_header(module_version_string=""):
- version_info = get_version_info(module_version_string)
- # NOTE: It is safe to generate these files here, since this is still executed serially.
- f = open("core/version_generated.gen.h", "w")
- f.write(
- """/* THIS FILE IS GENERATED DO NOT EDIT */
- #ifndef VERSION_GENERATED_GEN_H
- #define VERSION_GENERATED_GEN_H
- #define VERSION_SHORT_NAME "{short_name}"
- #define VERSION_NAME "{name}"
- #define VERSION_MAJOR {major}
- #define VERSION_MINOR {minor}
- #define VERSION_PATCH {patch}
- #define VERSION_STATUS "{status}"
- #define VERSION_BUILD "{build}"
- #define VERSION_MODULE_CONFIG "{module_config}"
- #define VERSION_YEAR {year}
- #define VERSION_WEBSITE "{website}"
- #define VERSION_DOCS_BRANCH "{docs_branch}"
- #define VERSION_DOCS_URL "https://docs.godotengine.org/en/" VERSION_DOCS_BRANCH
- #endif // VERSION_GENERATED_GEN_H
- """.format(
- **version_info
- )
- )
- f.close()
- fhash = open("core/version_hash.gen.cpp", "w")
- fhash.write(
- """/* THIS FILE IS GENERATED DO NOT EDIT */
- #include "core/version.h"
- const char *const VERSION_HASH = "{git_hash}";
- """.format(
- **version_info
- )
- )
- fhash.close()
- def parse_cg_file(fname, uniforms, sizes, conditionals):
- fs = open(fname, "r")
- line = fs.readline()
- while line:
- if re.match(r"^\s*uniform", line):
- res = re.match(r"uniform ([\d\w]*) ([\d\w]*)")
- type = res.groups(1)
- name = res.groups(2)
- uniforms.append(name)
- if type.find("texobj") != -1:
- sizes.append(1)
- else:
- t = re.match(r"float(\d)x(\d)", type)
- if t:
- sizes.append(int(t.groups(1)) * int(t.groups(2)))
- else:
- t = re.match(r"float(\d)", type)
- sizes.append(int(t.groups(1)))
- if line.find("[branch]") != -1:
- conditionals.append(name)
- line = fs.readline()
- fs.close()
- def get_cmdline_bool(option, default):
- """We use `ARGUMENTS.get()` to check if options were manually overridden on the command line,
- and SCons' _text2bool helper to convert them to booleans, otherwise they're handled as strings.
- """
- from SCons.Script import ARGUMENTS
- from SCons.Variables.BoolVariable import _text2bool
- cmdline_val = ARGUMENTS.get(option)
- if cmdline_val is not None:
- return _text2bool(cmdline_val)
- else:
- return default
- def detect_modules(search_path, recursive=False):
- """Detects and collects a list of C++ modules at specified path
- `search_path` - a directory path containing modules. The path may point to
- a single module, which may have other nested modules. A module must have
- "register_types.h", "SCsub", "config.py" files created to be detected.
- `recursive` - if `True`, then all subdirectories are searched for modules as
- specified by the `search_path`, otherwise collects all modules under the
- `search_path` directory. If the `search_path` is a module, it is collected
- in all cases.
- Returns an `OrderedDict` with module names as keys, and directory paths as
- values. If a path is relative, then it is a built-in module. If a path is
- absolute, then it is a custom module collected outside of the engine source.
- """
- modules = OrderedDict()
- def add_module(path):
- module_name = os.path.basename(path)
- module_path = path.replace("\\", "/") # win32
- modules[module_name] = module_path
- def is_engine(path):
- # Prevent recursively detecting modules in self and other
- # Godot sources when using `custom_modules` build option.
- version_path = os.path.join(path, "version.py")
- if os.path.exists(version_path):
- with open(version_path) as f:
- if 'short_name = "godot"' in f.read():
- return True
- return False
- def get_files(path):
- files = glob.glob(os.path.join(path, "*"))
- # Sort so that `register_module_types` does not change that often,
- # and plugins are registered in alphabetic order as well.
- files.sort()
- return files
- if not recursive:
- if is_module(search_path):
- add_module(search_path)
- for path in get_files(search_path):
- if is_engine(path):
- continue
- if is_module(path):
- add_module(path)
- else:
- to_search = [search_path]
- while to_search:
- path = to_search.pop()
- if is_module(path):
- add_module(path)
- for child in get_files(path):
- if not os.path.isdir(child):
- continue
- if is_engine(child):
- continue
- to_search.insert(0, child)
- return modules
- def is_module(path):
- if not os.path.isdir(path):
- return False
- must_exist = ["register_types.h", "SCsub", "config.py"]
- for f in must_exist:
- if not os.path.exists(os.path.join(path, f)):
- return False
- return True
- def write_disabled_classes(class_list):
- f = open("core/disabled_classes.gen.h", "w")
- f.write("/* THIS FILE IS GENERATED DO NOT EDIT */\n")
- f.write("#ifndef DISABLED_CLASSES_GEN_H\n")
- f.write("#define DISABLED_CLASSES_GEN_H\n\n")
- for c in class_list:
- cs = c.strip()
- if cs != "":
- f.write("#define ClassDB_Disable_" + cs + " 1\n")
- f.write("\n#endif\n")
- def write_modules(modules):
- includes_cpp = ""
- initialize_cpp = ""
- uninitialize_cpp = ""
- for name, path in modules.items():
- try:
- with open(os.path.join(path, "register_types.h")):
- includes_cpp += '#include "' + path + '/register_types.h"\n'
- initialize_cpp += "#ifdef MODULE_" + name.upper() + "_ENABLED\n"
- initialize_cpp += "\tinitialize_" + name + "_module(p_level);\n"
- initialize_cpp += "#endif\n"
- uninitialize_cpp += "#ifdef MODULE_" + name.upper() + "_ENABLED\n"
- uninitialize_cpp += "\tuninitialize_" + name + "_module(p_level);\n"
- uninitialize_cpp += "#endif\n"
- except OSError:
- pass
- modules_cpp = """// register_module_types.gen.cpp
- /* THIS FILE IS GENERATED DO NOT EDIT */
- #include "register_module_types.h"
- #include "modules/modules_enabled.gen.h"
- %s
- void initialize_modules(ModuleInitializationLevel p_level) {
- %s
- }
- void uninitialize_modules(ModuleInitializationLevel p_level) {
- %s
- }
- """ % (
- includes_cpp,
- initialize_cpp,
- uninitialize_cpp,
- )
- # NOTE: It is safe to generate this file here, since this is still executed serially
- with open("modules/register_module_types.gen.cpp", "w") as f:
- f.write(modules_cpp)
- def convert_custom_modules_path(path):
- if not path:
- return path
- path = os.path.realpath(os.path.expanduser(os.path.expandvars(path)))
- err_msg = "Build option 'custom_modules' must %s"
- if not os.path.isdir(path):
- raise ValueError(err_msg % "point to an existing directory.")
- if path == os.path.realpath("modules"):
- raise ValueError(err_msg % "be a directory other than built-in `modules` directory.")
- return path
- def disable_module(self):
- self.disabled_modules.append(self.current_module)
- def module_add_dependencies(self, module, dependencies, optional=False):
- """
- Adds dependencies for a given module.
- Meant to be used in module `can_build` methods.
- """
- if module not in self.module_dependencies:
- self.module_dependencies[module] = [[], []]
- if optional:
- self.module_dependencies[module][1].extend(dependencies)
- else:
- self.module_dependencies[module][0].extend(dependencies)
- def module_check_dependencies(self, module):
- """
- Checks if module dependencies are enabled for a given module,
- and prints a warning if they aren't.
- Meant to be used in module `can_build` methods.
- Returns a boolean (True if dependencies are satisfied).
- """
- missing_deps = []
- required_deps = self.module_dependencies[module][0] if module in self.module_dependencies else []
- for dep in required_deps:
- opt = "module_{}_enabled".format(dep)
- if not opt in self or not self[opt]:
- missing_deps.append(dep)
- if missing_deps != []:
- print(
- "Disabling '{}' module as the following dependencies are not satisfied: {}".format(
- module, ", ".join(missing_deps)
- )
- )
- return False
- else:
- return True
- def sort_module_list(env):
- out = OrderedDict()
- deps = {k: v[0] + list(filter(lambda x: x in env.module_list, v[1])) for k, v in env.module_dependencies.items()}
- frontier = list(env.module_list.keys())
- explored = []
- while len(frontier):
- cur = frontier.pop()
- deps_list = deps[cur] if cur in deps else []
- if len(deps_list) and any([d not in explored for d in deps_list]):
- # Will explore later, after its dependencies
- frontier.insert(0, cur)
- continue
- explored.append(cur)
- for k in explored:
- env.module_list.move_to_end(k)
- def use_windows_spawn_fix(self, platform=None):
- if os.name != "nt":
- return # not needed, only for windows
- # On Windows, due to the limited command line length, when creating a static library
- # from a very high number of objects SCons will invoke "ar" once per object file;
- # that makes object files with same names to be overwritten so the last wins and
- # the library loses symbols defined by overwritten objects.
- # By enabling quick append instead of the default mode (replacing), libraries will
- # got built correctly regardless the invocation strategy.
- # Furthermore, since SCons will rebuild the library from scratch when an object file
- # changes, no multiple versions of the same object file will be present.
- self.Replace(ARFLAGS="q")
- def mySubProcess(cmdline, env):
- startupinfo = subprocess.STARTUPINFO()
- startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW
- popen_args = {
- "stdin": subprocess.PIPE,
- "stdout": subprocess.PIPE,
- "stderr": subprocess.PIPE,
- "startupinfo": startupinfo,
- "shell": False,
- "env": env,
- }
- if sys.version_info >= (3, 7, 0):
- popen_args["text"] = True
- proc = subprocess.Popen(cmdline, **popen_args)
- _, err = proc.communicate()
- rv = proc.wait()
- if rv:
- print("=====")
- print(err)
- print("=====")
- return rv
- def mySpawn(sh, escape, cmd, args, env):
- newargs = " ".join(args[1:])
- cmdline = cmd + " " + newargs
- rv = 0
- env = {str(key): str(value) for key, value in iter(env.items())}
- if len(cmdline) > 32000 and cmd.endswith("ar"):
- cmdline = cmd + " " + args[1] + " " + args[2] + " "
- for i in range(3, len(args)):
- rv = mySubProcess(cmdline + args[i], env)
- if rv:
- break
- else:
- rv = mySubProcess(cmdline, env)
- return rv
- self["SPAWN"] = mySpawn
- def no_verbose(sys, env):
- colors = {}
- # Colors are disabled in non-TTY environments such as pipes. This means
- # that if output is redirected to a file, it will not contain color codes
- if sys.stdout.isatty():
- colors["blue"] = "\033[0;94m"
- colors["bold_blue"] = "\033[1;94m"
- colors["reset"] = "\033[0m"
- else:
- colors["blue"] = ""
- colors["bold_blue"] = ""
- colors["reset"] = ""
- # There is a space before "..." to ensure that source file names can be
- # Ctrl + clicked in the VS Code terminal.
- compile_source_message = "{}Compiling {}$SOURCE{} ...{}".format(
- colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"]
- )
- java_compile_source_message = "{}Compiling {}$SOURCE{} ...{}".format(
- colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"]
- )
- compile_shared_source_message = "{}Compiling shared {}$SOURCE{} ...{}".format(
- colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"]
- )
- link_program_message = "{}Linking Program {}$TARGET{} ...{}".format(
- colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"]
- )
- link_library_message = "{}Linking Static Library {}$TARGET{} ...{}".format(
- colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"]
- )
- ranlib_library_message = "{}Ranlib Library {}$TARGET{} ...{}".format(
- colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"]
- )
- link_shared_library_message = "{}Linking Shared Library {}$TARGET{} ...{}".format(
- colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"]
- )
- java_library_message = "{}Creating Java Archive {}$TARGET{} ...{}".format(
- colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"]
- )
- env.Append(CXXCOMSTR=[compile_source_message])
- env.Append(CCCOMSTR=[compile_source_message])
- env.Append(SHCCCOMSTR=[compile_shared_source_message])
- env.Append(SHCXXCOMSTR=[compile_shared_source_message])
- env.Append(ARCOMSTR=[link_library_message])
- env.Append(RANLIBCOMSTR=[ranlib_library_message])
- env.Append(SHLINKCOMSTR=[link_shared_library_message])
- env.Append(LINKCOMSTR=[link_program_message])
- env.Append(JARCOMSTR=[java_library_message])
- env.Append(JAVACCOMSTR=[java_compile_source_message])
- def detect_visual_c_compiler_version(tools_env):
- # tools_env is the variable scons uses to call tools that execute tasks, SCons's env['ENV'] that executes tasks...
- # (see the SCons documentation for more information on what it does)...
- # in order for this function to be well encapsulated i choose to force it to receive SCons's TOOLS env (env['ENV']
- # and not scons setup environment (env)... so make sure you call the right environment on it or it will fail to detect
- # the proper vc version that will be called
- # There is no flag to give to visual c compilers to set the architecture, i.e. scons arch argument (x86_32, x86_64, arm64, etc.).
- # There are many different cl.exe files that are run, and each one compiles & links to a different architecture
- # As far as I know, the only way to figure out what compiler will be run when Scons calls cl.exe via Program()
- # is to check the PATH variable and figure out which one will be called first. Code below does that and returns:
- # the following string values:
- # "" Compiler not detected
- # "amd64" Native 64 bit compiler
- # "amd64_x86" 64 bit Cross Compiler for 32 bit
- # "x86" Native 32 bit compiler
- # "x86_amd64" 32 bit Cross Compiler for 64 bit
- # There are other architectures, but Godot does not support them currently, so this function does not detect arm/amd64_arm
- # and similar architectures/compilers
- # Set chosen compiler to "not detected"
- vc_chosen_compiler_index = -1
- vc_chosen_compiler_str = ""
- # Start with Pre VS 2017 checks which uses VCINSTALLDIR:
- if "VCINSTALLDIR" in tools_env:
- # print("Checking VCINSTALLDIR")
- # find() works with -1 so big ifs below are needed... the simplest solution, in fact
- # First test if amd64 and amd64_x86 compilers are present in the path
- vc_amd64_compiler_detection_index = tools_env["PATH"].find(tools_env["VCINSTALLDIR"] + "BIN\\amd64;")
- if vc_amd64_compiler_detection_index > -1:
- vc_chosen_compiler_index = vc_amd64_compiler_detection_index
- vc_chosen_compiler_str = "amd64"
- vc_amd64_x86_compiler_detection_index = tools_env["PATH"].find(tools_env["VCINSTALLDIR"] + "BIN\\amd64_x86;")
- if vc_amd64_x86_compiler_detection_index > -1 and (
- vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_amd64_x86_compiler_detection_index
- ):
- vc_chosen_compiler_index = vc_amd64_x86_compiler_detection_index
- vc_chosen_compiler_str = "amd64_x86"
- # Now check the 32 bit compilers
- vc_x86_compiler_detection_index = tools_env["PATH"].find(tools_env["VCINSTALLDIR"] + "BIN;")
- if vc_x86_compiler_detection_index > -1 and (
- vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_x86_compiler_detection_index
- ):
- vc_chosen_compiler_index = vc_x86_compiler_detection_index
- vc_chosen_compiler_str = "x86"
- vc_x86_amd64_compiler_detection_index = tools_env["PATH"].find(tools_env["VCINSTALLDIR"] + "BIN\\x86_amd64;")
- if vc_x86_amd64_compiler_detection_index > -1 and (
- vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_x86_amd64_compiler_detection_index
- ):
- vc_chosen_compiler_index = vc_x86_amd64_compiler_detection_index
- vc_chosen_compiler_str = "x86_amd64"
- # and for VS 2017 and newer we check VCTOOLSINSTALLDIR:
- if "VCTOOLSINSTALLDIR" in tools_env:
- # Newer versions have a different path available
- vc_amd64_compiler_detection_index = (
- tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX64\\X64;")
- )
- if vc_amd64_compiler_detection_index > -1:
- vc_chosen_compiler_index = vc_amd64_compiler_detection_index
- vc_chosen_compiler_str = "amd64"
- vc_amd64_x86_compiler_detection_index = (
- tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX64\\X86;")
- )
- if vc_amd64_x86_compiler_detection_index > -1 and (
- vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_amd64_x86_compiler_detection_index
- ):
- vc_chosen_compiler_index = vc_amd64_x86_compiler_detection_index
- vc_chosen_compiler_str = "amd64_x86"
- vc_x86_compiler_detection_index = (
- tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX86\\X86;")
- )
- if vc_x86_compiler_detection_index > -1 and (
- vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_x86_compiler_detection_index
- ):
- vc_chosen_compiler_index = vc_x86_compiler_detection_index
- vc_chosen_compiler_str = "x86"
- vc_x86_amd64_compiler_detection_index = (
- tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX86\\X64;")
- )
- if vc_x86_amd64_compiler_detection_index > -1 and (
- vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_x86_amd64_compiler_detection_index
- ):
- vc_chosen_compiler_str = "x86_amd64"
- return vc_chosen_compiler_str
- def find_visual_c_batch_file(env):
- from SCons.Tool.MSCommon.vc import (
- get_default_version,
- get_host_target,
- find_batch_file,
- )
- # Syntax changed in SCons 4.4.0.
- from SCons import __version__ as scons_raw_version
- scons_ver = env._get_major_minor_revision(scons_raw_version)
- version = get_default_version(env)
- if scons_ver >= (4, 4, 0):
- (host_platform, target_platform, _) = get_host_target(env, version)
- else:
- (host_platform, target_platform, _) = get_host_target(env)
- return find_batch_file(env, version, host_platform, target_platform)[0]
- def generate_cpp_hint_file(filename):
- if os.path.isfile(filename):
- # Don't overwrite an existing hint file since the user may have customized it.
- pass
- else:
- try:
- with open(filename, "w") as fd:
- fd.write("#define GDCLASS(m_class, m_inherits)\n")
- except OSError:
- print("Could not write cpp.hint file.")
- def glob_recursive(pattern, node="."):
- from SCons import Node
- from SCons.Script import Glob
- results = []
- for f in Glob(str(node) + "/*", source=True):
- if type(f) is Node.FS.Dir:
- results += glob_recursive(pattern, f)
- results += Glob(str(node) + "/" + pattern, source=True)
- return results
- def add_to_vs_project(env, sources):
- for x in sources:
- if type(x) == type(""):
- fname = env.File(x).path
- else:
- fname = env.File(x)[0].path
- pieces = fname.split(".")
- if len(pieces) > 0:
- basename = pieces[0]
- basename = basename.replace("\\\\", "/")
- if os.path.isfile(basename + ".h"):
- env.vs_incs += [basename + ".h"]
- elif os.path.isfile(basename + ".hpp"):
- env.vs_incs += [basename + ".hpp"]
- if os.path.isfile(basename + ".c"):
- env.vs_srcs += [basename + ".c"]
- elif os.path.isfile(basename + ".cpp"):
- env.vs_srcs += [basename + ".cpp"]
- def generate_vs_project(env, original_args, project_name="godot"):
- batch_file = find_visual_c_batch_file(env)
- filtered_args = original_args.copy()
- # Ignore the "vsproj" option to not regenerate the VS project on every build
- filtered_args.pop("vsproj", None)
- # The "platform" option is ignored because only the Windows platform is currently supported for VS projects
- filtered_args.pop("platform", None)
- # The "target" option is ignored due to the way how targets configuration is performed for VS projects (there is a separate project configuration for each target)
- filtered_args.pop("target", None)
- # The "progress" option is ignored as the current compilation progress indication doesn't work in VS
- filtered_args.pop("progress", None)
- if batch_file:
- class ModuleConfigs(Mapping):
- # This version information (Win32, x64, Debug, Release) seems to be
- # required for Visual Studio to understand that it needs to generate an NMAKE
- # project. Do not modify without knowing what you are doing.
- PLATFORMS = ["Win32", "x64"]
- PLATFORM_IDS = ["x86_32", "x86_64"]
- CONFIGURATIONS = ["editor", "template_release", "template_debug"]
- DEV_SUFFIX = ".dev" if env["dev_build"] else ""
- @staticmethod
- def for_every_variant(value):
- return [value for _ in range(len(ModuleConfigs.CONFIGURATIONS) * len(ModuleConfigs.PLATFORMS))]
- def __init__(self):
- shared_targets_array = []
- self.names = []
- self.arg_dict = {
- "variant": [],
- "runfile": shared_targets_array,
- "buildtarget": shared_targets_array,
- "cpppaths": [],
- "cppdefines": [],
- "cmdargs": [],
- }
- self.add_mode() # default
- def add_mode(
- self,
- name: str = "",
- includes: str = "",
- cli_args: str = "",
- defines=None,
- ):
- if defines is None:
- defines = []
- self.names.append(name)
- self.arg_dict["variant"] += [
- f'{config}{f"_[{name}]" if name else ""}|{platform}'
- for config in ModuleConfigs.CONFIGURATIONS
- for platform in ModuleConfigs.PLATFORMS
- ]
- self.arg_dict["runfile"] += [
- f'bin\\godot.windows.{config}{ModuleConfigs.DEV_SUFFIX}{".double" if env["precision"] == "double" else ""}.{plat_id}{f".{name}" if name else ""}.exe'
- for config in ModuleConfigs.CONFIGURATIONS
- for plat_id in ModuleConfigs.PLATFORM_IDS
- ]
- self.arg_dict["cpppaths"] += ModuleConfigs.for_every_variant(env["CPPPATH"] + [includes])
- self.arg_dict["cppdefines"] += ModuleConfigs.for_every_variant(list(env["CPPDEFINES"]) + defines)
- self.arg_dict["cmdargs"] += ModuleConfigs.for_every_variant(cli_args)
- def build_commandline(self, commands):
- configuration_getter = (
- "$(Configuration"
- + "".join([f'.Replace("{name}", "")' for name in self.names[1:]])
- + '.Replace("_[]", "")'
- + ")"
- )
- common_build_prefix = [
- 'cmd /V /C set "plat=$(PlatformTarget)"',
- '(if "$(PlatformTarget)"=="x64" (set "plat=x86_amd64"))',
- 'call "' + batch_file + '" !plat!',
- ]
- # Windows allows us to have spaces in paths, so we need
- # to double quote off the directory. However, the path ends
- # in a backslash, so we need to remove this, lest it escape the
- # last double quote off, confusing MSBuild
- common_build_postfix = [
- "--directory=\"$(ProjectDir.TrimEnd('\\'))\"",
- "platform=windows",
- f"target={configuration_getter}",
- "progress=no",
- ]
- for arg, value in filtered_args.items():
- common_build_postfix.append(f"{arg}={value}")
- result = " ^& ".join(common_build_prefix + [" ".join([commands] + common_build_postfix)])
- return result
- # Mappings interface definitions
- def __iter__(self) -> Iterator[str]:
- for x in self.arg_dict:
- yield x
- def __len__(self) -> int:
- return len(self.names)
- def __getitem__(self, k: str):
- return self.arg_dict[k]
- add_to_vs_project(env, env.core_sources)
- add_to_vs_project(env, env.drivers_sources)
- add_to_vs_project(env, env.main_sources)
- add_to_vs_project(env, env.modules_sources)
- add_to_vs_project(env, env.scene_sources)
- add_to_vs_project(env, env.servers_sources)
- if env["tests"]:
- add_to_vs_project(env, env.tests_sources)
- if env.editor_build:
- add_to_vs_project(env, env.editor_sources)
- for header in glob_recursive("**/*.h"):
- env.vs_incs.append(str(header))
- module_configs = ModuleConfigs()
- if env.get("module_mono_enabled"):
- mono_defines = [("GD_MONO_HOT_RELOAD",)] if env.editor_build else []
- module_configs.add_mode(
- "mono",
- cli_args="module_mono_enabled=yes",
- defines=mono_defines,
- )
- scons_cmd = "scons"
- path_to_venv = os.getenv("VIRTUAL_ENV")
- path_to_scons_exe = Path(str(path_to_venv)) / "Scripts" / "scons.exe"
- if path_to_venv and path_to_scons_exe.exists():
- scons_cmd = str(path_to_scons_exe)
- env["MSVSBUILDCOM"] = module_configs.build_commandline(scons_cmd)
- env["MSVSREBUILDCOM"] = module_configs.build_commandline(f"{scons_cmd} vsproj=yes")
- env["MSVSCLEANCOM"] = module_configs.build_commandline(f"{scons_cmd} --clean")
- if not env.get("MSVS"):
- env["MSVS"]["PROJECTSUFFIX"] = ".vcxproj"
- env["MSVS"]["SOLUTIONSUFFIX"] = ".sln"
- env.MSVSProject(
- target=["#" + project_name + env["MSVSPROJECTSUFFIX"]],
- incs=env.vs_incs,
- srcs=env.vs_srcs,
- auto_build_solution=1,
- **module_configs,
- )
- else:
- print("Could not locate Visual Studio batch file to set up the build environment. Not generating VS project.")
- def precious_program(env, program, sources, **args):
- program = env.ProgramOriginal(program, sources, **args)
- env.Precious(program)
- return program
- def add_shared_library(env, name, sources, **args):
- library = env.SharedLibrary(name, sources, **args)
- env.NoCache(library)
- return library
- def add_library(env, name, sources, **args):
- library = env.Library(name, sources, **args)
- env.NoCache(library)
- return library
- def add_program(env, name, sources, **args):
- program = env.Program(name, sources, **args)
- env.NoCache(program)
- return program
- def CommandNoCache(env, target, sources, command, **args):
- result = env.Command(target, sources, command, **args)
- env.NoCache(result)
- return result
- def Run(env, function, short_message, subprocess=True):
- from SCons.Script import Action
- from platform_methods import run_in_subprocess
- output_print = short_message if not env["verbose"] else ""
- if not subprocess:
- return Action(function, output_print)
- else:
- return Action(run_in_subprocess(function), output_print)
- def detect_darwin_sdk_path(platform, env):
- sdk_name = ""
- if platform == "macos":
- sdk_name = "macosx"
- var_name = "MACOS_SDK_PATH"
- elif platform == "ios":
- sdk_name = "iphoneos"
- var_name = "IOS_SDK_PATH"
- elif platform == "iossimulator":
- sdk_name = "iphonesimulator"
- var_name = "IOS_SDK_PATH"
- else:
- raise Exception("Invalid platform argument passed to detect_darwin_sdk_path")
- if not env[var_name]:
- try:
- sdk_path = subprocess.check_output(["xcrun", "--sdk", sdk_name, "--show-sdk-path"]).strip().decode("utf-8")
- if sdk_path:
- env[var_name] = sdk_path
- except (subprocess.CalledProcessError, OSError):
- print("Failed to find SDK path while running xcrun --sdk {} --show-sdk-path.".format(sdk_name))
- raise
- def is_vanilla_clang(env):
- if not using_clang(env):
- return False
- try:
- version = subprocess.check_output([env.subst(env["CXX"]), "--version"]).strip().decode("utf-8")
- except (subprocess.CalledProcessError, OSError):
- print("Couldn't parse CXX environment variable to infer compiler version.")
- return False
- return not version.startswith("Apple")
- def get_compiler_version(env):
- """
- Returns a dictionary with various version information:
- - major, minor, patch: Version following semantic versioning system
- - metadata1, metadata2: Extra information
- - date: Date of the build
- """
- ret = {
- "major": -1,
- "minor": -1,
- "patch": -1,
- "metadata1": None,
- "metadata2": None,
- "date": None,
- }
- if not env.msvc:
- # Not using -dumpversion as some GCC distros only return major, and
- # Clang used to return hardcoded 4.2.1: # https://reviews.llvm.org/D56803
- try:
- version = subprocess.check_output([env.subst(env["CXX"]), "--version"]).strip().decode("utf-8")
- except (subprocess.CalledProcessError, OSError):
- print("Couldn't parse CXX environment variable to infer compiler version.")
- return ret
- else:
- # TODO: Implement for MSVC
- return ret
- match = re.search(
- r"(?:(?<=version )|(?<=\) )|(?<=^))"
- r"(?P<major>\d+)"
- r"(?:\.(?P<minor>\d*))?"
- r"(?:\.(?P<patch>\d*))?"
- r"(?:-(?P<metadata1>[0-9a-zA-Z-]*))?"
- r"(?:\+(?P<metadata2>[0-9a-zA-Z-]*))?"
- r"(?: (?P<date>[0-9]{8}|[0-9]{6})(?![0-9a-zA-Z]))?",
- version,
- )
- if match is not None:
- for key, value in match.groupdict().items():
- if value is not None:
- ret[key] = value
- # Transform semantic versioning to integers
- for key in ["major", "minor", "patch"]:
- ret[key] = int(ret[key] or -1)
- return ret
- def using_gcc(env):
- return "gcc" in os.path.basename(env["CC"])
- def using_clang(env):
- return "clang" in os.path.basename(env["CC"])
- def using_emcc(env):
- return "emcc" in os.path.basename(env["CC"])
- def show_progress(env):
- import sys
- from SCons.Script import Progress, Command, AlwaysBuild
- screen = sys.stdout
- # Progress reporting is not available in non-TTY environments since it
- # messes with the output (for example, when writing to a file)
- show_progress = env["progress"] and sys.stdout.isatty()
- node_count = 0
- node_count_max = 0
- node_count_interval = 1
- node_count_fname = str(env.Dir("#")) + "/.scons_node_count"
- import time, math
- class cache_progress:
- # The default is 1 GB cache and 12 hours half life
- def __init__(self, path=None, limit=1073741824, half_life=43200):
- self.path = path
- self.limit = limit
- self.exponent_scale = math.log(2) / half_life
- if env["verbose"] and path != None:
- screen.write(
- "Current cache limit is {} (used: {})\n".format(
- self.convert_size(limit), self.convert_size(self.get_size(path))
- )
- )
- self.delete(self.file_list())
- def __call__(self, node, *args, **kw):
- nonlocal node_count, node_count_max, node_count_interval, node_count_fname, show_progress
- if show_progress:
- # Print the progress percentage
- node_count += node_count_interval
- if node_count_max > 0 and node_count <= node_count_max:
- screen.write("\r[%3d%%] " % (node_count * 100 / node_count_max))
- screen.flush()
- elif node_count_max > 0 and node_count > node_count_max:
- screen.write("\r[100%] ")
- screen.flush()
- else:
- screen.write("\r[Initial build] ")
- screen.flush()
- def delete(self, files):
- if len(files) == 0:
- return
- if env["verbose"]:
- # Utter something
- screen.write("\rPurging %d %s from cache...\n" % (len(files), len(files) > 1 and "files" or "file"))
- [os.remove(f) for f in files]
- def file_list(self):
- if self.path is None:
- # Nothing to do
- return []
- # Gather a list of (filename, (size, atime)) within the
- # cache directory
- file_stat = [(x, os.stat(x)[6:8]) for x in glob.glob(os.path.join(self.path, "*", "*"))]
- if file_stat == []:
- # Nothing to do
- return []
- # Weight the cache files by size (assumed to be roughly
- # proportional to the recompilation time) times an exponential
- # decay since the ctime, and return a list with the entries
- # (filename, size, weight).
- current_time = time.time()
- file_stat = [(x[0], x[1][0], (current_time - x[1][1])) for x in file_stat]
- # Sort by the most recently accessed files (most sensible to keep) first
- file_stat.sort(key=lambda x: x[2])
- # Search for the first entry where the storage limit is
- # reached
- sum, mark = 0, None
- for i, x in enumerate(file_stat):
- sum += x[1]
- if sum > self.limit:
- mark = i
- break
- if mark is None:
- return []
- else:
- return [x[0] for x in file_stat[mark:]]
- def convert_size(self, size_bytes):
- if size_bytes == 0:
- return "0 bytes"
- size_name = ("bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB")
- i = int(math.floor(math.log(size_bytes, 1024)))
- p = math.pow(1024, i)
- s = round(size_bytes / p, 2)
- return "%s %s" % (int(s) if i == 0 else s, size_name[i])
- def get_size(self, start_path="."):
- total_size = 0
- for dirpath, dirnames, filenames in os.walk(start_path):
- for f in filenames:
- fp = os.path.join(dirpath, f)
- total_size += os.path.getsize(fp)
- return total_size
- def progress_finish(target, source, env):
- nonlocal node_count, progressor
- try:
- with open(node_count_fname, "w") as f:
- f.write("%d\n" % node_count)
- progressor.delete(progressor.file_list())
- except Exception:
- pass
- try:
- with open(node_count_fname) as f:
- node_count_max = int(f.readline())
- except Exception:
- pass
- cache_directory = os.environ.get("SCONS_CACHE")
- # Simple cache pruning, attached to SCons' progress callback. Trim the
- # cache directory to a size not larger than cache_limit.
- cache_limit = float(os.getenv("SCONS_CACHE_LIMIT", 1024)) * 1024 * 1024
- progressor = cache_progress(cache_directory, cache_limit)
- Progress(progressor, interval=node_count_interval)
- progress_finish_command = Command("progress_finish", [], progress_finish)
- AlwaysBuild(progress_finish_command)
- def dump(env):
- # Dumps latest build information for debugging purposes and external tools.
- from json import dump
- def non_serializable(obj):
- return "<<non-serializable: %s>>" % (type(obj).__qualname__)
- with open(".scons_env.json", "w") as f:
- dump(env.Dictionary(), f, indent=4, default=non_serializable)
|