Change to actual extra_script.py
This commit is contained in:
parent
ec7e1a35e7
commit
c614cae472
|
@ -10,7 +10,7 @@ from datetime import datetime
|
||||||
import re
|
import re
|
||||||
import pprint
|
import pprint
|
||||||
from platformio.project.config import ProjectConfig
|
from platformio.project.config import ProjectConfig
|
||||||
|
from platformio.project.exception import InvalidProjectConfError
|
||||||
|
|
||||||
Import("env")
|
Import("env")
|
||||||
#print(env.Dump())
|
#print(env.Dump())
|
||||||
|
@ -104,7 +104,18 @@ def writeFileIfChanged(fileName,data):
|
||||||
return True
|
return True
|
||||||
|
|
||||||
def mergeConfig(base,other):
|
def mergeConfig(base,other):
|
||||||
for cname in other:
|
try:
|
||||||
|
customconfig = env.GetProjectOption("custom_config")
|
||||||
|
except InvalidProjectConfError:
|
||||||
|
customconfig = None
|
||||||
|
for bdir in other:
|
||||||
|
if customconfig and os.path.exists(os.path.join(bdir,customconfig)):
|
||||||
|
cname=os.path.join(bdir,customconfig)
|
||||||
|
print("merge custom config {}".format(cname))
|
||||||
|
with open(cname,'rb') as ah:
|
||||||
|
base += json.load(ah)
|
||||||
|
continue
|
||||||
|
cname=os.path.join(bdir,"config.json")
|
||||||
if os.path.exists(cname):
|
if os.path.exists(cname):
|
||||||
print("merge config %s"%cname)
|
print("merge config %s"%cname)
|
||||||
with open(cname,'rb') as ah:
|
with open(cname,'rb') as ah:
|
||||||
|
@ -150,25 +161,13 @@ def expandConfig(config):
|
||||||
rt.append(replaceTexts(c,replace))
|
rt.append(replaceTexts(c,replace))
|
||||||
return rt
|
return rt
|
||||||
|
|
||||||
def createUserItemList(dirs,itemName,files):
|
def generateMergedConfig(inFile,outFile,addDirs=[]):
|
||||||
rt=[]
|
|
||||||
for d in dirs:
|
|
||||||
iname=os.path.join(d,itemName)
|
|
||||||
if os.path.exists(iname):
|
|
||||||
rt.append(iname)
|
|
||||||
for f in files:
|
|
||||||
if not os.path.exists(f):
|
|
||||||
raise Exception("user item %s not found"%f)
|
|
||||||
rt.append(f)
|
|
||||||
return rt
|
|
||||||
|
|
||||||
def generateMergedConfig(inFile,outFile,addFiles=[]):
|
|
||||||
if not os.path.exists(inFile):
|
if not os.path.exists(inFile):
|
||||||
raise Exception("unable to read cfg file %s"%inFile)
|
raise Exception("unable to read cfg file %s"%inFile)
|
||||||
data=""
|
data=""
|
||||||
with open(inFile,'rb') as ch:
|
with open(inFile,'rb') as ch:
|
||||||
config=json.load(ch)
|
config=json.load(ch)
|
||||||
config=mergeConfig(config,addFiles)
|
config=mergeConfig(config,addDirs)
|
||||||
config=expandConfig(config)
|
config=expandConfig(config)
|
||||||
data=json.dumps(config,indent=2)
|
data=json.dumps(config,indent=2)
|
||||||
writeFileIfChanged(outFile,data)
|
writeFileIfChanged(outFile,data)
|
||||||
|
@ -388,7 +387,12 @@ def getLibs():
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def joinFiles(target,flist):
|
def joinFiles(target,pattern,dirlist):
|
||||||
|
flist=[]
|
||||||
|
for dir in dirlist:
|
||||||
|
fn=os.path.join(dir,pattern)
|
||||||
|
if os.path.exists(fn):
|
||||||
|
flist.append(fn)
|
||||||
current=False
|
current=False
|
||||||
if os.path.exists(target):
|
if os.path.exists(target):
|
||||||
current=True
|
current=True
|
||||||
|
@ -459,28 +463,7 @@ def handleDeps(env):
|
||||||
)
|
)
|
||||||
env.AddBuildMiddleware(injectIncludes)
|
env.AddBuildMiddleware(injectIncludes)
|
||||||
|
|
||||||
def getOption(env,name,toArray=True):
|
|
||||||
try:
|
|
||||||
opt=env.GetProjectOption(name)
|
|
||||||
if toArray:
|
|
||||||
if opt is None:
|
|
||||||
return []
|
|
||||||
if isinstance(opt,list):
|
|
||||||
return opt
|
|
||||||
return opt.split("\n" if "\n" in opt else ",")
|
|
||||||
return opt
|
|
||||||
except:
|
|
||||||
pass
|
|
||||||
if toArray:
|
|
||||||
return []
|
|
||||||
|
|
||||||
def getFileList(files):
|
|
||||||
base=basePath()
|
|
||||||
rt=[]
|
|
||||||
for f in files:
|
|
||||||
if f is not None and f != "":
|
|
||||||
rt.append(os.path.join(base,f))
|
|
||||||
return rt
|
|
||||||
def prebuild(env):
|
def prebuild(env):
|
||||||
global userTaskDirs
|
global userTaskDirs
|
||||||
print("#prebuild running")
|
print("#prebuild running")
|
||||||
|
@ -490,18 +473,14 @@ def prebuild(env):
|
||||||
if ldf_mode == 'off':
|
if ldf_mode == 'off':
|
||||||
print("##ldf off - own dependency handling")
|
print("##ldf off - own dependency handling")
|
||||||
handleDeps(env)
|
handleDeps(env)
|
||||||
extraConfigs=getOption(env,'custom_config',toArray=True)
|
|
||||||
extraJs=getOption(env,'custom_js',toArray=True)
|
|
||||||
extraCss=getOption(env,'custom_css',toArray=True)
|
|
||||||
|
|
||||||
userTaskDirs=getUserTaskDirs()
|
userTaskDirs=getUserTaskDirs()
|
||||||
mergedConfig=os.path.join(outPath(),os.path.basename(CFG_FILE))
|
mergedConfig=os.path.join(outPath(),os.path.basename(CFG_FILE))
|
||||||
generateMergedConfig(os.path.join(basePath(),CFG_FILE),mergedConfig,createUserItemList(userTaskDirs,"config.json", getFileList(extraConfigs)))
|
generateMergedConfig(os.path.join(basePath(),CFG_FILE),mergedConfig,userTaskDirs)
|
||||||
compressFile(mergedConfig,mergedConfig+".gz")
|
compressFile(mergedConfig,mergedConfig+".gz")
|
||||||
generateCfg(mergedConfig,os.path.join(outPath(),CFG_INCLUDE),False)
|
generateCfg(mergedConfig,os.path.join(outPath(),CFG_INCLUDE),False)
|
||||||
generateCfg(mergedConfig,os.path.join(outPath(),CFG_INCLUDE_IMPL),True)
|
generateCfg(mergedConfig,os.path.join(outPath(),CFG_INCLUDE_IMPL),True)
|
||||||
joinFiles(os.path.join(outPath(),INDEXJS+".gz"),createUserItemList(["web"]+userTaskDirs,INDEXJS,getFileList(extraJs)))
|
joinFiles(os.path.join(outPath(),INDEXJS+".gz"),INDEXJS,["web"]+userTaskDirs)
|
||||||
joinFiles(os.path.join(outPath(),INDEXCSS+".gz"),createUserItemList(["web"]+userTaskDirs,INDEXCSS,getFileList(extraCss)))
|
joinFiles(os.path.join(outPath(),INDEXCSS+".gz"),INDEXCSS,["web"]+userTaskDirs)
|
||||||
embedded=getEmbeddedFiles(env)
|
embedded=getEmbeddedFiles(env)
|
||||||
filedefs=[]
|
filedefs=[]
|
||||||
for ef in embedded:
|
for ef in embedded:
|
||||||
|
|
|
@ -1,528 +0,0 @@
|
||||||
print("running extra...")
|
|
||||||
import gzip
|
|
||||||
import shutil
|
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
import inspect
|
|
||||||
import json
|
|
||||||
import glob
|
|
||||||
from datetime import datetime
|
|
||||||
import re
|
|
||||||
import pprint
|
|
||||||
from platformio.project.config import ProjectConfig
|
|
||||||
from platformio.project.exception import InvalidProjectConfError
|
|
||||||
|
|
||||||
Import("env")
|
|
||||||
#print(env.Dump())
|
|
||||||
OWN_FILE="extra_script.py"
|
|
||||||
GEN_DIR='lib/generated'
|
|
||||||
CFG_FILE='web/config.json'
|
|
||||||
XDR_FILE='web/xdrconfig.json'
|
|
||||||
INDEXJS="index.js"
|
|
||||||
INDEXCSS="index.css"
|
|
||||||
CFG_INCLUDE='GwConfigDefinitions.h'
|
|
||||||
CFG_INCLUDE_IMPL='GwConfigDefImpl.h'
|
|
||||||
XDR_INCLUDE='GwXdrTypeMappings.h'
|
|
||||||
TASK_INCLUDE='GwUserTasks.h'
|
|
||||||
GROVE_CONFIG="GwM5GroveGen.h"
|
|
||||||
GROVE_CONFIG_IN="lib/hardware/GwM5Grove.in"
|
|
||||||
EMBEDDED_INCLUDE="GwEmbeddedFiles.h"
|
|
||||||
|
|
||||||
def getEmbeddedFiles(env):
|
|
||||||
rt=[]
|
|
||||||
efiles=env.GetProjectOption("board_build.embed_files")
|
|
||||||
for f in efiles.split("\n"):
|
|
||||||
if f == '':
|
|
||||||
continue
|
|
||||||
rt.append(f)
|
|
||||||
return rt
|
|
||||||
|
|
||||||
def basePath():
|
|
||||||
#see: https://stackoverflow.com/questions/16771894/python-nameerror-global-name-file-is-not-defined
|
|
||||||
return os.path.dirname(inspect.getfile(lambda: None))
|
|
||||||
|
|
||||||
def outPath():
|
|
||||||
return os.path.join(basePath(),GEN_DIR)
|
|
||||||
def checkDir():
|
|
||||||
dn=outPath()
|
|
||||||
if not os.path.exists(dn):
|
|
||||||
os.makedirs(dn)
|
|
||||||
if not os.path.isdir(dn):
|
|
||||||
print("unable to create %s"%dn)
|
|
||||||
return False
|
|
||||||
return True
|
|
||||||
|
|
||||||
def isCurrent(infile,outfile):
|
|
||||||
if os.path.exists(outfile):
|
|
||||||
otime=os.path.getmtime(outfile)
|
|
||||||
itime=os.path.getmtime(infile)
|
|
||||||
if (otime >= itime):
|
|
||||||
own=os.path.join(basePath(),OWN_FILE)
|
|
||||||
if os.path.exists(own):
|
|
||||||
owntime=os.path.getmtime(own)
|
|
||||||
if owntime > otime:
|
|
||||||
return False
|
|
||||||
print("%s is newer then %s, no need to recreate"%(outfile,infile))
|
|
||||||
return True
|
|
||||||
return False
|
|
||||||
def compressFile(inFile,outfile):
|
|
||||||
if isCurrent(inFile,outfile):
|
|
||||||
return
|
|
||||||
print("compressing %s"%inFile)
|
|
||||||
with open(inFile, 'rb') as f_in:
|
|
||||||
with gzip.open(outfile, 'wb') as f_out:
|
|
||||||
shutil.copyfileobj(f_in, f_out)
|
|
||||||
|
|
||||||
def generateFile(infile,outfile,callback,inMode='rb',outMode='w'):
|
|
||||||
if isCurrent(infile,outfile):
|
|
||||||
return
|
|
||||||
print("creating %s"%outfile)
|
|
||||||
oh=None
|
|
||||||
with open(infile,inMode) as ch:
|
|
||||||
with open(outfile,outMode) as oh:
|
|
||||||
try:
|
|
||||||
callback(ch,oh,inFile=infile)
|
|
||||||
oh.close()
|
|
||||||
except Exception as e:
|
|
||||||
try:
|
|
||||||
oh.close()
|
|
||||||
except:
|
|
||||||
pass
|
|
||||||
os.unlink(outfile)
|
|
||||||
raise
|
|
||||||
|
|
||||||
def writeFileIfChanged(fileName,data):
|
|
||||||
if os.path.exists(fileName):
|
|
||||||
with open(fileName,"r") as ih:
|
|
||||||
old=ih.read()
|
|
||||||
ih.close()
|
|
||||||
if old == data:
|
|
||||||
return False
|
|
||||||
print("#generating %s"%fileName)
|
|
||||||
with open(fileName,"w") as oh:
|
|
||||||
oh.write(data)
|
|
||||||
return True
|
|
||||||
|
|
||||||
def mergeConfig(base,other):
|
|
||||||
try:
|
|
||||||
customconfig = env.GetProjectOption("custom_config")
|
|
||||||
except InvalidProjectConfError:
|
|
||||||
customconfig = None
|
|
||||||
for bdir in other:
|
|
||||||
if customconfig and os.path.exists(os.path.join(bdir,customconfig)):
|
|
||||||
cname=os.path.join(bdir,customconfig)
|
|
||||||
print("merge custom config {}".format(cname))
|
|
||||||
with open(cname,'rb') as ah:
|
|
||||||
base += json.load(ah)
|
|
||||||
continue
|
|
||||||
cname=os.path.join(bdir,"config.json")
|
|
||||||
if os.path.exists(cname):
|
|
||||||
print("merge config %s"%cname)
|
|
||||||
with open(cname,'rb') as ah:
|
|
||||||
merge=json.load(ah)
|
|
||||||
base=base+merge
|
|
||||||
return base
|
|
||||||
|
|
||||||
def replaceTexts(data,replacements):
|
|
||||||
if replacements is None:
|
|
||||||
return data
|
|
||||||
if isinstance(data,str):
|
|
||||||
for k,v in replacements.items():
|
|
||||||
data=data.replace("$"+k,str(v))
|
|
||||||
return data
|
|
||||||
if isinstance(data,list):
|
|
||||||
rt=[]
|
|
||||||
for e in data:
|
|
||||||
rt.append(replaceTexts(e,replacements))
|
|
||||||
return rt
|
|
||||||
if isinstance(data,dict):
|
|
||||||
rt={}
|
|
||||||
for k,v in data.items():
|
|
||||||
rt[replaceTexts(k,replacements)]=replaceTexts(v,replacements)
|
|
||||||
return rt
|
|
||||||
return data
|
|
||||||
def expandConfig(config):
|
|
||||||
rt=[]
|
|
||||||
for item in config:
|
|
||||||
type=item.get('type')
|
|
||||||
if type != 'array':
|
|
||||||
rt.append(item)
|
|
||||||
continue
|
|
||||||
replacements=item.get('replace')
|
|
||||||
children=item.get('children')
|
|
||||||
name=item.get('name')
|
|
||||||
if name is None:
|
|
||||||
name="#unknown#"
|
|
||||||
if not isinstance(replacements,list):
|
|
||||||
raise Exception("missing replacements at array %s"%name)
|
|
||||||
for replace in replacements:
|
|
||||||
if children is not None:
|
|
||||||
for c in children:
|
|
||||||
rt.append(replaceTexts(c,replace))
|
|
||||||
return rt
|
|
||||||
|
|
||||||
def generateMergedConfig(inFile,outFile,addDirs=[]):
|
|
||||||
if not os.path.exists(inFile):
|
|
||||||
raise Exception("unable to read cfg file %s"%inFile)
|
|
||||||
data=""
|
|
||||||
with open(inFile,'rb') as ch:
|
|
||||||
config=json.load(ch)
|
|
||||||
config=mergeConfig(config,addDirs)
|
|
||||||
config=expandConfig(config)
|
|
||||||
data=json.dumps(config,indent=2)
|
|
||||||
writeFileIfChanged(outFile,data)
|
|
||||||
|
|
||||||
def generateCfg(inFile,outFile,impl):
|
|
||||||
if not os.path.exists(inFile):
|
|
||||||
raise Exception("unable to read cfg file %s"%inFile)
|
|
||||||
data=""
|
|
||||||
with open(inFile,'rb') as ch:
|
|
||||||
config=json.load(ch)
|
|
||||||
data+="//generated from %s\n"%inFile
|
|
||||||
l=len(config)
|
|
||||||
idx=0
|
|
||||||
if not impl:
|
|
||||||
data+='#include "GwConfigItem.h"\n'
|
|
||||||
data+='class GwConfigDefinitions{\n'
|
|
||||||
data+=' public:\n'
|
|
||||||
data+=' int getNumConfig() const{return %d;}\n'%(l)
|
|
||||||
for item in config:
|
|
||||||
n=item.get('name')
|
|
||||||
if n is None:
|
|
||||||
continue
|
|
||||||
if len(n) > 15:
|
|
||||||
raise Exception("%s: config names must be max 15 caracters"%n)
|
|
||||||
data+=' static constexpr const char* %s="%s";\n'%(n,n)
|
|
||||||
data+="};\n"
|
|
||||||
else:
|
|
||||||
data+='void GwConfigHandler::populateConfigs(GwConfigInterface **config){\n'
|
|
||||||
for item in config:
|
|
||||||
name=item.get('name')
|
|
||||||
if name is None:
|
|
||||||
continue
|
|
||||||
data+=' configs[%d]='%(idx)
|
|
||||||
idx+=1
|
|
||||||
secret="false";
|
|
||||||
if item.get('type') == 'password':
|
|
||||||
secret="true"
|
|
||||||
data+=" new GwConfigInterface(%s,\"%s\",%s);\n"%(name,item.get('default'),secret)
|
|
||||||
data+='}\n'
|
|
||||||
writeFileIfChanged(outFile,data)
|
|
||||||
|
|
||||||
def labelFilter(label):
|
|
||||||
return re.sub("[^a-zA-Z0-9]","",re.sub("\([0-9]*\)","",label))
|
|
||||||
def generateXdrMappings(fp,oh,inFile=''):
|
|
||||||
jdoc=json.load(fp)
|
|
||||||
oh.write("static GwXDRTypeMapping* typeMappings[]={\n")
|
|
||||||
first=True
|
|
||||||
for cat in jdoc:
|
|
||||||
item=jdoc[cat]
|
|
||||||
cid=item.get('id')
|
|
||||||
if cid is None:
|
|
||||||
continue
|
|
||||||
tc=item.get('type')
|
|
||||||
if tc is not None:
|
|
||||||
if first:
|
|
||||||
first=False
|
|
||||||
else:
|
|
||||||
oh.write(",\n")
|
|
||||||
oh.write(" new GwXDRTypeMapping(%d,0,%d) /*%s*/"%(cid,tc,cat))
|
|
||||||
fields=item.get('fields')
|
|
||||||
if fields is None:
|
|
||||||
continue
|
|
||||||
idx=0
|
|
||||||
for fe in fields:
|
|
||||||
if not isinstance(fe,dict):
|
|
||||||
continue
|
|
||||||
tc=fe.get('t')
|
|
||||||
id=fe.get('v')
|
|
||||||
if id is None:
|
|
||||||
id=idx
|
|
||||||
idx+=1
|
|
||||||
l=fe.get('l') or ''
|
|
||||||
if tc is None or id is None:
|
|
||||||
continue
|
|
||||||
if first:
|
|
||||||
first=False
|
|
||||||
else:
|
|
||||||
oh.write(",\n")
|
|
||||||
oh.write(" new GwXDRTypeMapping(%d,%d,%d) /*%s:%s*/"%(cid,id,tc,cat,l))
|
|
||||||
oh.write("\n")
|
|
||||||
oh.write("};\n")
|
|
||||||
for cat in jdoc:
|
|
||||||
item=jdoc[cat]
|
|
||||||
cid=item.get('id')
|
|
||||||
if cid is None:
|
|
||||||
continue
|
|
||||||
selectors=item.get('selector')
|
|
||||||
if selectors is not None:
|
|
||||||
for selector in selectors:
|
|
||||||
label=selector.get('l')
|
|
||||||
value=selector.get('v')
|
|
||||||
if label is not None and value is not None:
|
|
||||||
label=labelFilter(label)
|
|
||||||
define=("GWXDRSEL_%s_%s"%(cat,label)).upper()
|
|
||||||
oh.write(" #define %s %s\n"%(define,value))
|
|
||||||
fields=item.get('fields')
|
|
||||||
if fields is not None:
|
|
||||||
idx=0
|
|
||||||
for field in fields:
|
|
||||||
v=field.get('v')
|
|
||||||
if v is None:
|
|
||||||
v=idx
|
|
||||||
else:
|
|
||||||
v=int(v)
|
|
||||||
label=field.get('l')
|
|
||||||
if v is not None and label is not None:
|
|
||||||
define=("GWXDRFIELD_%s_%s"%(cat,labelFilter(label))).upper();
|
|
||||||
oh.write(" #define %s %s\n"%(define,str(v)))
|
|
||||||
idx+=1
|
|
||||||
|
|
||||||
class Grove:
|
|
||||||
def __init__(self,name) -> None:
|
|
||||||
self.name=name
|
|
||||||
def _ss(self,z=False):
|
|
||||||
if z:
|
|
||||||
return self.name
|
|
||||||
return self.name if self.name != 'Z' else ''
|
|
||||||
def _suffix(self):
|
|
||||||
return '_'+self.name if self.name != 'Z' else ''
|
|
||||||
def replace(self,line):
|
|
||||||
if line is None:
|
|
||||||
return line
|
|
||||||
return line.replace('$G$',self._ss()).replace('$Z$',self._ss(True)).replace('$GS$',self._suffix())
|
|
||||||
def generateGroveDefs(inh,outh,inFile=''):
|
|
||||||
GROVES=[Grove('Z'),Grove('A'),Grove('B'),Grove('C')]
|
|
||||||
definition=[]
|
|
||||||
started=False
|
|
||||||
def writeConfig():
|
|
||||||
for grove in GROVES:
|
|
||||||
for cl in definition:
|
|
||||||
outh.write(grove.replace(cl))
|
|
||||||
|
|
||||||
for line in inh:
|
|
||||||
if re.match(" *#GROVE",line):
|
|
||||||
started=True
|
|
||||||
if len(definition) > 0:
|
|
||||||
writeConfig()
|
|
||||||
definition=[]
|
|
||||||
continue
|
|
||||||
if started:
|
|
||||||
definition.append(line)
|
|
||||||
if len(definition) > 0:
|
|
||||||
writeConfig()
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
userTaskDirs=[]
|
|
||||||
|
|
||||||
def getUserTaskDirs():
|
|
||||||
rt=[]
|
|
||||||
taskdirs=glob.glob(os.path.join( basePath(),'lib','*task*'))
|
|
||||||
for task in taskdirs:
|
|
||||||
rt.append(task)
|
|
||||||
return rt
|
|
||||||
|
|
||||||
def checkAndAdd(file,names,ilist):
|
|
||||||
if not file.endswith('.h'):
|
|
||||||
return
|
|
||||||
match=False
|
|
||||||
for cmp in names:
|
|
||||||
#print("##check %s<->%s"%(f.lower(),cmp))
|
|
||||||
if file.lower() == cmp:
|
|
||||||
match=True
|
|
||||||
if not match:
|
|
||||||
return
|
|
||||||
ilist.append(file)
|
|
||||||
def genereateUserTasks(outfile):
|
|
||||||
includes=[]
|
|
||||||
for task in userTaskDirs:
|
|
||||||
#print("##taskdir=%s"%task)
|
|
||||||
base=os.path.basename(task)
|
|
||||||
includeNames=[base.lower()+".h",'gw'+base.lower()+'.h']
|
|
||||||
for f in os.listdir(task):
|
|
||||||
checkAndAdd(f,includeNames,includes)
|
|
||||||
includeData=""
|
|
||||||
for i in includes:
|
|
||||||
print("#task include %s"%i)
|
|
||||||
includeData+="#include <%s>\n"%i
|
|
||||||
writeFileIfChanged(outfile,includeData)
|
|
||||||
|
|
||||||
def generateEmbedded(elist,outFile):
|
|
||||||
content=""
|
|
||||||
for entry in elist:
|
|
||||||
content+="EMBED_GZ_FILE(\"%s\",%s,\"%s\");\n"%entry
|
|
||||||
writeFileIfChanged(outFile,content)
|
|
||||||
|
|
||||||
def getContentType(fn):
|
|
||||||
if (fn.endswith('.gz')):
|
|
||||||
fn=fn[0:-3]
|
|
||||||
if (fn.endswith('html')):
|
|
||||||
return "text/html"
|
|
||||||
if (fn.endswith('json')):
|
|
||||||
return "application/json"
|
|
||||||
if (fn.endswith('js')):
|
|
||||||
return "text/javascript"
|
|
||||||
if (fn.endswith('css')):
|
|
||||||
return "text/css"
|
|
||||||
return "application/octet-stream"
|
|
||||||
|
|
||||||
|
|
||||||
def getLibs():
|
|
||||||
base=os.path.join(basePath(),"lib")
|
|
||||||
rt=[]
|
|
||||||
for sd in os.listdir(base):
|
|
||||||
if sd == '..':
|
|
||||||
continue
|
|
||||||
if sd == '.':
|
|
||||||
continue
|
|
||||||
fn=os.path.join(base,sd)
|
|
||||||
if os.path.isdir(fn):
|
|
||||||
rt.append(sd)
|
|
||||||
EXTRAS=['generated']
|
|
||||||
for e in EXTRAS:
|
|
||||||
if not e in rt:
|
|
||||||
rt.append(e)
|
|
||||||
return rt
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def joinFiles(target,pattern,dirlist):
|
|
||||||
flist=[]
|
|
||||||
for dir in dirlist:
|
|
||||||
fn=os.path.join(dir,pattern)
|
|
||||||
if os.path.exists(fn):
|
|
||||||
flist.append(fn)
|
|
||||||
current=False
|
|
||||||
if os.path.exists(target):
|
|
||||||
current=True
|
|
||||||
for f in flist:
|
|
||||||
if not isCurrent(f,target):
|
|
||||||
current=False
|
|
||||||
break
|
|
||||||
if current:
|
|
||||||
print("%s is up to date"%target)
|
|
||||||
return
|
|
||||||
print("creating %s"%target)
|
|
||||||
with gzip.open(target,"wb") as oh:
|
|
||||||
for fn in flist:
|
|
||||||
print("adding %s to %s"%(fn,target))
|
|
||||||
with open(fn,"rb") as rh:
|
|
||||||
shutil.copyfileobj(rh,oh)
|
|
||||||
|
|
||||||
|
|
||||||
OWNLIBS=getLibs()+["FS","WiFi"]
|
|
||||||
GLOBAL_INCLUDES=[]
|
|
||||||
|
|
||||||
def handleDeps(env):
|
|
||||||
#overwrite the GetProjectConfig
|
|
||||||
#to inject all our libs
|
|
||||||
oldGetProjectConfig=env.GetProjectConfig
|
|
||||||
def GetProjectConfigX(env):
|
|
||||||
rt=oldGetProjectConfig()
|
|
||||||
cenv="env:"+env['PIOENV']
|
|
||||||
libs=[]
|
|
||||||
for section,options in rt.as_tuple():
|
|
||||||
if section == cenv:
|
|
||||||
for key,values in options:
|
|
||||||
if key == 'lib_deps':
|
|
||||||
libs=values
|
|
||||||
|
|
||||||
mustUpdate=False
|
|
||||||
for lib in OWNLIBS:
|
|
||||||
if not lib in libs:
|
|
||||||
libs.append(lib)
|
|
||||||
mustUpdate=True
|
|
||||||
if mustUpdate:
|
|
||||||
update=[(cenv,[('lib_deps',libs)])]
|
|
||||||
rt.update(update)
|
|
||||||
return rt
|
|
||||||
env.AddMethod(GetProjectConfigX,"GetProjectConfig")
|
|
||||||
#store the list of all includes after we resolved
|
|
||||||
#the dependencies for our main project
|
|
||||||
#we will use them for all compilations afterwards
|
|
||||||
oldLibBuilder=env.ConfigureProjectLibBuilder
|
|
||||||
def ConfigureProjectLibBuilderX(env):
|
|
||||||
global GLOBAL_INCLUDES
|
|
||||||
project=oldLibBuilder()
|
|
||||||
#print("##ConfigureProjectLibBuilderX")
|
|
||||||
#pprint.pprint(project)
|
|
||||||
if project.depbuilders:
|
|
||||||
#print("##depbuilders %s"%",".join(map(lambda x: x.path,project.depbuilders)))
|
|
||||||
for db in project.depbuilders:
|
|
||||||
idirs=db.get_include_dirs()
|
|
||||||
for id in idirs:
|
|
||||||
if not id in GLOBAL_INCLUDES:
|
|
||||||
GLOBAL_INCLUDES.append(id)
|
|
||||||
return project
|
|
||||||
env.AddMethod(ConfigureProjectLibBuilderX,"ConfigureProjectLibBuilder")
|
|
||||||
def injectIncludes(env,node):
|
|
||||||
return env.Object(
|
|
||||||
node,
|
|
||||||
CPPPATH=env["CPPPATH"]+GLOBAL_INCLUDES
|
|
||||||
)
|
|
||||||
env.AddBuildMiddleware(injectIncludes)
|
|
||||||
|
|
||||||
|
|
||||||
def prebuild(env):
|
|
||||||
global userTaskDirs
|
|
||||||
print("#prebuild running")
|
|
||||||
if not checkDir():
|
|
||||||
sys.exit(1)
|
|
||||||
ldf_mode=env.GetProjectOption("lib_ldf_mode")
|
|
||||||
if ldf_mode == 'off':
|
|
||||||
print("##ldf off - own dependency handling")
|
|
||||||
handleDeps(env)
|
|
||||||
userTaskDirs=getUserTaskDirs()
|
|
||||||
mergedConfig=os.path.join(outPath(),os.path.basename(CFG_FILE))
|
|
||||||
generateMergedConfig(os.path.join(basePath(),CFG_FILE),mergedConfig,userTaskDirs)
|
|
||||||
compressFile(mergedConfig,mergedConfig+".gz")
|
|
||||||
generateCfg(mergedConfig,os.path.join(outPath(),CFG_INCLUDE),False)
|
|
||||||
generateCfg(mergedConfig,os.path.join(outPath(),CFG_INCLUDE_IMPL),True)
|
|
||||||
joinFiles(os.path.join(outPath(),INDEXJS+".gz"),INDEXJS,["web"]+userTaskDirs)
|
|
||||||
joinFiles(os.path.join(outPath(),INDEXCSS+".gz"),INDEXCSS,["web"]+userTaskDirs)
|
|
||||||
embedded=getEmbeddedFiles(env)
|
|
||||||
filedefs=[]
|
|
||||||
for ef in embedded:
|
|
||||||
print("#checking embedded file %s"%ef)
|
|
||||||
(dn,fn)=os.path.split(ef)
|
|
||||||
pureName=fn
|
|
||||||
if pureName.endswith('.gz'):
|
|
||||||
pureName=pureName[0:-3]
|
|
||||||
ct=getContentType(pureName)
|
|
||||||
usname=ef.replace('/','_').replace('.','_')
|
|
||||||
filedefs.append((pureName,usname,ct))
|
|
||||||
inFile=os.path.join(basePath(),"web",pureName)
|
|
||||||
if os.path.exists(inFile):
|
|
||||||
compressFile(inFile,ef)
|
|
||||||
else:
|
|
||||||
print("#WARNING: infile %s for %s not found"%(inFile,ef))
|
|
||||||
generateEmbedded(filedefs,os.path.join(outPath(),EMBEDDED_INCLUDE))
|
|
||||||
genereateUserTasks(os.path.join(outPath(), TASK_INCLUDE))
|
|
||||||
generateFile(os.path.join(basePath(),XDR_FILE),os.path.join(outPath(),XDR_INCLUDE),generateXdrMappings)
|
|
||||||
generateFile(os.path.join(basePath(),GROVE_CONFIG_IN),os.path.join(outPath(),GROVE_CONFIG),generateGroveDefs,inMode='r')
|
|
||||||
version="dev"+datetime.now().strftime("%Y%m%d")
|
|
||||||
env.Append(CPPDEFINES=[('GWDEVVERSION',version)])
|
|
||||||
|
|
||||||
def cleangenerated(source, target, env):
|
|
||||||
od=outPath()
|
|
||||||
if os.path.isdir(od):
|
|
||||||
print("#cleaning up %s"%od)
|
|
||||||
for f in os.listdir(od):
|
|
||||||
if f == "." or f == "..":
|
|
||||||
continue
|
|
||||||
fn=os.path.join(od,f)
|
|
||||||
os.unlink(f)
|
|
||||||
|
|
||||||
|
|
||||||
print("#prescript...")
|
|
||||||
prebuild(env)
|
|
||||||
board="PLATFORM_BOARD_%s"%env["BOARD"].replace("-","_").upper()
|
|
||||||
print("Board=#%s#"%board)
|
|
||||||
print("BuildFlags=%s"%(" ".join(env["BUILD_FLAGS"])))
|
|
||||||
env.Append(
|
|
||||||
LINKFLAGS=[ "-u", "custom_app_desc" ],
|
|
||||||
CPPDEFINES=[(board,"1")]
|
|
||||||
)
|
|
||||||
#script does not run on clean yet - maybe in the future
|
|
||||||
env.AddPostAction("clean",cleangenerated)
|
|
Loading…
Reference in New Issue