bzr revid: olt@tinyerp.com-20090115101542-zlgr0hb6xstgatfb
This commit is contained in:
Olivier Laurent 2009-01-15 11:15:42 +01:00
commit 61824218f8
12 changed files with 174 additions and 130 deletions

View File

@ -41,7 +41,8 @@ import release
import re
import base64
from zipfile import PyZipFile, ZIP_DEFLATED
import cStringIO
from cStringIO import StringIO
logger = netsvc.Logger()
@ -55,6 +56,7 @@ if ad != _ad:
# Modules already loaded
loaded = []
class Graph(dict):
def addNode(self, name, deps):
@ -78,6 +80,7 @@ class Graph(dict):
yield module
level += 1
class Singleton(object):
def __new__(cls, name, graph):
@ -89,6 +92,7 @@ class Singleton(object):
graph[name] = inst
return inst
class Node(Singleton):
def __init__(self, name, graph):
@ -106,7 +110,7 @@ class Node(Singleton):
for attr in ('init', 'update', 'demo'):
if hasattr(self, attr):
setattr(node, attr, True)
self.childs.sort(lambda x,y: cmp(x.name, y.name))
self.childs.sort(lambda x, y: cmp(x.name, y.name))
def hasChild(self, name):
return Node(name, self.graph) in self.childs or \
@ -134,9 +138,9 @@ class Node(Singleton):
s += '%s`-> %s' % (' ' * depth, c._pprint(depth+1))
return s
def get_module_path(module):
"""Return the path of the given module.
"""
"""Return the path of the given module."""
if os.path.exists(opj(ad, module)) or os.path.exists(opj(ad, '%s.zip' % module)):
return opj(ad, module)
@ -147,14 +151,16 @@ def get_module_path(module):
logger.notifyChannel('init', netsvc.LOG_WARNING, 'module %s: module not found' % (module,))
return False
def get_module_filetree(module, dir='.'):
path = get_module_path(module)
if not path:
return False
dir = os.path.normpath(dir)
if dir == '.': dir = ''
if dir.startswith('..') or dir[0] == '/':
if dir == '.':
dir = ''
if dir.startswith('..') or (dir and dir[0] == '/'):
raise Exception('Cannot access file outside the module')
if not os.path.isdir(path):
@ -163,25 +169,27 @@ def get_module_filetree(module, dir='.'):
files = ['/'.join(f.split('/')[1:]) for f in zip.namelist()]
else:
files = tools.osutil.listdir(path, True)
tree = {}
for f in files:
if not f.startswith(dir):
continue
f = f[len(dir)+int(not dir.endswith('/')):]
if dir:
f = f[len(dir)+int(not dir.endswith('/')):]
lst = f.split(os.sep)
current = tree
while len(lst) != 1:
current = current.setdefault(lst.pop(0), {})
current[lst.pop(0)] = None
return tree
def get_module_as_zip(modulename, b64enc=True, src=True):
RE_exclude = re.compile('(?:^\..+\.swp$)|(?:\.py[oc]$)|(?:\.bak$)|(?:\.~.~$)', re.I)
def _zippy(archive, path, src=True):
path = os.path.abspath(path)
base = os.path.basename(path)
@ -189,16 +197,16 @@ def get_module_as_zip(modulename, b64enc=True, src=True):
bf = os.path.basename(f)
if not RE_exclude.search(bf) and (src or bf == '__terp__.py' or not path.endswith('.py')):
archive.write(os.path.join(path, f), os.path.join(base, f))
ap = get_module_path(str(modulename))
if not ap:
raise Exception('Unable to find path for module %s' % modulename)
ap = ap.encode('utf8')
ap = ap.encode('utf8')
if os.path.isfile(ap + '.zip'):
val = file(ap + '.zip', 'rb').read()
else:
archname = cStringIO.StringIO('wb')
archname = StringIO()
archive = PyZipFile(archname, "w", ZIP_DEFLATED)
archive.writepy(ap)
_zippy(archive, ap, src=src)
@ -206,11 +214,6 @@ def get_module_as_zip(modulename, b64enc=True, src=True):
val = archname.getvalue()
archname.close()
### debug
f = file('/tmp/mod.zip', 'wb')
f.write(val)
f.close()
if b64enc:
val = base64.encodestring(val)
return val
@ -227,6 +230,7 @@ def get_module_resource(module, *args):
a = get_module_path(module)
return a and opj(a, *args) or False
def get_modules():
"""Returns the list of module names
"""
@ -244,9 +248,10 @@ def get_modules():
return list(set(listdir(ad) + listdir(_ad)))
def create_graph(module_list, force=None):
if not force:
force=[]
force = []
graph = Graph()
packages = []
@ -258,7 +263,8 @@ def create_graph(module_list, force=None):
except IOError:
continue
terp_file = get_module_resource(module, '__terp__.py')
if not terp_file: continue
if not terp_file:
continue
if os.path.isfile(terp_file) or zipfile.is_zipfile(mod_path+'.zip'):
try:
info = eval(tools.file_open(terp_file).read())
@ -267,14 +273,14 @@ def create_graph(module_list, force=None):
raise
if info.get('installable', True):
packages.append((module, info.get('depends', []), info))
dependencies = dict([(p, deps) for p, deps, data in packages])
current, later = Set([p for p, dep, data in packages]), Set()
while packages and current > later:
package, deps, data = packages[0]
# if all dependencies of 'package' are already in the graph, add 'package' in the graph
if reduce(lambda x,y: x and y in graph, deps, True):
if reduce(lambda x, y: x and y in graph, deps, True):
if not package in current:
packages.pop(0)
continue
@ -290,15 +296,15 @@ def create_graph(module_list, force=None):
later.add(package)
packages.append((package, deps, data))
packages.pop(0)
for package in later:
unmet_deps = filter(lambda p: p not in graph, dependencies[package])
logger.notifyChannel('init', netsvc.LOG_ERROR, 'module %s: Unmet dependencies: %s' % (package, ', '.join(unmet_deps)))
return graph
def init_module_objects(cr, module_name, obj_list):
pool = pooler.get_pool(cr.dbname)
logger.notifyChannel('init', netsvc.LOG_INFO, 'module %s: creating or updating database tables' % module_name)
todo = []
for obj in obj_list:
@ -313,6 +319,7 @@ def init_module_objects(cr, module_name, obj_list):
t[1](cr, *t[2])
cr.commit()
def register_class(m):
"""
Register module named m, if not already registered
@ -353,8 +360,8 @@ class MigrationManager(object):
This class manage the migration of modules
Migrations files must be python files containing a "migrate(cr, installed_version)" function.
Theses files must respect a directory tree structure: A 'migrations' folder which containt a
folder by version. Version can be 'module' version or 'server.module' version (in this case,
the files will only be processed by this version of the server). Python file names must start
folder by version. Version can be 'module' version or 'server.module' version (in this case,
the files will only be processed by this version of the server). Python file names must start
by 'pre' or 'post' and will be executed, respectively, before and after the module initialisation
Example:
@ -370,7 +377,7 @@ class MigrationManager(object):
| `-- post-clean-data.py
`-- foo.py # not processed
This similar structure is generated by the maintenance module with the migrations files get by
This similar structure is generated by the maintenance module with the migrations files get by
the maintenance contract
"""
@ -395,7 +402,6 @@ class MigrationManager(object):
self.migrations[pkg.name]['module'] = get_module_filetree(pkg.name, 'migrations') or {}
self.migrations[pkg.name]['maintenance'] = get_module_filetree('base', 'maintenance/migrations/' + pkg.name) or {}
def migrate_module(self, pkg, stage):
assert stage in ('pre', 'post')
stageformat = {'pre': '[>%s]',
@ -404,7 +410,7 @@ class MigrationManager(object):
if not (hasattr(pkg, 'update') or pkg.state == 'to upgrade'):
return
def convert_version(version):
if version.startswith(release.major_version) and version != release.major_version:
return version # the version number already containt the server version
@ -415,7 +421,7 @@ class MigrationManager(object):
return [d for d in tree if tree[d] is not None]
versions = list(set(
__get_dir(self.migrations[pkg.name]['module']) +
__get_dir(self.migrations[pkg.name]['module']) +
__get_dir(self.migrations[pkg.name]['maintenance'])
))
versions.sort(key=lambda k: parse_version(convert_version(k)))
@ -426,7 +432,7 @@ class MigrationManager(object):
"""
m = self.migrations[pkg.name]
lst = []
mapping = {'module': {'module': pkg.name, 'rootdir': opj('migrations')},
'maintenance': {'module': 'base', 'rootdir': opj('maintenance', 'migrations', pkg.name)},
}
@ -441,7 +447,7 @@ class MigrationManager(object):
lst.append((mapping[x]['module'], opj(mapping[x]['rootdir'], version, f)))
return lst
def mergedict(a,b):
def mergedict(a, b):
a = a.copy()
a.update(b)
return a
@ -450,7 +456,7 @@ class MigrationManager(object):
parsed_installed_version = parse_version(pkg.installed_version or '')
current_version = parse_version(convert_version(pkg.data.get('version', '0')))
versions = _get_migration_versions(pkg)
for version in versions:
@ -460,32 +466,44 @@ class MigrationManager(object):
'stage': stage,
'version': stageformat[stage] % version,
}
for modulename, pyfile in _get_migration_files(pkg, version, stage):
name, ext = os.path.splitext(os.path.basename(pyfile))
if ext.lower() != '.py':
continue
fp = tools.file_open(opj(modulename, pyfile))
mod = None
mod = fp = fp2 = None
try:
mod = imp.load_source(name, pyfile, fp)
logger.notifyChannel('migration', netsvc.LOG_INFO, 'module %(addon)s: Running migration %(version)s %(name)s"' % mergedict({'name': mod.__name__},strfmt))
mod.migrate(self.cr, pkg.installed_version)
except ImportError:
logger.notifyChannel('migration', netsvc.LOG_ERROR, 'module %(addon)s: Unable to load %(stage)-migration file %(file)s' % mergedict({'file': opj(modulename,pyfile)}, strfmt))
raise
except AttributeError:
logger.notifyChannel('migration', netsvc.LOG_ERROR, 'module %(addon)s: Each %(stage)-migration file must have a "migrate(cr, installed_version)" function' % strfmt)
except:
raise
fp.close()
del mod
fp = tools.file_open(opj(modulename, pyfile))
# imp.load_source need a real file object, so we create
# on from the file-like object we get from file_open
fp2 = os.tmpfile()
fp2.write(fp.read())
fp2.seek(0)
try:
mod = imp.load_source(name, pyfile, fp2)
logger.notifyChannel('migration', netsvc.LOG_INFO, 'module %(addon)s: Running migration %(version)s %(name)s"' % mergedict({'name': mod.__name__}, strfmt))
mod.migrate(self.cr, pkg.installed_version)
except ImportError:
logger.notifyChannel('migration', netsvc.LOG_ERROR, 'module %(addon)s: Unable to load %(stage)s-migration file %(file)s' % mergedict({'file': opj(modulename, pyfile)}, strfmt))
raise
except AttributeError:
logger.notifyChannel('migration', netsvc.LOG_ERROR, 'module %(addon)s: Each %(stage)s-migration file must have a "migrate(cr, installed_version)" function' % strfmt)
except:
raise
finally:
if fp:
fp.close()
if fp2:
fp2.close()
if mod:
del mod
def load_module_graph(cr, graph, status=None, perform_checks=True, **kwargs):
# **kwargs is passed directly to convert_xml_import
if not status:
status={}
status = {}
status = status.copy()
package_todo = []
@ -509,14 +527,14 @@ def load_module_graph(cr, graph, status=None, perform_checks=True, **kwargs):
for package in graph:
for k, v in additional_data[package.name].items():
setattr(package, k, v)
migrations = MigrationManager(cr, graph)
check_rules = False
modobj = None
for package in graph:
status['progress'] = (float(statusi)+0.1)/len(graph)
status['progress'] = (float(statusi)+0.1) / len(graph)
m = package.name
mid = package.id
@ -533,14 +551,14 @@ def load_module_graph(cr, graph, status=None, perform_checks=True, **kwargs):
modobj.check(cr, 1, [mid])
idref = {}
status['progress'] = (float(statusi)+0.4)/len(graph)
status['progress'] = (float(statusi)+0.4) / len(graph)
if hasattr(package, 'init') or hasattr(package, 'update') or package.state in ('to install', 'to upgrade'):
check_rules = True
init_module_objects(cr, m, modules)
for kind in ('init', 'update'):
for filename in package.data.get('%s_xml' % kind, []):
mode = 'update'
if hasattr(package, 'init') or package.state=='to install':
if hasattr(package, 'init') or package.state == 'to install':
mode = 'init'
logger.notifyChannel('init', netsvc.LOG_INFO, 'module %s: loading %s' % (m, filename))
name, ext = os.path.splitext(filename)
@ -557,7 +575,7 @@ def load_module_graph(cr, graph, status=None, perform_checks=True, **kwargs):
tools.convert_xml_import(cr, m, fp, idref, mode=mode, **kwargs)
fp.close()
if hasattr(package, 'demo') or (package.dbdemo and package.state != 'installed'):
status['progress'] = (float(statusi)+0.75)/len(graph)
status['progress'] = (float(statusi)+0.75) / len(graph)
for xml in package.data.get('demo_xml', []):
name, ext = os.path.splitext(xml)
logger.notifyChannel('init', netsvc.LOG_INFO, 'module %s: loading %s' % (m, xml))
@ -574,7 +592,7 @@ def load_module_graph(cr, graph, status=None, perform_checks=True, **kwargs):
#cr.execute("update ir_module_module set state='installed', latest_version=%s where id=%s", (ver, mid,))
# Set new modules and dependencies
modobj.write(cr, 1, [mid], {'state':'installed', 'latest_version':ver})
modobj.write(cr, 1, [mid], {'state': 'installed', 'latest_version': ver})
cr.commit()
# Update translations for all installed languages
@ -583,12 +601,12 @@ def load_module_graph(cr, graph, status=None, perform_checks=True, **kwargs):
cr.commit()
migrations.migrate_module(package, 'post')
statusi+=1
statusi += 1
if perform_checks and check_rules:
cr.execute("""select model,name from ir_model where id not in (select model_id from ir_model_access)""")
for (model,name) in cr.fetchall():
logger.notifyChannel('init', netsvc.LOG_WARNING, 'object %s (%s) has no access rules!' % (model,name))
for (model, name) in cr.fetchall():
logger.notifyChannel('init', netsvc.LOG_WARNING, 'object %s (%s) has no access rules!' % (model, name))
cr.execute('select model from ir_model where state=%s', ('manual',))
@ -598,9 +616,10 @@ def load_module_graph(cr, graph, status=None, perform_checks=True, **kwargs):
pool.get('ir.model.data')._process_end(cr, 1, package_todo)
cr.commit()
def load_modules(db, force_demo=False, status=None, update_module=False):
if not status:
status={}
status = {}
cr = db.cursor()
force = []
@ -622,20 +641,20 @@ def load_modules(db, force_demo=False, status=None, update_module=False):
ids = modobj.search(cr, 1, ['&', ('state', '=', 'uninstalled'), ('name', 'in', mods)])
if ids:
modobj.button_install(cr, 1, ids)
mods = [k for k in tools.config['update'] if tools.config['update'][k]]
if mods:
ids = modobj.search(cr, 1, ['&',('state', '=', 'installed'), ('name', 'in', mods)])
ids = modobj.search(cr, 1, ['&', ('state', '=', 'installed'), ('name', 'in', mods)])
if ids:
modobj.button_upgrade(cr, 1, ids)
cr.execute("update ir_module_module set state=%s where name=%s", ('installed', 'base'))
cr.execute("select name from ir_module_module where state in ('installed', 'to install', 'to upgrade')")
else:
cr.execute("select name from ir_module_module where state in ('installed', 'to upgrade')")
module_list = [name for (name,) in cr.fetchall()]
graph = create_graph(module_list, force)
# the 'base' module has already been updated
base = graph['base']
base.state = 'installed'
@ -648,7 +667,7 @@ def load_modules(db, force_demo=False, status=None, update_module=False):
logger.notifyChannel('init', netsvc.LOG_INFO, report)
for kind in ('init', 'demo', 'update'):
tools.config[kind]={}
tools.config[kind] = {}
cr.commit()
if update_module:
@ -656,7 +675,7 @@ def load_modules(db, force_demo=False, status=None, update_module=False):
for mod_id, mod_name in cr.fetchall():
pool = pooler.get_pool(cr.dbname)
cr.execute('select model,res_id from ir_model_data where noupdate=%s and module=%s order by id desc', (False, mod_name,))
for rmod,rid in cr.fetchall():
for rmod, rid in cr.fetchall():
uid = 1
pool.get(rmod).unlink(cr, uid, [rid])
cr.execute('delete from ir_model_data where noupdate=%s and module=%s', (False, mod_name,))

View File

@ -30,7 +30,8 @@ def _check_xml(self, cr, uid, ids, context={}):
for view in self.browse(cr, uid, ids, context):
eview = etree.fromstring(view.arch.encode('utf8'))
frng = tools.file_open(os.path.join('base','rng','view.rng'))
relaxng = etree.RelaxNG(file=frng)
relaxng_doc = etree.parse(frng)
relaxng = etree.RelaxNG(relaxng_doc)
if not relaxng.validate(eview):
logger = netsvc.Logger()
logger.notifyChannel('init', netsvc.LOG_ERROR, 'The view does not fit the required schema !')
@ -64,7 +65,7 @@ class view(osv.osv):
'field_parent': fields.char('Childs Field',size=64),
}
_defaults = {
'arch': lambda *a: '<?xml version="1.0"?>\n<tree title="Unknwown">\n\t<field name="name"/>\n</tree>',
'arch': lambda *a: '<?xml version="1.0"?>\n<tree string="Unknwown">\n\t<field name="name"/>\n</tree>',
'priority': lambda *a: 16
}
_order = "priority"
@ -73,7 +74,7 @@ class view(osv.osv):
]
def create(self, cr, uid, vals, context={}):
if 'inherit_id' in vals:
if 'inherit_id' in vals and vals['inherit_id']:
obj=self.browse(cr,uid,vals['inherit_id'])
child=self.pool.get(vals['model'])
error="Inherited view model [%s] and \

View File

@ -28,7 +28,7 @@ import os
import tools
import zipfile
import cStringIO
from StringIO import StringIO
import base64
finish_form ='''<?xml version="1.0"?>
@ -55,7 +55,8 @@ class move_module_wizard(wizard.interface):
module_data = data['form']['module_file']
val =base64.decodestring(module_data)
fp = cStringIO.StringIO(val)
fp = StringIO()
fp.write(val)
fdata = zipfile.ZipFile(fp, 'r')
fname = fdata.namelist()[0]
module_name = os.path.split(fname)[0]

View File

@ -158,8 +158,13 @@ class report_custom(report_int):
row = []
cond = []
for i in range(4):
if f['field_child'+str(i)]:
row.append(f['field_child'+str(i)][1])
field_child = f['field_child'+str(i)]
if field_child:
row.append(
service.execute(cr.dbname, uid,
'ir.model.fields', 'read', [field_child[0]],
['name'], context=context)[0]['name']
)
if f['fc'+str(i)+'_operande']:
fct_name = 'id'
cond_op = f['fc'+str(i)+'_op']

View File

@ -163,9 +163,10 @@ class document(object):
ext = fname.split('.')[-1].lower()
if ext in ('jpg','jpeg', 'png'):
import base64
import cStringIO
from StringIO import StringIO
dt = base64.decodestring(datas['datas'])
fp = cStringIO.StringIO(dt)
fp = StringIO()
fp.write(dt)
i = str(len(self.bin_datas))
self.bin_datas[i] = fp

View File

@ -40,7 +40,7 @@
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
import sys
import cStringIO
from StringIO import StringIO
import xml.dom.minidom
import copy
@ -351,26 +351,26 @@ class _rml_canvas(object):
import urllib
from reportlab.lib.utils import ImageReader
s = StringIO()
if not node.hasAttribute('file'):
if node.hasAttribute('name'):
image_data = self.images[node.getAttribute('name')]
s = cStringIO.StringIO(image_data)
s.write(image_data)
else:
import base64
image_data = base64.decodestring(node.firstChild.nodeValue)
if not image_data: return False
s = cStringIO.StringIO(image_data)
s.write(image_data)
else:
if node.getAttribute('file') in self.images:
s = cStringIO.StringIO(self.images[node.getAttribute('file')])
s.write(self.images[node.getAttribute('file')])
else:
try:
u = urllib.urlopen(str(node.getAttribute('file')))
s = cStringIO.StringIO(u.read())
except:
u = file(os.path.join(self.path,str(node.getAttribute('file'))), 'rb')
s = cStringIO.StringIO(u.read())
s.write(u.read())
img = ImageReader(s)
(sx,sy) = img.getSize()
@ -639,7 +639,8 @@ class _rml_flowable(object):
else:
import base64
image_data = base64.decodestring(node.firstChild.nodeValue)
image = cStringIO.StringIO(image_data)
image = StringIO()
image.write(image_data)
return platypus.Image(image, mask=(250,255,250,255,250,255), **(utils.attr_get(node, ['width','height'])))
else:
return platypus.Image(node.getAttribute('file'), mask=(250,255,250,255,250,255), **(utils.attr_get(node, ['width','height'])))
@ -794,7 +795,7 @@ def parseString(data, fout=None, images={}, path='.',title=None):
fp.close()
return fout
else:
fp = cStringIO.StringIO()
fp = StringIO()
r.render(fp)
return fp.getvalue()

View File

@ -598,6 +598,7 @@ class report_sxw(report_rml):
report_xml = None
title=''
attach = False
want_header = self.header
if report_xml_ids:
report_xml = ir_actions_report_xml_obj.browse(cr, uid, report_xml_ids[0],
context=context)
@ -605,6 +606,7 @@ class report_sxw(report_rml):
attach = report_xml.attachment
rml = report_xml.report_rml_content
report_type = report_xml.report_type
want_header = report_xml.header
else:
ir_menu_report_obj = pool.get('ir.ui.menu')
report_menu_ids = ir_menu_report_obj.search(cr, uid,
@ -657,14 +659,14 @@ class report_sxw(report_rml):
pe.childNodes[0].data=data['model']
meta = rml_dom_meta.documentElement.toxml('utf-8')
rml2 = rml_parser._parse(rml_dom, objs, data, header=self.header)
rml2 = rml_parser._parse(rml_dom, objs, data, header=want_header)
sxw_z = zipfile.ZipFile(sxw_io, mode='a')
sxw_z.writestr('content.xml', "<?xml version='1.0' encoding='UTF-8'?>" + \
rml2)
sxw_z.writestr('meta.xml', "<?xml version='1.0' encoding='UTF-8'?>" + \
meta)
if self.header:
if want_header:
#Add corporate header/footer
if report_type=='odt':
rml = tools.file_open('custom/corporate_odt_header.xml').read()
@ -676,7 +678,7 @@ class report_sxw(report_rml):
objs = self.getObjects(cr, uid, ids, context)
rml_parser.preprocess(objs, data, ids)
rml_dom = xml.dom.minidom.parseString(rml)
rml2 = rml_parser._parse(rml_dom, objs, data, header=self.header)
rml2 = rml_parser._parse(rml_dom, objs, data, header=want_header)
sxw_z.writestr('styles.xml',"<?xml version='1.0' encoding='UTF-8'?>" + \
rml2)
sxw_z.close()
@ -690,7 +692,7 @@ class report_sxw(report_rml):
objs = self.getObjects(cr, uid, ids, context)
rml_parser.preprocess(objs, data, ids)
rml_dom = xml.dom.minidom.parseString(rml)
rml2 = rml_parser._parse(rml_dom, objs, data, header=self.header)
rml2 = rml_parser._parse(rml_dom, objs, data, header=want_header)
if rml_parser.logo:
logo = base64.decodestring(rml_parser.logo)

View File

@ -68,12 +68,14 @@ class db(netsvc.Service):
self.actions[id] = {'clean': False}
db = sql_db.db_connect('template1', serialize=1)
db = sql_db.db_connect('template1')
cr = db.cursor()
cr.autocommit(True)
time.sleep(0.2)
cr.execute('CREATE DATABASE ' + db_name + ' ENCODING \'unicode\'')
cr.close()
try:
cr.autocommit(True)
cr.execute('CREATE DATABASE ' + db_name + ' ENCODING \'unicode\'')
finally:
cr.close()
sql_db.close_db('template1')
class DBInitialize(object):
def __call__(self, serv, id, db_name, demo, lang, user_password='admin'):
try:
@ -146,7 +148,7 @@ class db(netsvc.Service):
sql_db.close_db(db_name)
logger = netsvc.Logger()
db = sql_db.db_connect('template1', serialize=1)
db = sql_db.db_connect('template1')
cr = db.cursor()
cr.autocommit(True)
try:
@ -161,6 +163,7 @@ class db(netsvc.Service):
'DROP DB: %s' % (db_name))
finally:
cr.close()
sql_db.close_db('template1')
return True
def dump(self, password, db_name):
@ -197,11 +200,14 @@ class db(netsvc.Service):
'RESTORE DB: %s already exists' % (db_name,))
raise Exception, "Database already exists"
db = sql_db.db_connect('template1', serialize=1)
db = sql_db.db_connect('template1')
cr = db.cursor()
cr.autocommit(True)
cr.execute('CREATE DATABASE ' + db_name + ' ENCODING \'unicode\'')
cr.close()
try:
cr.execute('CREATE DATABASE ' + db_name + ' ENCODING \'unicode\'')
finally:
cr.close()
sql_db.close_db('template1')
cmd = ['pg_restore']
if tools.config['db_user']:
@ -240,24 +246,28 @@ class db(netsvc.Service):
def list(self):
db = sql_db.db_connect('template1')
cr = db.cursor()
try:
cr = db.cursor()
db_user = tools.config["db_user"]
if not db_user and os.name == 'posix':
import pwd
db_user = pwd.getpwuid(os.getuid())[0]
if not db_user:
cr.execute("select decode(usename, 'escape') from pg_user where usesysid=(select datdba from pg_database where datname=%s)", (tools.config["db_name"],))
res = cr.fetchone()
db_user = res and str(res[0])
if db_user:
cr.execute("select decode(datname, 'escape') from pg_database where datdba=(select usesysid from pg_user where usename=%s) and datname not in ('template0', 'template1', 'postgres')", (db_user,))
else:
cr.execute("select decode(datname, 'escape') from pg_database where datname not in('template0', 'template1','postgres')")
res = [str(name) for (name,) in cr.fetchall()]
try:
cr = db.cursor()
db_user = tools.config["db_user"]
if not db_user and os.name == 'posix':
import pwd
db_user = pwd.getpwuid(os.getuid())[0]
if not db_user:
cr.execute("select decode(usename, 'escape') from pg_user where usesysid=(select datdba from pg_database where datname=%s)", (tools.config["db_name"],))
res = cr.fetchone()
db_user = res and str(res[0])
if db_user:
cr.execute("select decode(datname, 'escape') from pg_database where datdba=(select usesysid from pg_user where usename=%s) and datname not in ('template0', 'template1', 'postgres')", (db_user,))
else:
cr.execute("select decode(datname, 'escape') from pg_database where datname not in('template0', 'template1','postgres')")
res = [str(name) for (name,) in cr.fetchall()]
except:
res = []
finally:
cr.close()
except:
res = []
sql_db.close_db('template1')
res.sort()
return res
@ -398,7 +408,10 @@ GNU Public Licence.
mp = addons.get_module_path(module)
if mp:
if os.path.isdir(mp):
rmtree(os.path.realpath(mp))
if os.path.islink(mp):
os.unlink(mp)
else:
rmtree(mp)
else:
os.unlink(mp + '.zip')

View File

@ -223,7 +223,7 @@ class PoolManager(object):
cls._pools[db_name].closeall()
del cls._pools[db_name]
def db_connect(db_name, serialize=0):
def db_connect(db_name):
return PoolManager.get(db_name)
def close_db(db_name):

View File

@ -44,7 +44,7 @@ from itertools import izip
def init_db(cr):
import addons
f = addons.get_module_resource('base', 'base.sql')
for line in file(f).read().split(';'):
for line in file_open(f).read().split(';'):
if (len(line)>0) and (not line.isspace()):
cr.execute(line)
cr.commit()
@ -55,12 +55,8 @@ def init_db(cr):
if not mod_path:
continue
info = False
if os.path.isfile(terp_file) and not os.path.isfile(mod_path+'.zip'):
info = eval(file(terp_file).read())
elif zipfile.is_zipfile(mod_path+'.zip'):
zfile = zipfile.ZipFile(mod_path+'.zip')
i = os.path.splitext(i)[0]
info = eval(zfile.read(os.path.join(i, '__terp__.py')))
if os.path.isfile(terp_file) or os.path.isfile(mod_path+'.zip'):
info = eval(file_open(terp_file).read())
if info:
categs = info.get('category', 'Uncategorized').split('/')
p_id = None
@ -224,13 +220,14 @@ def file_open(name, mode="r", subdir='addons', pathinfo=False):
else:
zipname = tail
if zipfile.is_zipfile(head+'.zip'):
import cStringIO
from cStringIO import StringIO
zfile = zipfile.ZipFile(head+'.zip')
try:
fo = cStringIO.StringIO(zfile.read(os.path.join(
fo = StringIO()
fo.write(zfile.read(os.path.join(
os.path.basename(head), zipname).replace(
os.sep, '/')))
fo.seek(0)
if pathinfo:
return fo, name
return fo
@ -721,19 +718,22 @@ def get_languages():
'bs_BS': u'Bosnian / bosanski jezik',
'ca_ES': u'Catalan / Català',
'cs_CZ': u'Czech / Čeština',
'da_DK': u'Danish / Dansk',
'de_DE': u'German / Deutsch',
'el_EL': u'Greek / Ελληνικά',
'en_CA': u'English (CA)',
'en_EN': u'English (default)',
'en_GB': u'English (UK)',
'en_US': u'English (US)',
'es_AR': u'Spanish (AR) / Español (AR)',
'es_ES': u'Spanish / Español',
'et_ET': u'Estonian / Eesti keel',
'et_EE': u'Estonian / Eesti keel',
'fr_BE': u'French (BE) / Français (BE)',
'fr_CH': u'French (CH) / Français (CH)',
'fr_FR': u'French / Français',
'hr_HR': u'Croatian / hrvatski jezik',
'hu_HU': u'Hungarian / Magyar',
'id_ID': u'Indonesian / Bahasa Indonesia',
'it_IT': u'Italian / Italiano',
'lt_LT': u'Lithuanian / Lietuvių kalba',
'nl_NL': u'Dutch / Nederlands',

View File

@ -335,7 +335,7 @@ def trans_generate(lang, modules, dbname=None):
_to_translate = []
def push_translation(module, type, name, id, source):
tuple = (module, type, name, id, source)
tuple = (module, source, name, id, type)
if source and tuple not in _to_translate:
_to_translate.append(tuple)
@ -486,8 +486,9 @@ def trans_generate(lang, modules, dbname=None):
out = [["module","type","name","res_id","src","value"]] # header
_to_translate.sort()
# translate strings marked as to be translated
for module, type, name, id, source in _to_translate:
for module, source, name, id, type in _to_translate:
trans = trans_obj._get_source(cr, uid, name, type, lang, source)
out.append([module, type, name, id, source, encode(trans) or ''])