718 lines
28 KiB
Python
718 lines
28 KiB
Python
# -*- coding: utf-8 -*-
|
|
import hashlib
|
|
import inspect
|
|
import itertools
|
|
import logging
|
|
import math
|
|
import re
|
|
import urlparse
|
|
|
|
import werkzeug
|
|
import werkzeug.exceptions
|
|
import werkzeug.utils
|
|
import werkzeug.wrappers
|
|
# optional python-slugify import (https://github.com/un33k/python-slugify)
|
|
try:
|
|
import slugify as slugify_lib
|
|
except ImportError:
|
|
slugify_lib = None
|
|
|
|
import openerp
|
|
from openerp.osv import orm, osv, fields
|
|
from openerp.tools.safe_eval import safe_eval
|
|
from openerp.addons.web.http import request
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
def url_for(path_or_uri, lang=None):
|
|
if isinstance(path_or_uri, unicode):
|
|
path_or_uri = path_or_uri.encode('utf-8')
|
|
current_path = request.httprequest.path
|
|
if isinstance(current_path, unicode):
|
|
current_path = current_path.encode('utf-8')
|
|
location = path_or_uri.strip()
|
|
force_lang = lang is not None
|
|
url = urlparse.urlparse(location)
|
|
|
|
if request and not url.netloc and not url.scheme and (url.path or force_lang):
|
|
location = urlparse.urljoin(current_path, location)
|
|
|
|
lang = lang or request.context.get('lang')
|
|
langs = [lg[0] for lg in request.website.get_languages()]
|
|
|
|
if (len(langs) > 1 or force_lang) and is_multilang_url(location, langs):
|
|
ps = location.split('/')
|
|
if ps[1] in langs:
|
|
# Replace the language only if we explicitly provide a language to url_for
|
|
if force_lang:
|
|
ps[1] = lang
|
|
# Remove the default language unless it's explicitly provided
|
|
elif ps[1] == request.website.default_lang_code:
|
|
ps.pop(1)
|
|
# Insert the context language or the provided language
|
|
elif lang != request.website.default_lang_code or force_lang:
|
|
ps.insert(1, lang)
|
|
location = '/'.join(ps)
|
|
|
|
return location.decode('utf-8')
|
|
|
|
def is_multilang_url(path, langs=None):
|
|
if not langs:
|
|
langs = [lg[0] for lg in request.website.get_languages()]
|
|
spath = path.split('/')
|
|
# if a language is already in the path, remove it
|
|
if spath[1] in langs:
|
|
spath.pop(1)
|
|
path = '/'.join(spath)
|
|
try:
|
|
router = request.httprequest.app.get_db_router(request.db).bind('')
|
|
func = router.match(path)[0]
|
|
return func.routing.get('multilang', False)
|
|
except Exception:
|
|
return False
|
|
|
|
def slugify(s, max_length=None):
|
|
if slugify_lib:
|
|
# There are 2 different libraries only python-slugify is supported
|
|
try:
|
|
return slugify_lib.slugify(s, max_length=max_length)
|
|
except TypeError:
|
|
pass
|
|
spaceless = re.sub(r'\s+', '-', s)
|
|
specialless = re.sub(r'[^-_A-Za-z0-9]', '', spaceless)
|
|
return specialless[:max_length]
|
|
|
|
def slug(value):
|
|
if isinstance(value, orm.browse_record):
|
|
# [(id, name)] = value.name_get()
|
|
id, name = value.id, value[value._rec_name]
|
|
else:
|
|
# assume name_search result tuple
|
|
id, name = value
|
|
slugname = slugify(name or '')
|
|
if not slugname:
|
|
return str(id)
|
|
return "%s-%d" % (slugname, id)
|
|
|
|
def urlplus(url, params):
|
|
return werkzeug.Href(url)(params or None)
|
|
|
|
class website(osv.osv):
|
|
def _get_menu_website(self, cr, uid, ids, context=None):
|
|
# IF a menu is changed, update all websites
|
|
return self.search(cr, uid, [], context=context)
|
|
|
|
def _get_menu(self, cr, uid, ids, name, arg, context=None):
|
|
root_domain = [('parent_id', '=', False)]
|
|
menus = self.pool.get('website.menu').search(cr, uid, root_domain, order='id', context=context)
|
|
menu = menus and menus[0] or False
|
|
return dict( map(lambda x: (x, menu), ids) )
|
|
|
|
def _get_public_user(self, cr, uid, ids, name='public_user', arg=(), context=None):
|
|
ref = self.get_public_user(cr, uid, context=context)
|
|
return dict( map(lambda x: (x, ref), ids) )
|
|
|
|
_name = "website" # Avoid website.website convention for conciseness (for new api). Got a special authorization from xmo and rco
|
|
_description = "Website"
|
|
_columns = {
|
|
'name': fields.char('Domain'),
|
|
'company_id': fields.many2one('res.company', string="Company"),
|
|
'language_ids': fields.many2many('res.lang', 'website_lang_rel', 'website_id', 'lang_id', 'Languages'),
|
|
'default_lang_id': fields.many2one('res.lang', string="Default language"),
|
|
'default_lang_code': fields.related('default_lang_id', 'code', type="char", string="Default language code", store=True),
|
|
'social_twitter': fields.char('Twitter Account'),
|
|
'social_facebook': fields.char('Facebook Account'),
|
|
'social_github': fields.char('GitHub Account'),
|
|
'social_linkedin': fields.char('LinkedIn Account'),
|
|
'social_youtube': fields.char('Youtube Account'),
|
|
'social_googleplus': fields.char('Google+ Account'),
|
|
'google_analytics_key': fields.char('Google Analytics Key'),
|
|
'user_id': fields.many2one('res.users', string='Public User'),
|
|
'public_user': fields.function(_get_public_user, relation='res.users', type='many2one', string='Public User'),
|
|
'menu_id': fields.function(_get_menu, relation='website.menu', type='many2one', string='Main Menu',
|
|
store= {
|
|
'website.menu': (_get_menu_website, ['sequence','parent_id','website_id'], 10)
|
|
})
|
|
}
|
|
|
|
# cf. Wizard hack in website_views.xml
|
|
def noop(self, *args, **kwargs):
|
|
pass
|
|
|
|
def write(self, cr, uid, ids, vals, context=None):
|
|
self._get_languages.clear_cache(self)
|
|
return super(website, self).write(cr, uid, ids, vals, context)
|
|
|
|
def new_page(self, cr, uid, name, template='website.default_page', ispage=True, context=None):
|
|
context = context or {}
|
|
imd = self.pool.get('ir.model.data')
|
|
view = self.pool.get('ir.ui.view')
|
|
template_module, template_name = template.split('.')
|
|
|
|
# completely arbitrary max_length
|
|
page_name = slugify(name, max_length=50)
|
|
page_xmlid = "%s.%s" % (template_module, page_name)
|
|
|
|
try:
|
|
# existing page
|
|
imd.get_object_reference(cr, uid, template_module, page_name)
|
|
except ValueError:
|
|
# new page
|
|
_, template_id = imd.get_object_reference(cr, uid, template_module, template_name)
|
|
page_id = view.copy(cr, uid, template_id, context=context)
|
|
page = view.browse(cr, uid, page_id, context=context)
|
|
page.write({
|
|
'arch': page.arch.replace(template, page_xmlid),
|
|
'name': page_name,
|
|
'page': ispage,
|
|
})
|
|
imd.create(cr, uid, {
|
|
'name': page_name,
|
|
'module': template_module,
|
|
'model': 'ir.ui.view',
|
|
'res_id': page_id,
|
|
'noupdate': True
|
|
}, context=context)
|
|
return page_xmlid
|
|
|
|
def page_for_name(self, cr, uid, ids, name, module='website', context=None):
|
|
# whatever
|
|
return '%s.%s' % (module, slugify(name, max_length=50))
|
|
|
|
def page_exists(self, cr, uid, ids, name, module='website', context=None):
|
|
try:
|
|
return self.pool["ir.model.data"].get_object_reference(cr, uid, module, name)
|
|
except:
|
|
return False
|
|
|
|
def get_public_user(self, cr, uid, context=None):
|
|
uid = openerp.SUPERUSER_ID
|
|
res = self.pool['ir.model.data'].get_object_reference(cr, uid, 'base', 'public_user')
|
|
return res and res[1] or False
|
|
|
|
@openerp.tools.ormcache(skiparg=3)
|
|
def _get_languages(self, cr, uid, id, context=None):
|
|
website = self.browse(cr, uid, id)
|
|
return [(lg.code, lg.name) for lg in website.language_ids]
|
|
|
|
def get_languages(self, cr, uid, ids, context=None):
|
|
return self._get_languages(cr, uid, ids[0])
|
|
|
|
def get_current_website(self, cr, uid, context=None):
|
|
# TODO: Select website, currently hard coded
|
|
return self.pool['website'].browse(cr, uid, 1, context=context)
|
|
|
|
def preprocess_request(self, cr, uid, ids, request, context=None):
|
|
# TODO FP: is_website_publisher and editable in context should be removed
|
|
# for performance reasons (1 query per image to load) but also to be cleaner
|
|
# I propose to replace this by a group 'base.group_website_publisher' on the
|
|
# view that requires it.
|
|
Access = request.registry['ir.model.access']
|
|
is_website_publisher = Access.check(cr, uid, 'ir.ui.view', 'write', False, context)
|
|
|
|
lang = request.context['lang']
|
|
is_master_lang = lang == request.website.default_lang_code
|
|
|
|
request.redirect = lambda url: werkzeug.utils.redirect(url_for(url))
|
|
request.context.update(
|
|
editable=is_website_publisher,
|
|
translatable=not is_master_lang,
|
|
)
|
|
|
|
def get_template(self, cr, uid, ids, template, context=None):
|
|
if '.' not in template:
|
|
template = 'website.%s' % template
|
|
module, xmlid = template.split('.', 1)
|
|
model, view_id = request.registry["ir.model.data"].get_object_reference(cr, uid, module, xmlid)
|
|
return self.pool["ir.ui.view"].browse(cr, uid, view_id, context=context)
|
|
|
|
def _render(self, cr, uid, ids, template, values=None, context=None):
|
|
# TODO: remove this. (just kept for backward api compatibility for saas-3)
|
|
return self.pool['ir.ui.view'].render(cr, uid, template, values=values, context=context)
|
|
|
|
def render(self, cr, uid, ids, template, values=None, status_code=None, context=None):
|
|
# TODO: remove this. (just kept for backward api compatibility for saas-3)
|
|
return request.render(template, values, uid=uid)
|
|
|
|
def pager(self, cr, uid, ids, url, total, page=1, step=30, scope=5, url_args=None, context=None):
|
|
# Compute Pager
|
|
page_count = int(math.ceil(float(total) / step))
|
|
|
|
page = max(1, min(int(page), page_count))
|
|
scope -= 1
|
|
|
|
pmin = max(page - int(math.floor(scope/2)), 1)
|
|
pmax = min(pmin + scope, page_count)
|
|
|
|
if pmax - pmin < scope:
|
|
pmin = pmax - scope if pmax - scope > 0 else 1
|
|
|
|
def get_url(page):
|
|
_url = "%s/page/%s" % (url, page) if page > 1 else url
|
|
if url_args:
|
|
_url = "%s?%s" % (_url, werkzeug.url_encode(url_args))
|
|
return _url
|
|
|
|
return {
|
|
"page_count": page_count,
|
|
"offset": (page - 1) * step,
|
|
"page": {
|
|
'url': get_url(page),
|
|
'num': page
|
|
},
|
|
"page_start": {
|
|
'url': get_url(pmin),
|
|
'num': pmin
|
|
},
|
|
"page_previous": {
|
|
'url': get_url(max(pmin, page - 1)),
|
|
'num': max(pmin, page - 1)
|
|
},
|
|
"page_next": {
|
|
'url': get_url(min(pmax, page + 1)),
|
|
'num': min(pmax, page + 1)
|
|
},
|
|
"page_end": {
|
|
'url': get_url(pmax),
|
|
'num': pmax
|
|
},
|
|
"pages": [
|
|
{'url': get_url(page), 'num': page}
|
|
for page in xrange(pmin, pmax+1)
|
|
]
|
|
}
|
|
|
|
def rule_is_enumerable(self, rule):
|
|
""" Checks that it is possible to generate sensible GET queries for
|
|
a given rule (if the endpoint matches its own requirements)
|
|
|
|
:type rule: werkzeug.routing.Rule
|
|
:rtype: bool
|
|
"""
|
|
endpoint = rule.endpoint
|
|
methods = rule.methods or ['GET']
|
|
converters = rule._converters.values()
|
|
|
|
return (
|
|
'GET' in methods
|
|
and endpoint.routing['type'] == 'http'
|
|
and endpoint.routing['auth'] in ('none', 'public')
|
|
and endpoint.routing.get('website', False)
|
|
# preclude combinatorial explosion by only allowing a single converter
|
|
and len(converters) <= 1
|
|
# ensure all converters on the rule are able to generate values for
|
|
# themselves
|
|
and all(hasattr(converter, 'generate') for converter in converters)
|
|
) and self.endpoint_is_enumerable(rule)
|
|
|
|
def endpoint_is_enumerable(self, rule):
|
|
""" Verifies that it's possible to generate a valid url for the rule's
|
|
endpoint
|
|
|
|
:type rule: werkzeug.routing.Rule
|
|
:rtype: bool
|
|
"""
|
|
spec = inspect.getargspec(rule.endpoint.method)
|
|
|
|
# if *args bail the fuck out, only dragons can live there
|
|
if spec.varargs:
|
|
return False
|
|
|
|
# remove all arguments with a default value from the list
|
|
defaults_count = len(spec.defaults or []) # spec.defaults can be None
|
|
# a[:-0] ~ a[:0] ~ [] -> replace defaults_count == 0 by None to get
|
|
# a[:None] ~ a
|
|
args = spec.args[:(-defaults_count or None)]
|
|
|
|
# params with defaults were removed, leftover allowed are:
|
|
# * self (technically should be first-parameter-of-instance-method but whatever)
|
|
# * any parameter mapping to a converter
|
|
return all(
|
|
(arg == 'self' or arg in rule._converters)
|
|
for arg in args)
|
|
|
|
def enumerate_pages(self, cr, uid, ids, query_string=None, context=None):
|
|
""" Available pages in the website/CMS. This is mostly used for links
|
|
generation and can be overridden by modules setting up new HTML
|
|
controllers for dynamic pages (e.g. blog).
|
|
|
|
By default, returns template views marked as pages.
|
|
|
|
:param str query_string: a (user-provided) string, fetches pages
|
|
matching the string
|
|
:returns: a list of mappings with two keys: ``name`` is the displayable
|
|
name of the resource (page), ``url`` is the absolute URL
|
|
of the same.
|
|
:rtype: list({name: str, url: str})
|
|
"""
|
|
router = request.httprequest.app.get_db_router(request.db)
|
|
# Force enumeration to be performed as public user
|
|
uid = self.get_public_user(cr, uid, context=context)
|
|
url_list = []
|
|
for rule in router.iter_rules():
|
|
if not self.rule_is_enumerable(rule):
|
|
continue
|
|
|
|
converters = rule._converters
|
|
filtered = bool(converters)
|
|
if converters:
|
|
# allow single converter as decided by fp, checked by
|
|
# rule_is_enumerable
|
|
[(name, converter)] = converters.items()
|
|
converter_values = converter.generate(
|
|
request.cr, uid, query=query_string, context=context)
|
|
generated = ({k: v} for k, v in itertools.izip(
|
|
itertools.repeat(name), converter_values))
|
|
else:
|
|
# force single iteration for literal urls
|
|
generated = [{}]
|
|
|
|
for values in generated:
|
|
domain_part, url = rule.build(values, append_unknown=False)
|
|
page = {'name': url, 'url': url}
|
|
if url in url_list:
|
|
continue
|
|
url_list.append(url)
|
|
if not filtered and query_string and not self.page_matches(cr, uid, page, query_string, context=context):
|
|
continue
|
|
yield page
|
|
|
|
def search_pages(self, cr, uid, ids, needle=None, limit=None, context=None):
|
|
return list(itertools.islice(
|
|
self.enumerate_pages(cr, uid, ids, query_string=needle, context=context),
|
|
limit))
|
|
|
|
def page_matches(self, cr, uid, page, needle, context=None):
|
|
""" Checks that a "page" matches a user-provide search string.
|
|
|
|
The default implementation attempts to perform a non-contiguous
|
|
substring match of the page's name.
|
|
|
|
:param page: {'name': str, 'url': str}
|
|
:param needle: str
|
|
:rtype: bool
|
|
"""
|
|
haystack = page['name'].lower()
|
|
|
|
needle = iter(needle.lower())
|
|
n = next(needle)
|
|
end = object()
|
|
|
|
for char in haystack:
|
|
if char != n: continue
|
|
|
|
n = next(needle, end)
|
|
# found all characters of needle in haystack in order
|
|
if n is end:
|
|
return True
|
|
|
|
return False
|
|
|
|
def kanban(self, cr, uid, ids, model, domain, column, template, step=None, scope=None, orderby=None, context=None):
|
|
step = step and int(step) or 10
|
|
scope = scope and int(scope) or 5
|
|
orderby = orderby or "name"
|
|
|
|
get_args = dict(request.httprequest.args or {})
|
|
model_obj = self.pool[model]
|
|
relation = model_obj._columns.get(column)._obj
|
|
relation_obj = self.pool[relation]
|
|
|
|
get_args.setdefault('kanban', "")
|
|
kanban = get_args.pop('kanban')
|
|
kanban_url = "?%s&kanban=" % werkzeug.url_encode(get_args)
|
|
|
|
pages = {}
|
|
for col in kanban.split(","):
|
|
if col:
|
|
col = col.split("-")
|
|
pages[int(col[0])] = int(col[1])
|
|
|
|
objects = []
|
|
for group in model_obj.read_group(cr, uid, domain, ["id", column], groupby=column):
|
|
obj = {}
|
|
|
|
# browse column
|
|
relation_id = group[column][0]
|
|
obj['column_id'] = relation_obj.browse(cr, uid, relation_id)
|
|
|
|
obj['kanban_url'] = kanban_url
|
|
for k, v in pages.items():
|
|
if k != relation_id:
|
|
obj['kanban_url'] += "%s-%s" % (k, v)
|
|
|
|
# pager
|
|
number = model_obj.search(cr, uid, group['__domain'], count=True)
|
|
obj['page_count'] = int(math.ceil(float(number) / step))
|
|
obj['page'] = pages.get(relation_id) or 1
|
|
if obj['page'] > obj['page_count']:
|
|
obj['page'] = obj['page_count']
|
|
offset = (obj['page']-1) * step
|
|
obj['page_start'] = max(obj['page'] - int(math.floor((scope-1)/2)), 1)
|
|
obj['page_end'] = min(obj['page_start'] + (scope-1), obj['page_count'])
|
|
|
|
# view data
|
|
obj['domain'] = group['__domain']
|
|
obj['model'] = model
|
|
obj['step'] = step
|
|
obj['orderby'] = orderby
|
|
|
|
# browse objects
|
|
object_ids = model_obj.search(cr, uid, group['__domain'], limit=step, offset=offset, order=orderby)
|
|
obj['object_ids'] = model_obj.browse(cr, uid, object_ids)
|
|
|
|
objects.append(obj)
|
|
|
|
values = {
|
|
'objects': objects,
|
|
'range': range,
|
|
'template': template,
|
|
}
|
|
return request.website._render("website.kanban_contain", values)
|
|
|
|
def kanban_col(self, cr, uid, ids, model, domain, page, template, step, orderby, context=None):
|
|
html = ""
|
|
model_obj = self.pool[model]
|
|
domain = safe_eval(domain)
|
|
step = int(step)
|
|
offset = (int(page)-1) * step
|
|
object_ids = model_obj.search(cr, uid, domain, limit=step, offset=offset, order=orderby)
|
|
object_ids = model_obj.browse(cr, uid, object_ids)
|
|
for object_id in object_ids:
|
|
html += request.website._render(template, {'object_id': object_id})
|
|
return html
|
|
|
|
class website_menu(osv.osv):
|
|
_name = "website.menu"
|
|
_description = "Website Menu"
|
|
_columns = {
|
|
'name': fields.char('Menu', size=64, required=True, translate=True),
|
|
'url': fields.char('Url', translate=True),
|
|
'new_window': fields.boolean('New Window'),
|
|
'sequence': fields.integer('Sequence'),
|
|
# TODO: support multiwebsite once done for ir.ui.views
|
|
'website_id': fields.many2one('website', 'Website'),
|
|
'parent_id': fields.many2one('website.menu', 'Parent Menu', select=True, ondelete="cascade"),
|
|
'child_id': fields.one2many('website.menu', 'parent_id', string='Child Menus'),
|
|
'parent_left': fields.integer('Parent Left', select=True),
|
|
'parent_right': fields.integer('Parent Right', select=True),
|
|
}
|
|
|
|
def __defaults_sequence(self, cr, uid, context):
|
|
menu = self.search_read(cr, uid, [(1,"=",1)], ["sequence"], limit=1, order="sequence DESC", context=context)
|
|
return menu and menu[0]["sequence"] or 0
|
|
|
|
_defaults = {
|
|
'url': '',
|
|
'sequence': __defaults_sequence,
|
|
'new_window': False,
|
|
}
|
|
_parent_store = True
|
|
_parent_order = 'sequence'
|
|
_order = "sequence"
|
|
|
|
# would be better to take a menu_id as argument
|
|
def get_tree(self, cr, uid, website_id, context=None):
|
|
def make_tree(node):
|
|
menu_node = dict(
|
|
id=node.id,
|
|
name=node.name,
|
|
url=node.url,
|
|
new_window=node.new_window,
|
|
sequence=node.sequence,
|
|
parent_id=node.parent_id.id,
|
|
children=[],
|
|
)
|
|
for child in node.child_id:
|
|
menu_node['children'].append(make_tree(child))
|
|
return menu_node
|
|
menu = self.pool.get('website').browse(cr, uid, website_id, context=context).menu_id
|
|
return make_tree(menu)
|
|
|
|
def save(self, cr, uid, website_id, data, context=None):
|
|
def replace_id(old_id, new_id):
|
|
for menu in data['data']:
|
|
if menu['id'] == old_id:
|
|
menu['id'] = new_id
|
|
if menu['parent_id'] == old_id:
|
|
menu['parent_id'] = new_id
|
|
to_delete = data['to_delete']
|
|
if to_delete:
|
|
self.unlink(cr, uid, to_delete, context=context)
|
|
for menu in data['data']:
|
|
mid = menu['id']
|
|
if isinstance(mid, str):
|
|
new_id = self.create(cr, uid, {'name': menu['name']}, context=context)
|
|
replace_id(mid, new_id)
|
|
for menu in data['data']:
|
|
self.write(cr, uid, [menu['id']], menu, context=context)
|
|
return True
|
|
|
|
class ir_attachment(osv.osv):
|
|
_inherit = "ir.attachment"
|
|
def _website_url_get(self, cr, uid, ids, name, arg, context=None):
|
|
result = {}
|
|
for attach in self.browse(cr, uid, ids, context=context):
|
|
if attach.type == 'url':
|
|
result[attach.id] = attach.url
|
|
else:
|
|
result[attach.id] = urlplus('/website/image', {
|
|
'model': 'ir.attachment',
|
|
'field': 'datas',
|
|
'id': attach.id,
|
|
'max_width': 1024,
|
|
'max_height': 768,
|
|
})
|
|
return result
|
|
def _datas_checksum(self, cr, uid, ids, name, arg, context=None):
|
|
return dict(
|
|
(attach['id'], self._compute_checksum(attach))
|
|
for attach in self.read(
|
|
cr, uid, ids, ['res_model', 'res_id', 'type', 'datas'],
|
|
context=context)
|
|
)
|
|
|
|
def _compute_checksum(self, attachment_dict):
|
|
if attachment_dict.get('res_model') == 'ir.ui.view'\
|
|
and not attachment_dict.get('res_id')\
|
|
and attachment_dict.get('type', 'binary') == 'binary'\
|
|
and attachment_dict.get('datas'):
|
|
return hashlib.new('sha1', attachment_dict['datas']).hexdigest()
|
|
return None
|
|
|
|
def _datas_big(self, cr, uid, ids, name, arg, context=None):
|
|
result = dict.fromkeys(ids, False)
|
|
if context and context.get('bin_size'):
|
|
return result
|
|
|
|
for record in self.browse(cr, uid, ids, context=context):
|
|
if not record.datas: continue
|
|
try:
|
|
result[record.id] = openerp.tools.image_resize_image_big(record.datas)
|
|
except IOError: # apparently the error PIL.Image.open raises
|
|
pass
|
|
|
|
return result
|
|
|
|
_columns = {
|
|
'datas_checksum': fields.function(_datas_checksum, size=40,
|
|
string="Datas checksum", type='char', store=True, select=True),
|
|
'website_url': fields.function(_website_url_get, string="Attachment URL", type='char'),
|
|
'datas_big': fields.function (_datas_big, type='binary', store=True,
|
|
string="Resized file content"),
|
|
}
|
|
|
|
def create(self, cr, uid, values, context=None):
|
|
chk = self._compute_checksum(values)
|
|
if chk:
|
|
match = self.search(cr, uid, [('datas_checksum', '=', chk)], context=context)
|
|
if match:
|
|
return match[0]
|
|
return super(ir_attachment, self).create(
|
|
cr, uid, values, context=context)
|
|
|
|
def try_remove(self, cr, uid, ids, context=None):
|
|
""" Removes a web-based image attachment if it is used by no view
|
|
(template)
|
|
|
|
Returns a dict mapping attachments which would not be removed (if any)
|
|
mapped to the views preventing their removal
|
|
"""
|
|
Views = self.pool['ir.ui.view']
|
|
attachments_to_remove = []
|
|
# views blocking removal of the attachment
|
|
removal_blocked_by = {}
|
|
|
|
for attachment in self.browse(cr, uid, ids, context=context):
|
|
# in-document URLs are html-escaped, a straight search will not
|
|
# find them
|
|
url = werkzeug.utils.escape(attachment.website_url)
|
|
ids = Views.search(cr, uid, [('arch', 'like', url)], context=context)
|
|
|
|
if ids:
|
|
removal_blocked_by[attachment.id] = Views.read(
|
|
cr, uid, ids, ['name'], context=context)
|
|
else:
|
|
attachments_to_remove.append(attachment.id)
|
|
if attachments_to_remove:
|
|
self.unlink(cr, uid, attachments_to_remove, context=context)
|
|
return removal_blocked_by
|
|
|
|
class res_partner(osv.osv):
|
|
_inherit = "res.partner"
|
|
|
|
def google_map_img(self, cr, uid, ids, zoom=8, width=298, height=298, context=None):
|
|
partner = self.browse(cr, uid, ids[0], context=context)
|
|
params = {
|
|
'center': '%s, %s %s, %s' % (partner.street or '', partner.city or '', partner.zip or '', partner.country_id and partner.country_id.name_get()[0][1] or ''),
|
|
'size': "%sx%s" % (height, width),
|
|
'zoom': zoom,
|
|
'sensor': 'false',
|
|
}
|
|
return urlplus('http://maps.googleapis.com/maps/api/staticmap' , params)
|
|
|
|
def google_map_link(self, cr, uid, ids, zoom=8, context=None):
|
|
partner = self.browse(cr, uid, ids[0], context=context)
|
|
params = {
|
|
'q': '%s, %s %s, %s' % (partner.street or '', partner.city or '', partner.zip or '', partner.country_id and partner.country_id.name_get()[0][1] or ''),
|
|
'z': 10
|
|
}
|
|
return urlplus('https://maps.google.com/maps' , params)
|
|
|
|
class res_company(osv.osv):
|
|
_inherit = "res.company"
|
|
def google_map_img(self, cr, uid, ids, zoom=8, width=298, height=298, context=None):
|
|
partner = self.browse(cr, openerp.SUPERUSER_ID, ids[0], context=context).partner_id
|
|
return partner and partner.google_map_img(zoom, width, height, context=context) or None
|
|
def google_map_link(self, cr, uid, ids, zoom=8, context=None):
|
|
partner = self.browse(cr, openerp.SUPERUSER_ID, ids[0], context=context).partner_id
|
|
return partner and partner.google_map_link(zoom, context=context) or None
|
|
|
|
class base_language_install(osv.osv_memory):
|
|
_inherit = "base.language.install"
|
|
_columns = {
|
|
'website_ids': fields.many2many('website', string='Websites to translate'),
|
|
}
|
|
|
|
def default_get(self, cr, uid, fields, context=None):
|
|
if context is None:
|
|
context = {}
|
|
defaults = super(base_language_install, self).default_get(cr, uid, fields, context)
|
|
website_id = context.get('params', {}).get('website_id')
|
|
if website_id:
|
|
if 'website_ids' not in defaults:
|
|
defaults['website_ids'] = []
|
|
defaults['website_ids'].append(website_id)
|
|
return defaults
|
|
|
|
def lang_install(self, cr, uid, ids, context=None):
|
|
if context is None:
|
|
context = {}
|
|
action = super(base_language_install, self).lang_install(cr, uid, ids, context)
|
|
language_obj = self.browse(cr, uid, ids)[0]
|
|
website_ids = [website.id for website in language_obj['website_ids']]
|
|
lang_id = self.pool['res.lang'].search(cr, uid, [('code', '=', language_obj['lang'])])
|
|
if website_ids and lang_id:
|
|
data = {'language_ids': [(4, lang_id[0])]}
|
|
self.pool['website'].write(cr, uid, website_ids, data)
|
|
params = context.get('params', {})
|
|
if 'url_return' in params:
|
|
return {
|
|
'url': params['url_return'].replace('[lang]', language_obj['lang']),
|
|
'type': 'ir.actions.act_url',
|
|
'target': 'self'
|
|
}
|
|
return action
|
|
|
|
class website_seo_metadata(osv.Model):
|
|
_name = 'website.seo.metadata'
|
|
_description = 'SEO metadata'
|
|
|
|
_columns = {
|
|
'website_meta_title': fields.char("Website meta title", translate=True),
|
|
'website_meta_description': fields.text("Website meta description", translate=True),
|
|
'website_meta_keywords': fields.char("Website meta keywords", translate=True),
|
|
}
|
|
|
|
# vim:et:
|