[IMP] set logging.getLogger(__name__)
bzr revid: kjo@tinyerp.com-20120622064854-5f6fss55k7305e8q
This commit is contained in:
parent
e02d8a68ef
commit
3c50eb017a
|
@ -29,6 +29,7 @@ import pooler
|
||||||
from osv import fields, osv
|
from osv import fields, osv
|
||||||
import decimal_precision as dp
|
import decimal_precision as dp
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
def check_cycle(self, cr, uid, ids, context=None):
|
def check_cycle(self, cr, uid, ids, context=None):
|
||||||
""" climbs the ``self._table.parent_id`` chains for 100 levels or
|
""" climbs the ``self._table.parent_id`` chains for 100 levels or
|
||||||
|
@ -294,7 +295,7 @@ class account_account(osv.osv):
|
||||||
if aml_query.strip():
|
if aml_query.strip():
|
||||||
wheres.append(aml_query.strip())
|
wheres.append(aml_query.strip())
|
||||||
filters = " AND ".join(wheres)
|
filters = " AND ".join(wheres)
|
||||||
logging.getLogger('account').debug('Filters: %s'%filters)
|
_logger.debug('Filters: %s'%filters)
|
||||||
# IN might not work ideally in case there are too many
|
# IN might not work ideally in case there are too many
|
||||||
# children_and_consolidated, in that case join on a
|
# children_and_consolidated, in that case join on a
|
||||||
# values() e.g.:
|
# values() e.g.:
|
||||||
|
@ -310,7 +311,7 @@ class account_account(osv.osv):
|
||||||
" GROUP BY l.account_id")
|
" GROUP BY l.account_id")
|
||||||
params = (tuple(children_and_consolidated),) + query_params
|
params = (tuple(children_and_consolidated),) + query_params
|
||||||
cr.execute(request, params)
|
cr.execute(request, params)
|
||||||
logging.getLogger('account').debug('Status: %s'%cr.statusmessage)
|
_logger.debug('Status: %s'%cr.statusmessage)
|
||||||
|
|
||||||
for res in cr.dictfetchall():
|
for res in cr.dictfetchall():
|
||||||
accounts[res['id']] = res
|
accounts[res['id']] = res
|
||||||
|
@ -2094,7 +2095,7 @@ class account_tax(osv.osv):
|
||||||
}
|
}
|
||||||
|
|
||||||
def compute(self, cr, uid, taxes, price_unit, quantity, product=None, partner=None):
|
def compute(self, cr, uid, taxes, price_unit, quantity, product=None, partner=None):
|
||||||
logging.getLogger('account').debug("Deprecated, use compute_all(...)['taxes'] instead of compute(...) to manage prices with tax included")
|
_logger.debug("Deprecated, use compute_all(...)['taxes'] instead of compute(...) to manage prices with tax included")
|
||||||
return self._compute(cr, uid, taxes, price_unit, quantity, product, partner)
|
return self._compute(cr, uid, taxes, price_unit, quantity, product, partner)
|
||||||
|
|
||||||
def _compute(self, cr, uid, taxes, price_unit, quantity, product=None, partner=None):
|
def _compute(self, cr, uid, taxes, price_unit, quantity, product=None, partner=None):
|
||||||
|
|
|
@ -30,11 +30,11 @@ from tools.translate import _
|
||||||
from osv import fields, osv
|
from osv import fields, osv
|
||||||
import netsvc
|
import netsvc
|
||||||
import tools
|
import tools
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class account_installer(osv.osv_memory):
|
class account_installer(osv.osv_memory):
|
||||||
_name = 'account.installer'
|
_name = 'account.installer'
|
||||||
_inherit = 'res.config.installer'
|
_inherit = 'res.config.installer'
|
||||||
__logger = logging.getLogger(_name)
|
|
||||||
|
|
||||||
def _get_charts(self, cr, uid, context=None):
|
def _get_charts(self, cr, uid, context=None):
|
||||||
modules = self.pool.get('ir.module.module')
|
modules = self.pool.get('ir.module.module')
|
||||||
|
@ -149,7 +149,7 @@ class account_installer(osv.osv_memory):
|
||||||
cr, uid, ids, context=context)
|
cr, uid, ids, context=context)
|
||||||
chart = self.read(cr, uid, ids, ['charts'],
|
chart = self.read(cr, uid, ids, ['charts'],
|
||||||
context=context)[0]['charts']
|
context=context)[0]['charts']
|
||||||
self.__logger.debug('Installing chart of accounts %s', chart)
|
self._logger.debug('Installing chart of accounts %s', chart)
|
||||||
return modules | set([chart])
|
return modules | set([chart])
|
||||||
|
|
||||||
account_installer()
|
account_installer()
|
||||||
|
|
|
@ -24,11 +24,12 @@ import time
|
||||||
from report import report_sxw
|
from report import report_sxw
|
||||||
import pooler
|
import pooler
|
||||||
import logging
|
import logging
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class bank_statement_balance_report(report_sxw.rml_parse):
|
class bank_statement_balance_report(report_sxw.rml_parse):
|
||||||
|
|
||||||
def set_context(self, objects, data, ids, report_type=None):
|
def set_context(self, objects, data, ids, report_type=None):
|
||||||
#logging('bank.statement.balance.report').warning('addons.'+__name__, 'set_context, objects = %s, data = %s, ids = %s' % (objects, data, ids))
|
#_logger.warning('addons.'+__name__, 'set_context, objects = %s, data = %s, ids = %s' % (objects, data, ids))
|
||||||
cr = self.cr
|
cr = self.cr
|
||||||
uid = self.uid
|
uid = self.uid
|
||||||
context = self.context
|
context = self.context
|
||||||
|
|
|
@ -28,6 +28,7 @@ import logging
|
||||||
import re
|
import re
|
||||||
from traceback import format_exception
|
from traceback import format_exception
|
||||||
from sys import exc_info
|
from sys import exc_info
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class account_coda_import(osv.osv_memory):
|
class account_coda_import(osv.osv_memory):
|
||||||
_name = 'account.coda.import'
|
_name = 'account.coda.import'
|
||||||
|
@ -815,7 +816,7 @@ class account_coda_import(osv.osv_memory):
|
||||||
ttype = line['type'] == 'supplier' and 'payment' or 'receipt',
|
ttype = line['type'] == 'supplier' and 'payment' or 'receipt',
|
||||||
date = line['val_date'],
|
date = line['val_date'],
|
||||||
context = context)
|
context = context)
|
||||||
#logging('account.coda').warning('voucher_dict = %s' % voucher_dict)
|
#_logger.warning('voucher_dict = %s' % voucher_dict)
|
||||||
voucher_line_vals = False
|
voucher_line_vals = False
|
||||||
if voucher_dict['value']['line_ids']:
|
if voucher_dict['value']['line_ids']:
|
||||||
for line_dict in voucher_dict['value']['line_ids']:
|
for line_dict in voucher_dict['value']['line_ids']:
|
||||||
|
@ -888,19 +889,19 @@ class account_coda_import(osv.osv_memory):
|
||||||
nb_err += 1
|
nb_err += 1
|
||||||
err_string += _('\nError ! ') + str(e)
|
err_string += _('\nError ! ') + str(e)
|
||||||
tb = ''.join(format_exception(*exc_info()))
|
tb = ''.join(format_exception(*exc_info()))
|
||||||
logging('account.coda').error('Application Error while processing Statement %s\n%s' % (statement.get('name', '/'),tb))
|
_logger.error('Application Error while processing Statement %s\n%s' % (statement.get('name', '/'),tb))
|
||||||
except Exception, e:
|
except Exception, e:
|
||||||
cr.rollback()
|
cr.rollback()
|
||||||
nb_err += 1
|
nb_err += 1
|
||||||
err_string += _('\nSystem Error : ') + str(e)
|
err_string += _('\nSystem Error : ') + str(e)
|
||||||
tb = ''.join(format_exception(*exc_info()))
|
tb = ''.join(format_exception(*exc_info()))
|
||||||
logging('account.coda').error('System Error while processing Statement %s\n%s' % (statement.get('name', '/'),tb))
|
_logger.error('System Error while processing Statement %s\n%s' % (statement.get('name', '/'),tb))
|
||||||
except :
|
except :
|
||||||
cr.rollback()
|
cr.rollback()
|
||||||
nb_err += 1
|
nb_err += 1
|
||||||
err_string = _('\nUnknown Error : ') + str(e)
|
err_string = _('\nUnknown Error : ') + str(e)
|
||||||
tb = ''.join(format_exception(*exc_info()))
|
tb = ''.join(format_exception(*exc_info()))
|
||||||
logging('account.coda').error('Unknown Error while processing Statement %s\n%s' % (statement.get('name', '/'),tb))
|
_logger.error('Unknown Error while processing Statement %s\n%s' % (statement.get('name', '/'),tb))
|
||||||
|
|
||||||
# end 'for statement in coda_statements'
|
# end 'for statement in coda_statements'
|
||||||
|
|
||||||
|
|
|
@ -44,8 +44,8 @@ from .. import utils
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
_logger = logging.getLogger('web.auth_openid')
|
_logger = logging.getLogger(__name__)
|
||||||
oidutil.log = logging.getLogger('openid').debug
|
oidutil.log = _logger.debug
|
||||||
|
|
||||||
|
|
||||||
class GoogleAppsAwareConsumer(consumer.GenericConsumer):
|
class GoogleAppsAwareConsumer(consumer.GenericConsumer):
|
||||||
|
|
|
@ -42,8 +42,10 @@ from osv import fields,osv
|
||||||
import pooler
|
import pooler
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
from service import security
|
from service import security
|
||||||
|
import logging
|
||||||
|
|
||||||
magic_md5 = '$1$'
|
magic_md5 = '$1$'
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
def gen_salt( length=8, symbols=ascii_letters + digits ):
|
def gen_salt( length=8, symbols=ascii_letters + digits ):
|
||||||
seed()
|
seed()
|
||||||
|
@ -179,8 +181,7 @@ class users(osv.osv):
|
||||||
cr = pooler.get_db(db).cursor()
|
cr = pooler.get_db(db).cursor()
|
||||||
return self._login(cr, db, login, password)
|
return self._login(cr, db, login, password)
|
||||||
except Exception:
|
except Exception:
|
||||||
import logging
|
_logger.exception('Could not authenticate')
|
||||||
logging.getLogger('netsvc').exception('Could not authenticate')
|
|
||||||
return Exception('Access Denied')
|
return Exception('Access Denied')
|
||||||
finally:
|
finally:
|
||||||
if cr is not None:
|
if cr is not None:
|
||||||
|
|
|
@ -27,21 +27,20 @@ LOG_INFO='info'
|
||||||
LOG_WARNING='warn'
|
LOG_WARNING='warn'
|
||||||
LOG_ERROR='error'
|
LOG_ERROR='error'
|
||||||
LOG_CRITICAL='critical'
|
LOG_CRITICAL='critical'
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
def log_detail(self):
|
def log_detail(self):
|
||||||
import os
|
import os
|
||||||
logger = logging.getLogger()
|
|
||||||
logfile_name = os.path.join(tempfile.gettempdir(), "openerp_report_designer.log")
|
logfile_name = os.path.join(tempfile.gettempdir(), "openerp_report_designer.log")
|
||||||
hdlr = logging.FileHandler(logfile_name)
|
hdlr = logging.FileHandler(logfile_name)
|
||||||
formatter = logging.Formatter('%(asctime)s %(levelname)s %(message)s')
|
formatter = logging.Formatter('%(asctime)s %(levelname)s %(message)s')
|
||||||
hdlr.setFormatter(formatter)
|
hdlr.setFormatter(formatter)
|
||||||
logger.addHandler(hdlr)
|
_logger.addHandler(hdlr)
|
||||||
logger.setLevel(logging.INFO)
|
_logger.setLevel(logging.INFO)
|
||||||
|
|
||||||
class Logger(object):
|
class Logger(object):
|
||||||
def log_write(self,name,level,msg):
|
def log_write(self,name,level,msg):
|
||||||
log = logging.getLogger(name)
|
getattr(_logger,level)(msg)
|
||||||
getattr(log,level)(msg)
|
|
||||||
|
|
||||||
def shutdown(self):
|
def shutdown(self):
|
||||||
logging.shutdown()
|
logging.shutdown()
|
||||||
|
|
|
@ -23,11 +23,12 @@ import logging
|
||||||
import string
|
import string
|
||||||
import datetime
|
import datetime
|
||||||
import re
|
import re
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import vatnumber
|
import vatnumber
|
||||||
except ImportError:
|
except ImportError:
|
||||||
logging.getLogger('base_vat').warning("VAT validation partially unavailable because the `vatnumber` Python library cannot be found. "
|
_logger.warning("VAT validation partially unavailable because the `vatnumber` Python library cannot be found. "
|
||||||
"Install it to support more countries, for example with `easy_install vatnumber`.")
|
"Install it to support more countries, for example with `easy_install vatnumber`.")
|
||||||
vatnumber = None
|
vatnumber = None
|
||||||
|
|
||||||
|
|
|
@ -23,6 +23,7 @@ from document_webdav import nodes
|
||||||
from document.nodes import _str2time, nodefd_static
|
from document.nodes import _str2time, nodefd_static
|
||||||
import logging
|
import logging
|
||||||
from orm_utils import get_last_modified
|
from orm_utils import get_last_modified
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from tools.dict_tools import dict_merge2
|
from tools.dict_tools import dict_merge2
|
||||||
|
@ -223,7 +224,6 @@ class node_calendar(nodes.node_class):
|
||||||
res = []
|
res = []
|
||||||
if not filters:
|
if not filters:
|
||||||
return res
|
return res
|
||||||
_log = logging.getLogger('caldav.query')
|
|
||||||
if filters.localName == 'calendar-query':
|
if filters.localName == 'calendar-query':
|
||||||
res = []
|
res = []
|
||||||
for filter_child in filters.childNodes:
|
for filter_child in filters.childNodes:
|
||||||
|
|
|
@ -34,6 +34,7 @@ import logging
|
||||||
from caldav_node import res_node_calendar
|
from caldav_node import res_node_calendar
|
||||||
from orm_utils import get_last_modified
|
from orm_utils import get_last_modified
|
||||||
from tools.safe_eval import safe_eval as eval
|
from tools.safe_eval import safe_eval as eval
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import vobject
|
import vobject
|
||||||
|
@ -240,7 +241,6 @@ def map_data(cr, uid, obj, context=None):
|
||||||
|
|
||||||
class CalDAV(object):
|
class CalDAV(object):
|
||||||
__attribute__ = {}
|
__attribute__ = {}
|
||||||
_logger = logging.getLogger('document.caldav')
|
|
||||||
|
|
||||||
def ical_set(self, name, value, type):
|
def ical_set(self, name, value, type):
|
||||||
""" set calendar Attribute
|
""" set calendar Attribute
|
||||||
|
|
|
@ -23,6 +23,7 @@ from osv import osv, fields
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
import caldav_node
|
import caldav_node
|
||||||
import logging
|
import logging
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class calendar_collection(osv.osv):
|
class calendar_collection(osv.osv):
|
||||||
_inherit = 'document.directory'
|
_inherit = 'document.directory'
|
||||||
|
@ -44,8 +45,7 @@ class calendar_collection(osv.osv):
|
||||||
root_cal_dir = self.browse(cr,uid, root_id, context=context)
|
root_cal_dir = self.browse(cr,uid, root_id, context=context)
|
||||||
return root_cal_dir.name
|
return root_cal_dir.name
|
||||||
except Exception:
|
except Exception:
|
||||||
logger = logging.getLogger('document')
|
_logger.warning('Cannot set root directory for Calendars:', exc_info=True)
|
||||||
logger.warning('Cannot set root directory for Calendars:', exc_info=True)
|
|
||||||
return False
|
return False
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
|
|
@ -26,6 +26,7 @@ import logging
|
||||||
from osv import fields, osv
|
from osv import fields, osv
|
||||||
import tools
|
import tools
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class crm_lead(base_stage, osv.osv):
|
class crm_lead(base_stage, osv.osv):
|
||||||
""" CRM Leads """
|
""" CRM Leads """
|
||||||
|
@ -180,7 +181,7 @@ class res_users(osv.osv):
|
||||||
'user_id': user_id}, context=context)
|
'user_id': user_id}, context=context)
|
||||||
except:
|
except:
|
||||||
# Tolerate a missing shortcut. See product/product.py for similar code.
|
# Tolerate a missing shortcut. See product/product.py for similar code.
|
||||||
logging.getLogger('orm').debug('Skipped meetings shortcut for user "%s"', data.get('name','<new'))
|
_logger.debug('Skipped meetings shortcut for user "%s"', data.get('name','<new'))
|
||||||
return user_id
|
return user_id
|
||||||
|
|
||||||
res_users()
|
res_users()
|
||||||
|
|
|
@ -23,12 +23,13 @@ import logging
|
||||||
import models
|
import models
|
||||||
import edi_service
|
import edi_service
|
||||||
from models.edi import EDIMixin, edi_document
|
from models.edi import EDIMixin, edi_document
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
# web
|
# web
|
||||||
try:
|
try:
|
||||||
import controllers
|
import controllers
|
||||||
except ImportError:
|
except ImportError:
|
||||||
logging.getLogger('init.load').warn(
|
_logger.warn(
|
||||||
"""Could not load openerp-web section of EDI, EDI will not behave correctly
|
"""Could not load openerp-web section of EDI, EDI will not behave correctly
|
||||||
|
|
||||||
To fix, launch openerp-web in embedded mode""")
|
To fix, launch openerp-web in embedded mode""")
|
||||||
|
|
|
@ -23,7 +23,7 @@ import logging
|
||||||
import netsvc
|
import netsvc
|
||||||
import openerp
|
import openerp
|
||||||
|
|
||||||
_logger = logging.getLogger('edi.service')
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class edi(netsvc.ExportService):
|
class edi(netsvc.ExportService):
|
||||||
|
|
||||||
|
|
|
@ -35,6 +35,7 @@ import pooler
|
||||||
from osv import osv,fields,orm
|
from osv import osv,fields,orm
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
from tools.safe_eval import safe_eval as eval
|
from tools.safe_eval import safe_eval as eval
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
EXTERNAL_ID_PATTERN = re.compile(r'^([^.:]+)(?::([^.]+))?\.(\S+)$')
|
EXTERNAL_ID_PATTERN = re.compile(r'^([^.:]+)(?::([^.]+))?\.(\S+)$')
|
||||||
EDI_VIEW_WEB_URL = '%s/edi/view?db=%s&token=%s'
|
EDI_VIEW_WEB_URL = '%s/edi/view?db=%s&token=%s'
|
||||||
|
@ -72,7 +73,6 @@ def last_update_for(record):
|
||||||
return record_log.get('write_date') or record_log.get('create_date') or False
|
return record_log.get('write_date') or record_log.get('create_date') or False
|
||||||
return False
|
return False
|
||||||
|
|
||||||
_logger = logging.getLogger('edi')
|
|
||||||
|
|
||||||
class edi_document(osv.osv):
|
class edi_document(osv.osv):
|
||||||
_name = 'edi.document'
|
_name = 'edi.document'
|
||||||
|
|
|
@ -24,6 +24,7 @@ from osv import fields,osv
|
||||||
from edi import EDIMixin
|
from edi import EDIMixin
|
||||||
from openerp import SUPERUSER_ID
|
from openerp import SUPERUSER_ID
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
RES_PARTNER_EDI_STRUCT = {
|
RES_PARTNER_EDI_STRUCT = {
|
||||||
'name': True,
|
'name': True,
|
||||||
|
@ -63,7 +64,7 @@ class res_partner(osv.osv, EDIMixin):
|
||||||
code, label = 'edi_generic', 'Generic Bank Type (auto-created for EDI)'
|
code, label = 'edi_generic', 'Generic Bank Type (auto-created for EDI)'
|
||||||
bank_code_ids = res_partner_bank_type.search(cr, uid, [('code','=',code)], context=context)
|
bank_code_ids = res_partner_bank_type.search(cr, uid, [('code','=',code)], context=context)
|
||||||
if not bank_code_ids:
|
if not bank_code_ids:
|
||||||
logging.getLogger('edi.res_partner').info('Normal bank account type is missing, creating '
|
_logger.info('Normal bank account type is missing, creating '
|
||||||
'a generic bank account type for EDI.')
|
'a generic bank account type for EDI.')
|
||||||
self.res_partner_bank_type.create(cr, SUPERUSER_ID, {'name': label,
|
self.res_partner_bank_type.create(cr, SUPERUSER_ID, {'name': label,
|
||||||
'code': label})
|
'code': label})
|
||||||
|
@ -84,7 +85,7 @@ class res_partner(osv.osv, EDIMixin):
|
||||||
bank_name, ext_bank_id, context=import_ctx)
|
bank_name, ext_bank_id, context=import_ctx)
|
||||||
except osv.except_osv:
|
except osv.except_osv:
|
||||||
# failed to import it, try again with unrestricted default type
|
# failed to import it, try again with unrestricted default type
|
||||||
logging.getLogger('edi.res_partner').warning('Failed to import bank account using'
|
_logger.warning('Failed to import bank account using'
|
||||||
'bank type: %s, ignoring', import_ctx['default_state'],
|
'bank type: %s, ignoring', import_ctx['default_state'],
|
||||||
exc_info=True)
|
exc_info=True)
|
||||||
return contact_id
|
return contact_id
|
||||||
|
|
|
@ -29,11 +29,12 @@ from osv import fields
|
||||||
import tools
|
import tools
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
from urllib import quote as quote
|
from urllib import quote as quote
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from mako.template import Template as MakoTemplate
|
from mako.template import Template as MakoTemplate
|
||||||
except ImportError:
|
except ImportError:
|
||||||
logging.getLogger('init').warning("email_template: mako templates not available, templating features will not work!")
|
_logger.warning("email_template: mako templates not available, templating features will not work!")
|
||||||
|
|
||||||
class email_template(osv.osv):
|
class email_template(osv.osv):
|
||||||
"Templates for sending email"
|
"Templates for sending email"
|
||||||
|
@ -75,7 +76,7 @@ class email_template(osv.osv):
|
||||||
result = u''
|
result = u''
|
||||||
return result
|
return result
|
||||||
except Exception:
|
except Exception:
|
||||||
logging.exception("failed to render mako template value %r", template)
|
_logger.exception("failed to render mako template value %r", template)
|
||||||
return u""
|
return u""
|
||||||
|
|
||||||
def get_email_template(self, cr, uid, template_id=False, record_id=None, context=None):
|
def get_email_template(self, cr, uid, template_id=False, record_id=None, context=None):
|
||||||
|
|
|
@ -39,7 +39,7 @@ from osv import osv, fields
|
||||||
import tools
|
import tools
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
|
|
||||||
logger = logging.getLogger('fetchmail')
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class fetchmail_server(osv.osv):
|
class fetchmail_server(osv.osv):
|
||||||
"""Incoming POP/IMAP mail server account"""
|
"""Incoming POP/IMAP mail server account"""
|
||||||
|
|
|
@ -25,6 +25,7 @@ import logging
|
||||||
from osv import fields, osv
|
from osv import fields, osv
|
||||||
from PIL import Image
|
from PIL import Image
|
||||||
import StringIO
|
import StringIO
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class hr_employee_category(osv.osv):
|
class hr_employee_category(osv.osv):
|
||||||
|
|
||||||
|
@ -304,7 +305,7 @@ class res_users(osv.osv):
|
||||||
'user_id': user_id}, context=context)
|
'user_id': user_id}, context=context)
|
||||||
except:
|
except:
|
||||||
# Tolerate a missing shortcut. See product/product.py for similar code.
|
# Tolerate a missing shortcut. See product/product.py for similar code.
|
||||||
logging.getLogger('orm').debug('Skipped meetings shortcut for user "%s"', data.get('name','<new'))
|
_logger.debug('Skipped meetings shortcut for user "%s"', data.get('name','<new'))
|
||||||
|
|
||||||
return user_id
|
return user_id
|
||||||
|
|
||||||
|
|
|
@ -30,7 +30,7 @@ import logging
|
||||||
import StringIO
|
import StringIO
|
||||||
import traceback
|
import traceback
|
||||||
pp = pprint.PrettyPrinter(indent=4)
|
pp = pprint.PrettyPrinter(indent=4)
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
@ -60,7 +60,6 @@ class import_framework(Thread):
|
||||||
self.context = context or {}
|
self.context = context or {}
|
||||||
self.email = email_to_notify
|
self.email = email_to_notify
|
||||||
self.table_list = []
|
self.table_list = []
|
||||||
self.logger = logging.getLogger(module_name)
|
|
||||||
self.initialize()
|
self.initialize()
|
||||||
|
|
||||||
"""
|
"""
|
||||||
|
@ -165,7 +164,7 @@ class import_framework(Thread):
|
||||||
data_i is a map external field_name => value
|
data_i is a map external field_name => value
|
||||||
and each data_i have a external id => in data_id['id']
|
and each data_i have a external id => in data_id['id']
|
||||||
"""
|
"""
|
||||||
self.logger.info(' Importing %s into %s' % (table, model))
|
_logger.info(' Importing %s into %s' % (table, model))
|
||||||
if not datas:
|
if not datas:
|
||||||
return (0, 'No data found')
|
return (0, 'No data found')
|
||||||
mapping['id'] = 'id_new'
|
mapping['id'] = 'id_new'
|
||||||
|
@ -188,7 +187,7 @@ class import_framework(Thread):
|
||||||
model_obj = self.obj.pool.get(model)
|
model_obj = self.obj.pool.get(model)
|
||||||
if not model_obj:
|
if not model_obj:
|
||||||
raise ValueError(_("%s is not a valid model name") % model)
|
raise ValueError(_("%s is not a valid model name") % model)
|
||||||
self.logger.debug(_(" fields imported : ") + str(fields))
|
_logger.debug(_(" fields imported : ") + str(fields))
|
||||||
(p, r, warning, s) = model_obj.import_data(self.cr, self.uid, fields, res, mode='update', current_module=self.module_name, noupdate=True, context=self.context)
|
(p, r, warning, s) = model_obj.import_data(self.cr, self.uid, fields, res, mode='update', current_module=self.module_name, noupdate=True, context=self.context)
|
||||||
for (field, field_name) in self_dependencies:
|
for (field, field_name) in self_dependencies:
|
||||||
self._import_self_dependencies(model_obj, field, datas)
|
self._import_self_dependencies(model_obj, field, datas)
|
||||||
|
@ -431,9 +430,9 @@ class import_framework(Thread):
|
||||||
'auto_delete' : True})
|
'auto_delete' : True})
|
||||||
email_obj.send(self.cr, self.uid, [email_id])
|
email_obj.send(self.cr, self.uid, [email_id])
|
||||||
if error:
|
if error:
|
||||||
self.logger.error(_("Import failed due to an unexpected error"))
|
_logger.error(_("Import failed due to an unexpected error"))
|
||||||
else:
|
else:
|
||||||
self.logger.info(_("Import finished, notification email sended"))
|
_logger.info(_("Import finished, notification email sended"))
|
||||||
|
|
||||||
def get_email_subject(self, result, error=False):
|
def get_email_subject(self, result, error=False):
|
||||||
"""
|
"""
|
||||||
|
|
|
@ -33,7 +33,7 @@ import import_sugarcrm
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
import sys
|
import sys
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
debug = False
|
debug = False
|
||||||
|
|
||||||
|
@ -119,7 +119,7 @@ def get_contact_by_email(portType, username, password, email_address=None):
|
||||||
email_list.append(list.Email_address)
|
email_list.append(list.Email_address)
|
||||||
return email_list
|
return email_list
|
||||||
except Exception,e:
|
except Exception,e:
|
||||||
logging.getLogger('sugarcrm_soap').error('Exception: %s\n' % (tools.ustr(e)))
|
_logger.error('Exception: %s\n' % (tools.ustr(e)))
|
||||||
return False
|
return False
|
||||||
|
|
||||||
def get_document_revision_search(portType, sessionid, module_id=None):
|
def get_document_revision_search(portType, sessionid, module_id=None):
|
||||||
|
|
|
@ -24,6 +24,7 @@ import re, time, random
|
||||||
from osv import fields, osv
|
from osv import fields, osv
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
import logging
|
import logging
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
"""
|
"""
|
||||||
account.invoice object:
|
account.invoice object:
|
||||||
|
@ -40,7 +41,7 @@ class account_invoice(osv.osv):
|
||||||
context=context)
|
context=context)
|
||||||
res[[i for i,x in enumerate(res) if x[0] == 'none'][0]] = ('none', 'Free Communication')
|
res[[i for i,x in enumerate(res) if x[0] == 'none'][0]] = ('none', 'Free Communication')
|
||||||
res.append(('bba', 'BBA Structured Communication'))
|
res.append(('bba', 'BBA Structured Communication'))
|
||||||
#logging('l1on.be.invoice.bba').warning('reference_type = %s' %res )
|
#l_logger.warning('reference_type = %s' %res )
|
||||||
return res
|
return res
|
||||||
|
|
||||||
def check_bbacomm(self, val):
|
def check_bbacomm(self, val):
|
||||||
|
@ -67,7 +68,7 @@ class account_invoice(osv.osv):
|
||||||
result = super(account_invoice, self).onchange_partner_id(cr, uid, ids, type, partner_id,
|
result = super(account_invoice, self).onchange_partner_id(cr, uid, ids, type, partner_id,
|
||||||
date_invoice, payment_term, partner_bank_id, company_id)
|
date_invoice, payment_term, partner_bank_id, company_id)
|
||||||
# reference_type = self.default_get(cr, uid, ['reference_type'])['reference_type']
|
# reference_type = self.default_get(cr, uid, ['reference_type'])['reference_type']
|
||||||
# logging('l1on.be.invoice.bba').warning('partner_id %s' % partner_id)
|
# _logger.warning('partner_id %s' % partner_id)
|
||||||
reference = False
|
reference = False
|
||||||
reference_type = 'none'
|
reference_type = 'none'
|
||||||
if partner_id:
|
if partner_id:
|
||||||
|
|
|
@ -23,6 +23,7 @@ from osv import fields, osv
|
||||||
import os
|
import os
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
import logging
|
import logging
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class wizard_multi_charts_accounts(osv.osv_memory):
|
class wizard_multi_charts_accounts(osv.osv_memory):
|
||||||
"""
|
"""
|
||||||
|
@ -79,7 +80,7 @@ class wizard_multi_charts_accounts(osv.osv_memory):
|
||||||
if context.get('lang') == lang:
|
if context.get('lang') == lang:
|
||||||
self.pool.get(out_obj._name).write(cr, uid, out_ids[j], {in_field: value[in_id]})
|
self.pool.get(out_obj._name).write(cr, uid, out_ids[j], {in_field: value[in_id]})
|
||||||
else:
|
else:
|
||||||
logging.getLogger('l10n.multilang').info('Language: %s. Translation from template: there is no translation available for %s!' %(lang, src[in_id]))#out_obj._name))
|
_logger.info('Language: %s. Translation from template: there is no translation available for %s!' %(lang, src[in_id]))#out_obj._name))
|
||||||
return True
|
return True
|
||||||
|
|
||||||
def execute(self, cr, uid, ids, context=None):
|
def execute(self, cr, uid, ids, context=None):
|
||||||
|
|
|
@ -36,7 +36,7 @@ from osv import fields
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
from openerp import SUPERUSER_ID
|
from openerp import SUPERUSER_ID
|
||||||
|
|
||||||
_logger = logging.getLogger('mail')
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
def format_date_tz(date, tz=None):
|
def format_date_tz(date, tz=None):
|
||||||
if not date:
|
if not date:
|
||||||
|
|
|
@ -27,7 +27,7 @@ from tools.misc import email_re
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
|
|
||||||
from base.res.res_users import _lang_get
|
from base.res.res_users import _lang_get
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
# welcome email sent to new portal users (note that calling tools.translate._
|
# welcome email sent to new portal users (note that calling tools.translate._
|
||||||
|
@ -174,7 +174,7 @@ class wizard(osv.osv_memory):
|
||||||
body = _(WELCOME_EMAIL_BODY) % data
|
body = _(WELCOME_EMAIL_BODY) % data
|
||||||
res = mail_message_obj.schedule_with_attach(cr, uid, email_from , [email_to], subject, body, context=context)
|
res = mail_message_obj.schedule_with_attach(cr, uid, email_from , [email_to], subject, body, context=context)
|
||||||
if not res:
|
if not res:
|
||||||
logging.getLogger('res.portal.wizard').warning(
|
_logger.warning(
|
||||||
'Failed to send email from %s to %s', email_from, email_to)
|
'Failed to send email from %s to %s', email_from, email_to)
|
||||||
|
|
||||||
return {'type': 'ir.actions.act_window_close'}
|
return {'type': 'ir.actions.act_window_close'}
|
||||||
|
|
|
@ -51,7 +51,7 @@ import tools
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
from osv.osv import except_osv
|
from osv.osv import except_osv
|
||||||
|
|
||||||
logger = logging.getLogger('report_webkit')
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
def mako_template(text):
|
def mako_template(text):
|
||||||
"""Build a Mako template.
|
"""Build a Mako template.
|
||||||
|
@ -248,7 +248,7 @@ class WebKitParser(report_sxw):
|
||||||
htmls.append(html)
|
htmls.append(html)
|
||||||
except Exception, e:
|
except Exception, e:
|
||||||
msg = exceptions.text_error_template().render()
|
msg = exceptions.text_error_template().render()
|
||||||
logger.error(msg)
|
_logger.error(msg)
|
||||||
raise except_osv(_('Webkit render'), msg)
|
raise except_osv(_('Webkit render'), msg)
|
||||||
else:
|
else:
|
||||||
try :
|
try :
|
||||||
|
@ -259,7 +259,7 @@ class WebKitParser(report_sxw):
|
||||||
htmls.append(html)
|
htmls.append(html)
|
||||||
except Exception, e:
|
except Exception, e:
|
||||||
msg = exceptions.text_error_template().render()
|
msg = exceptions.text_error_template().render()
|
||||||
logger.error(msg)
|
_logger.error(msg)
|
||||||
raise except_osv(_('Webkit render'), msg)
|
raise except_osv(_('Webkit render'), msg)
|
||||||
head_mako_tpl = mako_template(header)
|
head_mako_tpl = mako_template(header)
|
||||||
try :
|
try :
|
||||||
|
@ -281,7 +281,7 @@ class WebKitParser(report_sxw):
|
||||||
**self.parser_instance.localcontext)
|
**self.parser_instance.localcontext)
|
||||||
except:
|
except:
|
||||||
msg = exceptions.text_error_template().render()
|
msg = exceptions.text_error_template().render()
|
||||||
logger.error(msg)
|
_logger.error(msg)
|
||||||
raise except_osv(_('Webkit render'), msg)
|
raise except_osv(_('Webkit render'), msg)
|
||||||
if report_xml.webkit_debug :
|
if report_xml.webkit_debug :
|
||||||
try :
|
try :
|
||||||
|
@ -292,7 +292,7 @@ class WebKitParser(report_sxw):
|
||||||
**self.parser_instance.localcontext)
|
**self.parser_instance.localcontext)
|
||||||
except Exception, e:
|
except Exception, e:
|
||||||
msg = exceptions.text_error_template().render()
|
msg = exceptions.text_error_template().render()
|
||||||
logger.error(msg)
|
_logger.error(msg)
|
||||||
raise except_osv(_('Webkit render'), msg)
|
raise except_osv(_('Webkit render'), msg)
|
||||||
return (deb, 'html')
|
return (deb, 'html')
|
||||||
bin = self.get_lib(cursor, uid)
|
bin = self.get_lib(cursor, uid)
|
||||||
|
|
|
@ -32,6 +32,7 @@ from osv import expression
|
||||||
from tools.translate import _
|
from tools.translate import _
|
||||||
from tools.safe_eval import safe_eval
|
from tools.safe_eval import safe_eval
|
||||||
import openerp
|
import openerp
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
FULL_ACCESS = ('perm_read', 'perm_write', 'perm_create', 'perm_unlink')
|
FULL_ACCESS = ('perm_read', 'perm_write', 'perm_create', 'perm_unlink')
|
||||||
READ_WRITE_ACCESS = ('perm_read', 'perm_write')
|
READ_WRITE_ACCESS = ('perm_read', 'perm_write')
|
||||||
|
@ -48,7 +49,6 @@ def generate_random_pass():
|
||||||
return ''.join(random.sample(RANDOM_PASS_CHARACTERS,10))
|
return ''.join(random.sample(RANDOM_PASS_CHARACTERS,10))
|
||||||
|
|
||||||
class share_wizard(osv.osv_memory):
|
class share_wizard(osv.osv_memory):
|
||||||
_logger = logging.getLogger('share.wizard')
|
|
||||||
_name = 'share.wizard'
|
_name = 'share.wizard'
|
||||||
_description = 'Share Wizard'
|
_description = 'Share Wizard'
|
||||||
|
|
||||||
|
@ -335,7 +335,7 @@ class share_wizard(osv.osv_memory):
|
||||||
except Exception:
|
except Exception:
|
||||||
# Note: must catch all exceptions, as UnquoteEvalContext may cause many
|
# Note: must catch all exceptions, as UnquoteEvalContext may cause many
|
||||||
# different exceptions, as it shadows builtins.
|
# different exceptions, as it shadows builtins.
|
||||||
self._logger.debug("Failed to cleanup action context as it does not parse server-side", exc_info=True)
|
_logger.debug("Failed to cleanup action context as it does not parse server-side", exc_info=True)
|
||||||
result = context_str
|
result = context_str
|
||||||
return result
|
return result
|
||||||
|
|
||||||
|
@ -496,8 +496,8 @@ class share_wizard(osv.osv_memory):
|
||||||
[x.id for x in current_user.groups_id], target_model_ids, context=context)
|
[x.id for x in current_user.groups_id], target_model_ids, context=context)
|
||||||
group_access_map = self._get_access_map_for_groups_and_models(cr, uid,
|
group_access_map = self._get_access_map_for_groups_and_models(cr, uid,
|
||||||
[group_id], target_model_ids, context=context)
|
[group_id], target_model_ids, context=context)
|
||||||
self._logger.debug("Current user access matrix: %r", current_user_access_map)
|
_logger.debug("Current user access matrix: %r", current_user_access_map)
|
||||||
self._logger.debug("New group current access matrix: %r", group_access_map)
|
_logger.debug("New group current access matrix: %r", group_access_map)
|
||||||
|
|
||||||
# Create required rights if allowed by current user rights and not
|
# Create required rights if allowed by current user rights and not
|
||||||
# already granted
|
# already granted
|
||||||
|
@ -520,7 +520,7 @@ class share_wizard(osv.osv_memory):
|
||||||
need_creation = True
|
need_creation = True
|
||||||
if need_creation:
|
if need_creation:
|
||||||
model_access_obj.create(cr, UID_ROOT, values)
|
model_access_obj.create(cr, UID_ROOT, values)
|
||||||
self._logger.debug("Creating access right for model %s with values: %r", model.model, values)
|
_logger.debug("Creating access right for model %s with values: %r", model.model, values)
|
||||||
|
|
||||||
def _link_or_copy_current_user_rules(self, cr, current_user, group_id, fields_relations, context=None):
|
def _link_or_copy_current_user_rules(self, cr, current_user, group_id, fields_relations, context=None):
|
||||||
rule_obj = self.pool.get('ir.rule')
|
rule_obj = self.pool.get('ir.rule')
|
||||||
|
@ -542,13 +542,13 @@ class share_wizard(osv.osv_memory):
|
||||||
'groups': [(6,0,[group_id])],
|
'groups': [(6,0,[group_id])],
|
||||||
'domain_force': rule.domain, # evaluated version!
|
'domain_force': rule.domain, # evaluated version!
|
||||||
})
|
})
|
||||||
self._logger.debug("Copying rule %s (%s) on model %s with domain: %s", rule.name, rule.id, model.model, rule.domain_force)
|
_logger.debug("Copying rule %s (%s) on model %s with domain: %s", rule.name, rule.id, model.model, rule.domain_force)
|
||||||
else:
|
else:
|
||||||
# otherwise we can simply link the rule to keep it dynamic
|
# otherwise we can simply link the rule to keep it dynamic
|
||||||
rule_obj.write(cr, 1, [rule.id], {
|
rule_obj.write(cr, 1, [rule.id], {
|
||||||
'groups': [(4,group_id)]
|
'groups': [(4,group_id)]
|
||||||
})
|
})
|
||||||
self._logger.debug("Linking rule %s (%s) on model %s with domain: %s", rule.name, rule.id, model.model, rule.domain_force)
|
_logger.debug("Linking rule %s (%s) on model %s with domain: %s", rule.name, rule.id, model.model, rule.domain_force)
|
||||||
|
|
||||||
def _check_personal_rule_or_duplicate(self, cr, group_id, rule, context=None):
|
def _check_personal_rule_or_duplicate(self, cr, group_id, rule, context=None):
|
||||||
"""Verifies that the given rule only belongs to the given group_id, otherwise
|
"""Verifies that the given rule only belongs to the given group_id, otherwise
|
||||||
|
@ -567,7 +567,7 @@ class share_wizard(osv.osv_memory):
|
||||||
'groups': [(6,0,[group_id])],
|
'groups': [(6,0,[group_id])],
|
||||||
'domain_force': rule.domain_force, # non evaluated!
|
'domain_force': rule.domain_force, # non evaluated!
|
||||||
})
|
})
|
||||||
self._logger.debug("Duplicating rule %s (%s) (domain: %s) for modified access ", rule.name, rule.id, rule.domain_force)
|
_logger.debug("Duplicating rule %s (%s) (domain: %s) for modified access ", rule.name, rule.id, rule.domain_force)
|
||||||
# then disconnect from group_id:
|
# then disconnect from group_id:
|
||||||
rule.write({'groups':[(3,group_id)]}) # disconnects, does not delete!
|
rule.write({'groups':[(3,group_id)]}) # disconnects, does not delete!
|
||||||
return rule_obj.browse(cr, UID_ROOT, new_id, context=context)
|
return rule_obj.browse(cr, UID_ROOT, new_id, context=context)
|
||||||
|
@ -602,7 +602,7 @@ class share_wizard(osv.osv_memory):
|
||||||
if restrict:
|
if restrict:
|
||||||
continue
|
continue
|
||||||
else:
|
else:
|
||||||
self._logger.debug("Ignoring sharing rule on model %s with domain: %s the same rule exists already", model_id, domain)
|
_logger.debug("Ignoring sharing rule on model %s with domain: %s the same rule exists already", model_id, domain)
|
||||||
return
|
return
|
||||||
if restrict:
|
if restrict:
|
||||||
# restricting existing rules is done by adding the clause
|
# restricting existing rules is done by adding the clause
|
||||||
|
@ -614,7 +614,7 @@ class share_wizard(osv.osv_memory):
|
||||||
new_clause = expression.normalize(eval(domain, eval_ctx))
|
new_clause = expression.normalize(eval(domain, eval_ctx))
|
||||||
combined_domain = expression.AND([new_clause, org_domain])
|
combined_domain = expression.AND([new_clause, org_domain])
|
||||||
rule.write({'domain_force': combined_domain, 'name': rule.name + _('(Modified)')})
|
rule.write({'domain_force': combined_domain, 'name': rule.name + _('(Modified)')})
|
||||||
self._logger.debug("Combining sharing rule %s on model %s with domain: %s", rule.id, model_id, domain)
|
_logger.debug("Combining sharing rule %s on model %s with domain: %s", rule.id, model_id, domain)
|
||||||
if not rule_ids or not restrict:
|
if not rule_ids or not restrict:
|
||||||
# Adding the new rule in the group is ok for normal cases, because rules
|
# Adding the new rule in the group is ok for normal cases, because rules
|
||||||
# in the same group and for the same model will be combined with OR
|
# in the same group and for the same model will be combined with OR
|
||||||
|
@ -625,7 +625,7 @@ class share_wizard(osv.osv_memory):
|
||||||
'domain_force': domain,
|
'domain_force': domain,
|
||||||
'groups': [(4,group_id)]
|
'groups': [(4,group_id)]
|
||||||
})
|
})
|
||||||
self._logger.debug("Created sharing rule on model %s with domain: %s", model_id, domain)
|
_logger.debug("Created sharing rule on model %s with domain: %s", model_id, domain)
|
||||||
|
|
||||||
def _create_indirect_sharing_rules(self, cr, current_user, wizard_data, group_id, fields_relations, context=None):
|
def _create_indirect_sharing_rules(self, cr, current_user, wizard_data, group_id, fields_relations, context=None):
|
||||||
rule_name = _('Indirect sharing filter created by user %s (%s) for group %s') % \
|
rule_name = _('Indirect sharing filter created by user %s (%s) for group %s') % \
|
||||||
|
@ -648,7 +648,7 @@ class share_wizard(osv.osv_memory):
|
||||||
group_id, model_id=model.id, domain=str(related_domain),
|
group_id, model_id=model.id, domain=str(related_domain),
|
||||||
rule_name=rule_name, restrict=True, context=context)
|
rule_name=rule_name, restrict=True, context=context)
|
||||||
except Exception:
|
except Exception:
|
||||||
self._logger.exception('Failed to create share access')
|
_logger.exception('Failed to create share access')
|
||||||
raise osv.except_osv(_('Sharing access could not be created'),
|
raise osv.except_osv(_('Sharing access could not be created'),
|
||||||
_('Sorry, the current screen and filter you are trying to share are not supported at the moment.\nYou may want to try a simpler filter.'))
|
_('Sorry, the current screen and filter you are trying to share are not supported at the moment.\nYou may want to try a simpler filter.'))
|
||||||
|
|
||||||
|
@ -852,7 +852,7 @@ class share_wizard(osv.osv_memory):
|
||||||
notification_obj.create(cr, uid, {'user_id': result_line.user_id.id, 'message_id': msg_id}, context=context)
|
notification_obj.create(cr, uid, {'user_id': result_line.user_id.id, 'message_id': msg_id}, context=context)
|
||||||
|
|
||||||
def send_emails(self, cr, uid, wizard_data, context=None):
|
def send_emails(self, cr, uid, wizard_data, context=None):
|
||||||
self._logger.info('Sending share notifications by email...')
|
_logger.info('Sending share notifications by email...')
|
||||||
mail_message = self.pool.get('mail.message')
|
mail_message = self.pool.get('mail.message')
|
||||||
user = self.pool.get('res.users').browse(cr, UID_ROOT, uid)
|
user = self.pool.get('res.users').browse(cr, UID_ROOT, uid)
|
||||||
if not user.user_email:
|
if not user.user_email:
|
||||||
|
@ -885,7 +885,7 @@ class share_wizard(osv.osv_memory):
|
||||||
msg_ids.append(mail_message.schedule_with_attach(cr, uid, user.user_email, [email_to], subject, body, model='share.wizard', context=context))
|
msg_ids.append(mail_message.schedule_with_attach(cr, uid, user.user_email, [email_to], subject, body, model='share.wizard', context=context))
|
||||||
# force direct delivery, as users expect instant notification
|
# force direct delivery, as users expect instant notification
|
||||||
mail_message.send(cr, uid, msg_ids, context=context)
|
mail_message.send(cr, uid, msg_ids, context=context)
|
||||||
self._logger.info('%d share notification(s) sent.', len(msg_ids))
|
_logger.info('%d share notification(s) sent.', len(msg_ids))
|
||||||
|
|
||||||
def onchange_embed_options(self, cr, uid, ids, opt_title, opt_search, context=None):
|
def onchange_embed_options(self, cr, uid, ids, opt_title, opt_search, context=None):
|
||||||
wizard = self.browse(cr, uid, ids[0], context)
|
wizard = self.browse(cr, uid, ids[0], context)
|
||||||
|
|
|
@ -33,7 +33,7 @@ import tools
|
||||||
from tools import float_compare
|
from tools import float_compare
|
||||||
import decimal_precision as dp
|
import decimal_precision as dp
|
||||||
import logging
|
import logging
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
#----------------------------------------------------------
|
#----------------------------------------------------------
|
||||||
# Incoterms
|
# Incoterms
|
||||||
|
@ -419,9 +419,8 @@ class stock_location(osv.osv):
|
||||||
# so we ROLLBACK to the SAVEPOINT to restore the transaction to its earlier
|
# so we ROLLBACK to the SAVEPOINT to restore the transaction to its earlier
|
||||||
# state, we return False as if the products were not available, and log it:
|
# state, we return False as if the products were not available, and log it:
|
||||||
cr.execute("ROLLBACK TO stock_location_product_reserve")
|
cr.execute("ROLLBACK TO stock_location_product_reserve")
|
||||||
logger = logging.getLogger('stock.location')
|
_logger.warn("Failed attempt to reserve %s x product %s, likely due to another transaction already in progress. Next attempt is likely to work. Detailed error available at DEBUG level.", product_qty, product_id)
|
||||||
logger.warn("Failed attempt to reserve %s x product %s, likely due to another transaction already in progress. Next attempt is likely to work. Detailed error available at DEBUG level.", product_qty, product_id)
|
_logger.debug("Trace of the failed product reservation attempt: ", exc_info=True)
|
||||||
logger.debug("Trace of the failed product reservation attempt: ", exc_info=True)
|
|
||||||
return False
|
return False
|
||||||
|
|
||||||
# XXX TODO: rewrite this with one single query, possibly even the quantity conversion
|
# XXX TODO: rewrite this with one single query, possibly even the quantity conversion
|
||||||
|
|
|
@ -29,7 +29,7 @@ from tools.translate import _
|
||||||
import logging
|
import logging
|
||||||
import decimal_precision as dp
|
import decimal_precision as dp
|
||||||
|
|
||||||
_logger = logging.getLogger('mps')
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
def rounding(fl, round_value):
|
def rounding(fl, round_value):
|
||||||
|
|
|
@ -27,6 +27,7 @@ import pooler
|
||||||
import tools
|
import tools
|
||||||
from osv import fields, osv
|
from osv import fields, osv
|
||||||
from openerp import SUPERUSER_ID
|
from openerp import SUPERUSER_ID
|
||||||
|
_logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class CompanyLDAP(osv.osv):
|
class CompanyLDAP(osv.osv):
|
||||||
_name = 'res.company.ldap'
|
_name = 'res.company.ldap'
|
||||||
|
@ -107,8 +108,7 @@ class CompanyLDAP(osv.osv):
|
||||||
except ldap.INVALID_CREDENTIALS:
|
except ldap.INVALID_CREDENTIALS:
|
||||||
return False
|
return False
|
||||||
except ldap.LDAPError, e:
|
except ldap.LDAPError, e:
|
||||||
logger = logging.getLogger('orm.ldap')
|
_logger.error('An LDAP exception occurred: %s', e)
|
||||||
logger.error('An LDAP exception occurred: %s', e)
|
|
||||||
return entry
|
return entry
|
||||||
|
|
||||||
def query(self, conf, filter, retrieve_attributes=None):
|
def query(self, conf, filter, retrieve_attributes=None):
|
||||||
|
@ -135,7 +135,6 @@ class CompanyLDAP(osv.osv):
|
||||||
"""
|
"""
|
||||||
|
|
||||||
results = []
|
results = []
|
||||||
logger = logging.getLogger('orm.ldap')
|
|
||||||
try:
|
try:
|
||||||
conn = self.connect(conf)
|
conn = self.connect(conf)
|
||||||
conn.simple_bind_s(conf['ldap_binddn'] or '',
|
conn.simple_bind_s(conf['ldap_binddn'] or '',
|
||||||
|
@ -144,9 +143,9 @@ class CompanyLDAP(osv.osv):
|
||||||
filter, retrieve_attributes, timeout=60)
|
filter, retrieve_attributes, timeout=60)
|
||||||
conn.unbind()
|
conn.unbind()
|
||||||
except ldap.INVALID_CREDENTIALS:
|
except ldap.INVALID_CREDENTIALS:
|
||||||
logger.error('LDAP bind failed.')
|
_logger.error('LDAP bind failed.')
|
||||||
except ldap.LDAPError, e:
|
except ldap.LDAPError, e:
|
||||||
logger.error('An LDAP exception occurred: %s', e)
|
_logger.error('An LDAP exception occurred: %s', e)
|
||||||
return results
|
return results
|
||||||
|
|
||||||
def map_ldap_attributes(self, cr, uid, conf, login, ldap_entry):
|
def map_ldap_attributes(self, cr, uid, conf, login, ldap_entry):
|
||||||
|
@ -188,8 +187,7 @@ class CompanyLDAP(osv.osv):
|
||||||
if res[1]:
|
if res[1]:
|
||||||
user_id = res[0]
|
user_id = res[0]
|
||||||
elif conf['create_user']:
|
elif conf['create_user']:
|
||||||
logger = logging.getLogger('orm.ldap')
|
_logger.debug("Creating new OpenERP user \"%s\" from LDAP" % login)
|
||||||
logger.debug("Creating new OpenERP user \"%s\" from LDAP" % login)
|
|
||||||
user_obj = self.pool.get('res.users')
|
user_obj = self.pool.get('res.users')
|
||||||
values = self.map_ldap_attributes(cr, uid, conf, login, ldap_entry)
|
values = self.map_ldap_attributes(cr, uid, conf, login, ldap_entry)
|
||||||
if conf['user']:
|
if conf['user']:
|
||||||
|
|
Loading…
Reference in New Issue