website/models/website_page.py

304 lines
13 KiB
Python

# -*- coding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
from psycopg2 import sql
import re
from odoo.addons.http_routing.models.ir_http import slugify
from odoo.addons.website.tools import text_from_html
from odoo import api, fields, models
from odoo.osv import expression
from odoo.tools import escape_psql
from odoo.tools.translate import _
class Page(models.Model):
_name = 'website.page'
_inherits = {'ir.ui.view': 'view_id'}
_inherit = [
'website.published.multi.mixin',
'website.searchable.mixin',
]
_description = 'Page'
_order = 'website_id'
url = fields.Char('Page URL', required=True)
view_id = fields.Many2one('ir.ui.view', string='View', required=True, ondelete="cascade")
website_indexed = fields.Boolean('Is Indexed', default=True)
date_publish = fields.Datetime('Publishing Date')
menu_ids = fields.One2many('website.menu', 'page_id', 'Related Menus')
# This is needed to be able to control if page is a menu in page properties.
# TODO this should be reviewed entirely so that we use a transient model.
is_in_menu = fields.Boolean(compute='_compute_website_menu', inverse='_inverse_website_menu')
is_homepage = fields.Boolean(compute='_compute_is_homepage', inverse='_set_is_homepage', string='Homepage')
is_visible = fields.Boolean(compute='_compute_visible', string='Is Visible')
# Page options
header_overlay = fields.Boolean()
header_color = fields.Char()
header_text_color = fields.Char()
header_visible = fields.Boolean(default=True)
footer_visible = fields.Boolean(default=True)
# don't use mixin website_id but use website_id on ir.ui.view instead
website_id = fields.Many2one(related='view_id.website_id', store=True, readonly=False, ondelete='cascade')
arch = fields.Text(related='view_id.arch', readonly=False, depends_context=('website_id',))
def _compute_is_homepage(self):
website = self.env['website'].get_current_website()
for page in self:
page.is_homepage = page.url == (website.homepage_url or page.website_id == website and '/')
def _set_is_homepage(self):
website = self.env['website'].get_current_website()
for page in self:
if page.is_homepage:
if website.homepage_url != page.url:
website.homepage_url = page.url
else:
if website.homepage_url == page.url:
website.homepage_url = ''
def _compute_visible(self):
for page in self:
page.is_visible = page.website_published and (
not page.date_publish or page.date_publish < fields.Datetime.now()
)
@api.depends('menu_ids')
def _compute_website_menu(self):
for page in self:
page.is_in_menu = bool(page.menu_ids)
def _inverse_website_menu(self):
for page in self:
if page.is_in_menu:
if not page.menu_ids:
self.env['website.menu'].create({
'name': page.name,
'url': page.url,
'page_id': page.id,
'parent_id': page.website_id.menu_id.id,
'website_id': page.website_id.id,
})
elif page.menu_ids:
# If the page is no longer in menu, we should remove its website_menu
page.menu_ids.unlink()
# This update was added to make sure the mixin calculations are correct
# (page.website_url > page.url).
@api.depends('url')
def _compute_website_url(self):
for page in self:
page.website_url = page.url
def _get_most_specific_pages(self):
''' Returns the most specific pages in self. '''
ids = []
previous_page = None
# Iterate a single time on the whole list sorted on specific-website first.
for page in self.sorted(key=lambda p: (p.url, not p.website_id)):
if not previous_page or page.url != previous_page.url:
ids.append(page.id)
previous_page = page
return self.filtered(lambda page: page.id in ids)
@api.returns('self', lambda value: value.id)
def copy(self, default=None):
if default:
if not default.get('view_id'):
view = self.env['ir.ui.view'].browse(self.view_id.id)
new_view = view.copy({'website_id': default.get('website_id')})
default['view_id'] = new_view.id
default['url'] = default.get('url', self.env['website'].get_unique_path(self.url))
return super(Page, self).copy(default=default)
@api.model
def clone_page(self, page_id, page_name=None, clone_menu=True):
""" Clone a page, given its identifier
:param page_id : website.page identifier
"""
page = self.browse(int(page_id))
copy_param = dict(name=page_name or page.name, website_id=self.env['website'].get_current_website().id)
if page_name:
url = '/' + slugify(page_name, max_length=1024, path=True)
copy_param['url'] = self.env['website'].get_unique_path(url)
new_page = page.copy(copy_param)
# Should not clone menu if the page was cloned from one website to another
# Eg: Cloning a generic page (no website) will create a page with a website, we can't clone menu (not same container)
if clone_menu and new_page.website_id == page.website_id:
menu = self.env['website.menu'].search([('page_id', '=', page_id)], limit=1)
if menu:
# If the page being cloned has a menu, clone it too
menu.copy({'url': new_page.url, 'name': new_page.name, 'page_id': new_page.id})
return new_page.url
def unlink(self):
# When a website_page is deleted, the ORM does not delete its
# ir_ui_view. So we got to delete it ourself, but only if the
# ir_ui_view is not used by another website_page.
views_to_delete = self.view_id.filtered(
lambda v: v.page_ids <= self and not v.inherit_children_ids
)
# Rebind self to avoid unlink already deleted records from `ondelete="cascade"`
self = self - views_to_delete.page_ids
views_to_delete.unlink()
# Make sure website._get_menu_ids() will be recomputed
self.env.registry.clear_cache()
return super().unlink()
def write(self, vals):
for page in self:
website_id = False
if vals.get('website_id') or page.website_id:
website_id = vals.get('website_id') or page.website_id.id
# If URL has been edited, slug it
if 'url' in vals:
url = vals['url'] or ''
redirect_old_url = redirect_type = None
# TODO This should be done another way after the backend/frontend merge
if isinstance(url, dict):
redirect_old_url = url.get('redirect_old_url')
redirect_type = url.get('redirect_type')
url = url.get('url')
url = '/' + slugify(url, max_length=1024, path=True)
if page.url != url:
url = self.env['website'].with_context(website_id=website_id).get_unique_path(url)
page.menu_ids.write({'url': url})
if redirect_old_url:
self.env['website.rewrite'].create({
'name': vals.get('name') or page.name,
'redirect_type': redirect_type,
'url_from': page.url,
'url_to': url,
'website_id': website_id,
})
# Sync website's homepage URL
website = self.env['website'].get_current_website()
page_url_normalized = {'homepage_url': page.url}
website._handle_homepage_url(page_url_normalized)
if website.homepage_url == page_url_normalized['homepage_url']:
website.homepage_url = url
vals['url'] = url
# If name has changed, check for key uniqueness
if 'name' in vals and page.name != vals['name']:
vals['key'] = self.env['website'].with_context(website_id=website_id).get_unique_key(slugify(vals['name']))
if 'visibility' in vals:
if vals['visibility'] != 'restricted_group':
vals['groups_id'] = False
self.env.registry.clear_cache() # write on page == write on view that invalid cache
return super(Page, self).write(vals)
def get_website_meta(self):
self.ensure_one()
return self.view_id.get_website_meta()
@api.model
def _search_get_detail(self, website, order, options):
with_description = options['displayDescription']
# Read access on website.page requires sudo.
requires_sudo = True
domain = [website.website_domain()]
if not self.env.user.has_group('website.group_website_designer'):
# Rule must be reinforced because of sudo.
domain.append([('website_published', '=', True)])
search_fields = ['name', 'url']
fetch_fields = ['id', 'name', 'url']
mapping = {
'name': {'name': 'name', 'type': 'text', 'match': True},
'website_url': {'name': 'url', 'type': 'text', 'truncate': False},
}
if with_description:
search_fields.append('arch_db')
fetch_fields.append('arch')
mapping['description'] = {'name': 'arch', 'type': 'text', 'html': True, 'match': True}
return {
'model': 'website.page',
'base_domain': domain,
'requires_sudo': requires_sudo,
'search_fields': search_fields,
'fetch_fields': fetch_fields,
'mapping': mapping,
'icon': 'fa-file-o',
}
@api.model
def _search_fetch(self, search_detail, search, limit, order):
with_description = 'description' in search_detail['mapping']
# Cannot rely on the super's _search_fetch because the search must be
# performed among the most specific pages only.
fields = search_detail['search_fields']
base_domain = search_detail['base_domain']
domain = self._search_build_domain(base_domain, search, fields, search_detail.get('search_extra'))
most_specific_pages = self.env['website']._get_website_pages(
domain=expression.AND(base_domain), order=order
)
results = most_specific_pages.filtered_domain(domain) # already sudo
if with_description and search:
# Perform search in translations
# TODO Remove when domains will support xml_translate fields
query = sql.SQL("""
SELECT DISTINCT {table}.id
FROM {table}
LEFT JOIN ir_ui_view v ON {table}.view_id = v.id
WHERE (v.name ILIKE {search}
OR COALESCE(v.arch_db->>{lang}, v.arch_db->>'en_US') ILIKE {search})
AND {table}.id IN {ids}
LIMIT {limit}
""").format(
table=sql.Identifier(self._table),
search=sql.Placeholder('search'),
lang=sql.Literal(self.env.lang or 'en_US'),
ids=sql.Placeholder('ids'),
limit=sql.Placeholder('limit'),
)
self.env.cr.execute(query, {
'search': '%%%s%%' % escape_psql(search),
'ids': tuple(most_specific_pages.ids),
'limit': len(most_specific_pages.ids),
})
ids = {row[0] for row in self.env.cr.fetchall()}
if ids:
ids.update(results.ids)
domains = search_detail['base_domain'].copy()
domains.append([('id', 'in', list(ids))])
domain = expression.AND(domains)
model = self.sudo() if search_detail.get('requires_sudo') else self
results = model.search(
domain,
limit=len(ids),
order=search_detail.get('order', order)
)
def filter_page(search, page, all_pages):
# Search might have matched words in the xml tags and parameters therefore we make
# sure the terms actually appear inside the text.
text = '%s %s %s' % (page.name, page.url, text_from_html(page.arch))
pattern = '|'.join([re.escape(search_term) for search_term in search.split()])
return re.findall('(%s)' % pattern, text, flags=re.I) if pattern else False
if search and with_description:
results = results.filtered(lambda result: filter_page(search, result, results))
return results[:limit], len(results)
def action_page_debug_view(self):
return {
'type': 'ir.actions.act_window',
'res_model': 'ir.ui.view',
'res_id': self.view_id.id,
'view_mode': 'form',
'view_id': self.env.ref('website.view_view_form_extend').id,
}
# this is just a dummy function to be used as ormcache key
def _cached_response():
pass