# -*- coding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. from unittest.mock import patch from odoo.addons.base.models.ir_mail_server import IrMailServer from odoo.addons.survey.tests import common from odoo.tests import tagged from odoo.tests.common import HttpCase @tagged('-at_install', 'post_install', 'functional') class TestCertificationFlow(common.TestSurveyCommon, HttpCase): def test_flow_certification(self): # Step: survey user creates the certification # -------------------------------------------------- with self.with_user('survey_user'): certification = self.env['survey.survey'].create({ 'title': 'User Certification for SO lines', 'access_mode': 'public', 'users_login_required': True, 'questions_layout': 'page_per_question', 'users_can_go_back': True, 'scoring_type': 'scoring_with_answers', 'scoring_success_min': 85.0, 'certification': True, 'certification_mail_template_id': self.env.ref('survey.mail_template_certification').id, 'is_time_limited': True, 'time_limit': 10, }) q01 = self._add_question( None, 'When do you know it\'s the right time to use the SO line model?', 'simple_choice', sequence=1, constr_mandatory=True, constr_error_msg='Please select an answer', survey_id=certification.id, labels=[ {'value': 'Please stop'}, {'value': 'Only on the SO form'}, {'value': 'Only on the Survey form'}, {'value': 'Easy, all the time!!!', 'is_correct': True, 'answer_score': 2.0} ]) q02 = self._add_question( None, 'On average, how many lines of code do you need when you use SO line widgets?', 'simple_choice', sequence=2, constr_mandatory=True, constr_error_msg='Please select an answer', survey_id=certification.id, labels=[ {'value': '1'}, {'value': '5', 'is_correct': True, 'answer_score': 2.0}, {'value': '100'}, {'value': '1000'} ]) q03 = self._add_question( None, 'What do you think about SO line widgets (not rated)?', 'text_box', sequence=3, constr_mandatory=True, constr_error_msg='Please tell us what you think', survey_id=certification.id) q04 = self._add_question( None, 'On a scale of 1 to 10, how much do you like SO line widgets (not rated)?', 'simple_choice', sequence=4, constr_mandatory=True, constr_error_msg='Please tell us what you think', survey_id=certification.id, labels=[ {'value': '-1'}, {'value': '0'}, {'value': '100'} ]) q05 = self._add_question( None, 'Select all the correct "types" of SO lines', 'multiple_choice', sequence=5, constr_mandatory=False, survey_id=certification.id, labels=[ {'value': 'sale_order', 'is_correct': True, 'answer_score': 1.0}, {'value': 'survey_page', 'is_correct': True, 'answer_score': 1.0}, {'value': 'survey_question', 'is_correct': True, 'answer_score': 1.0}, {'value': 'a_future_and_yet_unknown_model', 'is_correct': True, 'answer_score': 1.0}, {'value': 'none', 'answer_score': -1.0} ]) # Step: employee takes the certification # -------------------------------------------------- self.authenticate('user_emp', 'user_emp') # Employee opens start page response = self._access_start(certification) self.assertResponse(response, 200, [certification.title, 'Time limit for this certification', '10 minutes']) # -> this should have generated a new user_input with a token user_inputs = self.env['survey.user_input'].search([('survey_id', '=', certification.id)]) self.assertEqual(len(user_inputs), 1) self.assertEqual(user_inputs.partner_id, self.user_emp.partner_id) answer_token = user_inputs.access_token # Employee begins survey with first page response = self._access_page(certification, answer_token) self.assertResponse(response, 200) csrf_token = self._find_csrf_token(response.text) r = self._access_begin(certification, answer_token) self.assertResponse(r, 200) with patch.object(IrMailServer, 'connect'): self._answer_question(q01, q01.suggested_answer_ids.ids[3], answer_token, csrf_token) self._answer_question(q02, q02.suggested_answer_ids.ids[1], answer_token, csrf_token) self._answer_question(q03, "I think they're great!", answer_token, csrf_token) self._answer_question(q04, q04.suggested_answer_ids.ids[0], answer_token, csrf_token, button_submit='previous') self._answer_question(q03, "Just kidding, I don't like it...", answer_token, csrf_token) self._answer_question(q04, q04.suggested_answer_ids.ids[0], answer_token, csrf_token) self._answer_question(q05, [q05.suggested_answer_ids.ids[0], q05.suggested_answer_ids.ids[1], q05.suggested_answer_ids.ids[3]], answer_token, csrf_token) user_inputs.invalidate_recordset() # Check that certification is successfully passed self.assertEqual(user_inputs.scoring_percentage, 87.5) self.assertTrue(user_inputs.scoring_success) # assert statistics statistics = user_inputs._prepare_statistics()[user_inputs] total_statistics = statistics['totals'] self.assertEqual( sorted( total_statistics, key=lambda item: item['text'] ), sorted( [ {'text': 'Correct', 'count': 2}, {'text': 'Partially', 'count': 1}, {'text': 'Incorrect', 'count': 0}, {'text': 'Unanswered', 'count': 0}, ], key=lambda item: item['text'] ) ) # Check that the certification is still successful even if scoring_success_min of certification is modified certification.write({'scoring_success_min': 90}) self.assertTrue(user_inputs.scoring_success) # Check answer correction is taken into account self.assertNotIn("I think they're great!", user_inputs.mapped('user_input_line_ids.value_text_box')) self.assertIn("Just kidding, I don't like it...", user_inputs.mapped('user_input_line_ids.value_text_box')) certification_email = self.env['mail.mail'].sudo().search([], limit=1, order="create_date desc") # Check certification email correctly sent and contains document self.assertIn("User Certification for SO lines", certification_email.subject) self.assertIn("employee@example.com", certification_email.email_to) self.assertEqual(len(certification_email.attachment_ids), 1) self.assertEqual(certification_email.attachment_ids[0].name, f'Certification - {certification.title}.html', 'Default certification report print_report_name is "Certification - %s" % (object.survey_id.display_name)') def test_randomized_certification(self): # Step: survey user creates the randomized certification # -------------------------------------------------- with self.with_user('survey_user'): certification = self.env['survey.survey'].create({ 'title': 'User randomized Certification', 'questions_layout': 'page_per_section', 'questions_selection': 'random', 'scoring_type': 'scoring_without_answers', }) page1 = self._add_question( None, 'Page 1', None, sequence=1, survey_id=certification.id, is_page=True, random_questions_count=1, ) q101 = self._add_question( None, 'What is the answer to the first question?', 'simple_choice', sequence=2, constr_mandatory=True, constr_error_msg='Please select an answer', survey_id=certification.id, labels=[ {'value': 'The correct answer', 'is_correct': True, 'answer_score': 1.0}, {'value': 'The wrong answer'}, ]) q102 = self._add_question( None, 'What is the answer to the second question?', 'simple_choice', sequence=3, constr_mandatory=True, constr_error_msg='Please select an answer', survey_id=certification.id, labels=[ {'value': 'The correct answer', 'is_correct': True, 'answer_score': 1.0}, {'value': 'The wrong answer'}, ]) # Step: employee takes the randomized certification # -------------------------------------------------- self.authenticate('user_emp', 'user_emp') # Employee opens start page response = self._access_start(certification) # -> this should have generated a new user_input with a token user_inputs = self.env['survey.user_input'].search([('survey_id', '=', certification.id)]) self.assertEqual(len(user_inputs), 1) self.assertEqual(user_inputs.partner_id, self.user_emp.partner_id) answer_token = user_inputs.access_token # Employee begins survey with first page response = self._access_page(certification, answer_token) self.assertResponse(response, 200) csrf_token = self._find_csrf_token(response.text) r = self._access_begin(certification, answer_token) self.assertResponse(r, 200) with patch.object(IrMailServer, 'connect'): question_ids = user_inputs.predefined_question_ids self.assertEqual(len(question_ids), 1, 'Only one question should have been selected by the randomization') # Whatever which question was selected, the correct answer is the first one self._answer_question(question_ids, question_ids.suggested_answer_ids.ids[0], answer_token, csrf_token) statistics = user_inputs._prepare_statistics()[user_inputs] total_statistics = statistics['totals'] self.assertEqual( sorted( total_statistics, key=lambda item: item['text'] ), sorted( [ {'text': 'Correct', 'count': 1}, {'text': 'Partially', 'count': 0}, {'text': 'Incorrect', 'count': 0}, {'text': 'Unanswered', 'count': 0}, ], key=lambda item: item['text'] ), "With the configured randomization, there should be exactly 1 correctly answered question and none skipped." ) section_statistics = statistics['by_section'] self.assertEqual(section_statistics, { 'Page 1': { 'question_count': 1, 'correct': 1, 'partial': 0, 'incorrect': 0, 'skipped': 0, } }, "With the configured randomization, there should be exactly 1 correctly answered question in the 'Page 1' section.")