Turkesh Patel (openERP) has proposed merging 
lp:~openerp-dev/openobject-addons/trunk-clean-yml-hr_evalution-tpa into 
lp:openobject-addons.

Requested reviews:
  Raphael Collet (OpenERP) (rco-openerp)
  Meera Trambadia (OpenERP) (mtr-openerp)
  Harry (OpenERP) (hmo-tinyerp)

For more details, see:
https://code.launchpad.net/~openerp-dev/openobject-addons/trunk-clean-yml-hr_evalution-tpa/+merge/83567

clean YMl files for hr_evaluation module.
-- 
https://code.launchpad.net/~openerp-dev/openobject-addons/trunk-clean-yml-hr_evalution-tpa/+merge/83567
Your team OpenERP R&D Team is subscribed to branch 
lp:~openerp-dev/openobject-addons/trunk-clean-yml-hr_evalution-tpa.
=== modified file 'hr_evaluation/__openerp__.py'
--- hr_evaluation/__openerp__.py	2011-12-06 11:21:27 +0000
+++ hr_evaluation/__openerp__.py	2011-12-08 10:25:33 +0000
@@ -50,7 +50,9 @@
         'hr_evaluation_data.xml',
         'hr_evaluation_installer.xml',
     ],
-    "test": ["test/test_hr_evaluation.yml"],
+    "test": ["test/test_hr_evaluation.yml",
+             "test/hr_evalution_demo.yml"
+             ],
     "active": False,
     "installable": True,
     "certificate" : "00883207679172998429",

=== modified file 'hr_evaluation/hr_evaluation_demo.xml'
--- hr_evaluation/hr_evaluation_demo.xml	2011-11-21 12:17:18 +0000
+++ hr_evaluation/hr_evaluation_demo.xml	2011-12-08 10:25:33 +0000
@@ -14,5 +14,11 @@
             <field name="evaluation_plan_id" ref="hr_evaluation_plan_managersevaluationplan0"/>
         </record>
 
+        <record id="hr_evaluation_evaluation_0" model="hr_evaluation.evaluation">
+            <field name="date">2011-12-24</field>
+            <field name="employee_id" ref="hr.employee1"/>
+            <field name="plan_id" ref="hr_evaluation.hr_evaluation_plan_managersevaluationplan0"/>
+        </record>
+
     </data>
 </openerp>
\ No newline at end of file

=== added file 'hr_evaluation/test/hr_evalution_demo.yml'
--- hr_evaluation/test/hr_evalution_demo.yml	1970-01-01 00:00:00 +0000
+++ hr_evaluation/test/hr_evalution_demo.yml	2011-12-08 10:25:33 +0000
@@ -0,0 +1,6 @@
+-
+  !record {model: hr.employee, id: hr.employee1, view: False}:
+    evaluation_plan_id: hr_evaluation_plan_managersevaluationplan0
+-
+  !record {model: hr_evaluation.evaluation, id: hr_evaluation_evaluation_0, view: False}:
+    plan_id: hr_evaluation.hr_evaluation_plan_managersevaluationplan0
\ No newline at end of file

=== modified file 'hr_evaluation/test/test_hr_evaluation.yml'
--- hr_evaluation/test/test_hr_evaluation.yml	2011-11-08 09:33:29 +0000
+++ hr_evaluation/test/test_hr_evaluation.yml	2011-12-08 10:25:33 +0000
@@ -1,144 +1,73 @@
 -
-  In order to test hr_evaluation module for OpenERP, I will create plan then create evaluation under that plan.
--
-
-  I create new Department.
--
-  !record {model: hr.department, id: hr_department_rd0}:
-    manager_id: base.user_root
-    name: 'R & D'
--
-  I create a new employee.
--
-  !record {model: hr.employee, id: hr_employee_employee0}:
-    address_home_id: base.res_partner_address_1
-    company_id: base.main_company
-    gender: male
-    marital: single
-    name: Mark Johnson
-    user_id: base.user_root
-    department_id: 'hr_department_rd0'
--
-  I create another new employee and assign first one as it's Manager.
--
-  !record {model: hr.employee, id: hr_employee_employee1}:
-    address_home_id: base.res_partner_address_3000
-    company_id: base.main_company
-    gender: male
-    name: Phil Graves
-    user_id: base.user_demo
-    parent_id: 'hr_employee_employee0'
-
--
-  I Create an "Employee Evaluation" survey for Manager's Evaluation Plan.
--
-    !record {model: 'survey', id: survey_0}:
-        title: 'Employee Evaluation'
-        max_response_limit: 20
-        response_user: 2
--
-  I Create an "Employee Evaluation" page in "Employee Evaluation" survey.
--
-    !record {model: 'survey.page', id: survey_employee_page_0}:
-        title: 'Employee Evaluation'
-        survey_id: survey_0
--
-  I Create "What is your Name" question in "Employee Evaluation" survey page.
--
-    !record {model: 'survey.question', id: survey_p_question_0}:
-        question: 'What is your Name?'
-        type: 'single_textbox'
-        sequence: 1
-        page_id: survey_employee_page_0
--
-  I Create "What is your gender" Question in "Employee Evaluation" survey page.
--
-    !record {model: 'survey.question', id: survey_p_question_1}:
-        question: 'What is your gender?'
-        type: multiple_choice_only_one_ans
-        sequence: 2
-        is_require_answer: true
-        page_id: survey_employee_page_0
--
-  I Create "Male" answer in question "What is your gender?"
--
-    !record {model: 'survey.answer', id: survey_p_1_1}:
-        answer: 'Male'
-        sequence: 1
-        question_id : survey_p_question_1
--
-  I Create "Female" answer in question "What is your gender?"
--
-    !record {model: 'survey.answer', id: survey_p_1_2}:
-        answer: 'Female'
-        sequence: 2
-        question_id : survey_p_question_1
-
--
-  I set the survey in open state.
+  I set the "Employee Evaluation" survey in open state.
 -
   !python {model: survey}: |
-       self.survey_open(cr, uid, [ref("survey_0")], context)
-
--
-  I create an Evaluation plan and select "Employee Evaluation" survey for "Send to Subordinates" and "Final interview with Manager" Phase.
--
-  !record {model: hr_evaluation.plan, id: hr_evaluation_plan_managersplan0}:
-    company_id: base.main_company
-    month_first: 3
-    month_next: 6
-    name: Manager's  Plan
-    phase_ids:
-      - action: bottom-up
-        name: Send to Subordinates
-        survey_id: 'survey_0'
-      - action: top-down
-        name: Final Interview with manager
-        sequence: 2
-        survey_id: 'survey_0'
--
-  I assign the evaluation plan to the employee "Mark Johnson".
--
-  !python {model: hr.employee}: |
-    res = self.onchange_evaluation_plan_id(cr, uid, [ref('hr_employee_employee0')], ref('hr_evaluation_plan_managersplan0'), False, None)
-    values = dict([('evaluation_plan_id', ref('hr_evaluation_plan_managersplan0'))] + res['value'].items())
-    self.write(cr, uid, [ref('hr_employee_employee0')], values, None)
--
-  I create an Evaluation for employee under "Manager Evaluation Plan".
--
-  !record {model: hr_evaluation.evaluation, id: hr_evaluation_evaluation_0}:
-    date: !eval time.strftime('%Y-%m-%d')
-    employee_id: 'hr_employee_employee1'
-    plan_id: 'hr_evaluation_plan_managersplan0'
-    progress: 0.0
-    state: draft
--
-  I change the employee on Evaluation.
--
-  !python {model: hr_evaluation.evaluation}: |
-    res = self.onchange_employee_id(cr, uid, [ref('hr_evaluation_evaluation_0')], ref('hr_employee_employee0'), None)
-    values = dict([('employee_id', ref('hr_employee_employee0'))] + res['value'].items())
-    self.write(cr, uid, [ref('hr_evaluation_evaluation_0')], values, None)
+    self.survey_open(cr, uid, [ref("survey_2")], context)
+-
+  I check that state of "Employee Evaluation" survey is Open.
+-
+  !assert {model: survey, id: survey_2, severity: error, string: Survey should be in OPen state}:
+    - state == 'open'
 -
   I start the evaluation process by click on "Start Evaluation" button.
 -
   !python {model: hr_evaluation.evaluation}: |
      self.button_plan_in_progress(cr, uid, [ref('hr_evaluation_evaluation_0')])
 -
+  I check that state is "Plan in progress".
+-
+  !assert {model: hr_evaluation.evaluation, id: hr_evaluation_evaluation_0, severity: error, string: Evaluation should be 'Plan in progress' state}:
+    - state == 'wait'
+-
   I find a mistake on evaluation form. So I cancel the evaluation and again start it.
 -
   !python {model: hr_evaluation.evaluation}: |
-     self.button_cancel(cr, uid, [ref('hr_evaluation_evaluation_0')])
-     self.button_draft(cr, uid, [ref('hr_evaluation_evaluation_0')])
-     self.button_plan_in_progress(cr, uid, [ref('hr_evaluation_evaluation_0')])
--
-  I close this survey request by giving answer of survey question.
--
-  !python {model: hr_evaluation.evaluation}: |
+    evaluation = self.browse(cr, uid, ref('hr_evaluation_evaluation_0') , context)
+    self.button_cancel(cr, uid, [ref('hr_evaluation_evaluation_0')])
+    assert evaluation.state == 'cancel', 'Evaluation should be in cancel state'
+    self.button_draft(cr, uid, [ref('hr_evaluation_evaluation_0')])
+    evaluation = self.browse(cr, uid, ref('hr_evaluation_evaluation_0') , context)
+    assert evaluation.state == 'draft', 'Evaluation should be in draft state'
+    self.button_plan_in_progress(cr, uid, [ref('hr_evaluation_evaluation_0')])
+-
+  I check that state is "Plan in progress" and "Interview Request" record is created
+-
+  !python {model: hr_evaluation.evaluation}: |
+    interview_obj = self.pool.get('hr.evaluation.interview')
+    evaluation = self.browse(cr, uid, ref('hr_evaluation_evaluation_0') , context)
+    assert evaluation.state == 'wait', "Evaluation should be 'Plan in progress' state"
+    interview_ids = interview_obj.search(cr, uid, [('evaluation_id','=', ref('hr_evaluation_evaluation_0'))])
+    assert len(interview_ids), "Interview evaluation survey not created"
+-
+  Give answer of the first page in "Employee Evaluation" survey.
+-
+  !python {model: survey.question.wiz}: |
+    name_wiz_obj=self.pool.get('survey.name.wiz')
+    interview_obj = self.pool.get('hr.evaluation.interview')
+    interview_ids = interview_obj.search(cr, uid, [('evaluation_id','=', ref('hr_evaluation_evaluation_0'))])
+    assert len(interview_ids), "Interview evaluation survey not created"
+    ctx = {'active_model':'hr.evaluation.interview', 'active_id': interview_ids[0], 'active_ids': [interview_ids], 'survey_id': ref("survey_2")}
+    name_id = name_wiz_obj.create(cr, uid, {'survey_id': ref("survey_2")})
+    ctx ["sur_name_id"] = name_id
+    self.create(cr, uid, {str(ref("survey_question_2")) +"_" +str(ref("survey_answer_1")) + "_multi" :'tpa',
+               str(ref("survey_question_2")) +"_" +str(ref("survey_answer_10")) + "_multi" :'application eng',
+               str(ref("survey_question_2")) +"_" +str(ref("survey_answer_20")) + "_multi" :'3',
+               str(ref("survey_question_2")) +"_" +str(ref("survey_answer_25")) + "_multi" :'2011-12-02 16:42:00',
+               str(ref("survey_question_2")) +"_" +str(ref("survey_answer_43")) + "_multi" :'HR',
+               str(ref("survey_question_2")) +"_" +str(ref("survey_answer_98")) + "_multi" :'tpa review'
+               }, context = ctx)
+-
+  I close this Evaluation survey by giving answer of questions.
+-
+  !python {model: hr_evaluation.evaluation}: |
+    interview_obj = self.pool.get('hr.evaluation.interview')
     evaluation = self.browse(cr, uid, ref('hr_evaluation_evaluation_0'))
-    self.pool.get('hr.evaluation.interview').survey_req_done(cr, uid, [r.id for r in evaluation.survey_request_ids])
+    interview_obj.survey_req_done(cr, uid, [r.id for r in evaluation.survey_request_ids])
+    for survey in evaluation.survey_request_ids:
+        interview = interview_obj.browse(cr, uid, survey.id, context)
+        assert interview.state == "done", 'survey must be in done state'
 -
-  I print the survey.
+  I print the evaluation.
 -
   !python {model: hr_evaluation.evaluation}: |
     evaluation = self.browse(cr, uid, ref('hr_evaluation_evaluation_0'))
@@ -147,22 +76,33 @@
   I click on "Final Validation" button to finalise evaluation.
 -
   !python {model: hr_evaluation.evaluation}: |
-    self.button_final_validation(cr, uid, [ref("hr_evaluation.hr_evaluation_evaluation_0")],
-      {"active_ids": [ref("hr_evaluation.menu_open_view_hr_evaluation_tree")]})
-
+    self.button_final_validation(cr, uid, [ref("hr_evaluation_evaluation_0")])
 -
-  I check that state is "Final Validation".
+  I check that state is "Waiting Appreciation".
 -
   !assert {model: hr_evaluation.evaluation, id: hr_evaluation_evaluation_0}:
       - state == 'progress'
 -
   Give Rating "Meet expectations" by selecting overall Rating.
 -
-  !record {model: hr_evaluation.evaluation, id: hr_evaluation.hr_evaluation_evaluation_0}:
+  !record {model: hr_evaluation.evaluation, id: hr_evaluation_evaluation_0}:
     rating: '2'
-
 -
   I close this Evaluation by click on "Done" button of this wizard.
 -
   !python {model: hr_evaluation.evaluation}: |
-    self.button_done(cr, uid, [ref("hr_evaluation.hr_evaluation_evaluation_0")], {"active_ids": [ref("hr_evaluation.menu_open_view_hr_evaluation_tree")]})
+    self.button_done(cr, uid, [ref("hr_evaluation_evaluation_0")])
+-
+  I check that state of Evaluation is done.
+-
+  !assert {model: hr_evaluation.evaluation, id: hr_evaluation_evaluation_0, severity: error, string: Evaluation should be in done state}:
+      - state == 'done'
+-
+  Print Evaluations Statistics Report
+-
+  !python {model: hr.evaluation.report}: |
+    import netsvc, tools, os, time
+    ctx={}
+    data_dict={'state': 'done', 'rating': 2, 'employee_id': ref("hr.employee1")}
+    from tools import test_reports
+    test_reports.try_report_action(cr, uid, 'hr_evaluation_evaluation_0',wiz_data=data_dict, context=ctx, our_module='hr_evaluation')
\ No newline at end of file

_______________________________________________
Mailing list: https://launchpad.net/~openerp-dev-gtk
Post to     : [email protected]
Unsubscribe : https://launchpad.net/~openerp-dev-gtk
More help   : https://help.launchpad.net/ListHelp

Reply via email to