Manual for PIP Project Management
|
|
|
- Virginia Douglas
- 9 years ago
- Views:
Transcription
1 Manual for PIP Project Management Section IV PIP Project Assessment and Evaluation ontents 1. Project Evaluation in General Evaluation in General Terms The 5 Evaluation riteria Project Assessment and Evaluation Absolute Assessment / Evaluation Definition of Absolute Assessment and Evaluation Absolute Assessment and Evaluation Forms Absolute Assessment and Evaluation Methods Evaluation of a ompleted Project omparative Assessment Definition of omparative Assessment (ompass) omparative Assessment Workshop omprehensive Results and Recommendations omprehensive Rating Results IV-1
2 Section IV Project Assessment and Evaluation Section IV explains the approach and methods of assessment and evaluation of PIP projects, conducted by DPI, MPI-DOE and/or the Planning Department in central government organizations. Although the contents are focused on requirements of the assessor/evaluator, the Project Owner also has to understand these procedures and contents to ensure that the projects proposed meet the assessment/evaluation requirements. hapter 1 explains about evaluation in general and the usage of the 5 evaluation criteria. 1. Project Evaluation in General This chapter explains the definition of evaluation as general terms, and the 5 evaluation criteria Evaluation in General Terms Evaluation is a periodic study conducted at certain stages of a project. Its ultimate objective is to analyze the situation of the project at a certain point, and to find out whether further improvements are necessary for the PIP project evaluated, and/or the project surroundings. When commencing an evaluation study, evaluation objectives are set up. Without evaluation objectives, the evaluation study lacks its direction, and evaluation results may not be utilized properly. A good evaluation needs to be credible and useful, at the same time independent and impartial. Evaluation results may be utilized for decision making, therefore the findings require to be proficient but still comprehensible. Evaluation must be done through thorough communication with stakeholders, but analyzed done without unnecessary bias. To produce accurate and fair evaluation results, evaluation must be done as systematically and objectively as possible The 5 Evaluation riteria The 5 Evaluation criteria are standard judgement requirements that are used in Absolute Assessment and Evaluation. It is essential to have a series of common evaluation criteria so that the assessment or evaluation stays consistent. (1) Definition of 5 Evaluation riteria The 5 Evaluation criteria is a comprehensive framework for assessment and evaluation that is commonly used by many international donors. It is an international evaluation standard introduced and recommended by Development Assistance ommittee (DA) in the IV-2
3 Organization for Economic ooperation and Development (OED) 1.y using the 5 evaluation criteria in evaluation for all PIP projects in all stages, there would be consistency in management of both the project itself and in comparison to other projects. The following are the 5 evaluation criteria, along with its basic definitions hart 1: Definition of the Five Evaluation riteria riteria Definition Relevance onsistency of the Project Purpose and Overall Goal, with development and policy targets. Appropriateness of the Project Purpose in line with the needs of beneficiaries, region and economy etc. Effectiveness Achievement of Project Purpose, or expectations of its achievement. Efficiency Appropriateness of project Inputs and its utilization, including; ost planned, used and disbursed Schedule of input and implementation (construction), Quality of Material and works, and, Actions for social and environmental approaches. Impact Any indirect effect caused by the project implementation. In project assessment, negative impacts on social and environmental aspects are stressed. Sustainability Sustainability of the project Outputs and its direct effect after completion. The project is evaluated from these five perspectives to verify whether it is/was necessary to implement the project, what effects the project has on the beneficiaries, whether the project is/was efficient in terms of effective use of resources, and how long the effects will be sustained. The following are the specific questions, or issues to be addressed for PIP projects by criteria. hart 2: Application of the 5 Evaluation riteria for PIP Projects riteria Relevance Issues to be addressed in PIP project assessment and evaluation Whether a project matches the priority of the NSEDP, Provincial and Sector SEDP, targeted beneficiaries, and other national and regional policies at the time of assessment/evaluation. Appropriateness of the Project Purpose (targeted beneficiary and region, etc) onsistency of the Project Purpose and Overall Goal, between the NSEDP, Provincial and 1 DA Principles for Evaluation of Development Assistance (1991), mentions that the 5 evaluation criteria is considered as an international standard when evaluating development projects. IV-3
4 Sustainability Impact Relevance Efficiency Effectiveness Manual for PIP Project Management (Version3.0, August 2010) Effectiveness Efficiency Impact (Negative) Sustainability Sector SEDP, and other policies. *Since development plans, needs and policies change in the course of time, it is important that the project is always evaluated on relevance based on the latest information. Whether or to what extent the Project Purpose is achieved. *In the case of newly proposed projects, Feasibility of Effectiveness will be evaluated. Whether project Inputs are utilized appropriately and efficiently. Whether the Inputs invested through the PIP budget efficiently develops to Outputs. *In case of newly proposed projects, Feasibility of Efficiency will be evaluated Main points are; Total cost, including financial schedule and actual disbursement Implementation plan and actual schedule of the project Quality of works and material, and Action taken for social and environmental issues Whether or how negative effect is caused through project implementation, and expectations of negative effect after completion. Social impacts such as resettlement and regional conflict Environmental impacts such as pollution Whether the outputs and the direct effect produced by the project can be sustained after the project is completed. Existence of Operation and Maintenance (O&M) Plans. (1) Responsible organization of O&M (2) O&M schedule (3) Material and equipment needed for O&M (4) O&M tasks and technical aspects (5) osts required for O&M and its source of budget (2) Project Framework and Evaluation riteria When evaluating a project that has a structured project framework, understanding the relations between the evaluation criteria and the narrative summary helps to analyze evaluation topics in an appropriate manner. Relations between the evaluation criterion and its levels in the Narrative Summary are as follows. Relevance checks the consistency and appropriateness of the project as a whole. Therefore, it relates to the Overall Goal, Project Purpose and Outputs. Effectiveness checks the achievement and its expectations to the Project Purpose. Since the achievement of the Project Purpose depends on the achievement of Outputs as components, Narrative Effectiveness relates to the Summary 5 Evaluation riteria Project Purpose and Outputs. Overall Goal Project Purpose Outputs Efficiency checks the appropriateness of project Inputs and its utilization. Activities considered as a Input, Activities IV-4 Figure 1: Narrative Summary and the 5 Evaluation riteria
5 process from Input to Output. Therefore, efficiency covers Outputs, Activities and Input. Impact checks any indirect negative effect caused by the project implementation. It relates to the Overall Goal and Project Purpose. Sustainability checks whether project results are sustained after completion. It relates to all levels in the Narrative Summary. 2. Project Assessment and Evaluation In PIP management, evaluation in general is broken up into 2 classifications; Assessment and Evaluation. Assessment and Evaluation of a PIP project have roles to examine the PIP project in its essential stages. The basic approach of assessment and evaluation is the same. However, the objectives vary depending on the category and stage of the project in subject. The following indicates the definition of assessment and evaluation respectively. Project Assessment is an inquiry to examine the PIP project before and during its progress, on whether it is necessary to start and/or continue the project implementation (construction). It also examines whether it is worthwhile to allocate the PIP budget, and if so, how much it should be allocated in the next fiscal year. There are two stages in project assessment; Absolute Assessment and omparative Assessment. Absolute Assessment is an intensive assessment focused to each individual project, while omparative Assessment compares projects within the same sector or region. While Absolute Assessment focuses on the performance of a PIP project, omparative Assessment focuses to its importance. Project Evaluation is an inquiry to examine the completion status and Operation / Maintenance (O&M) perspectives of the PIP project. It is done so that the project is, or would be effective and sustainable. Evaluation does not have direct links to PIP budgeting. It is normally focused to each individual project. 3. Absolute Assessment / Evaluation This chapter explains the concept and methods of Absolute Assessment and Evaluation of PIP projects. It first explains the concept and theory, followed by explanation of actual assessment/evaluation methods through certain formats Definition of Absolute Assessment and Evaluation Absolute Assessment and Evaluation are both commenced by focusing on one project. The differences between the two are in their objectives. While Absolute Assessment has an objective IV-5
6 to improve new and ongoing projects and allocate PIP budget, Evaluation has an objective to check the completed or operational status. Absolute Assessment and Evaluation are done in its respective PIP project management stage. Absolute Assessment is done by MPI/DPI with the cooperation from the PO. It is also ideal to commence Evaluation through MPI/DPI, although at times the PO can conduct Evaluation internally. The following chart shows assessment and evaluation done at each stage. Figure 2: Absolute Assessment and Evaluation done at each project stage PIP Project Stage Project Planning Project Monitoring Project ompletion Operations & Maintenance A S S E S S M E N T E V A L U A T I O N Implementation Assessment/Evaluation (1) Assessment for Feasibility Study and asic/detailed Design (2) Assessment for Project Implementation (onstruction / Technical Promotion) (3) Assessment for Ongoing Implementation (annual) (4) Evaluation at ompletion (5) Ex-Post Evaluation (2-3 years after operation) 3.2. Absolute Assessment and Evaluation Forms Simplified Project Assessment Sheet (SPAS) forms are used for Absolute Assessment, and Simplified Project Evaluation Sheet (SPES) forms are used for evaluation. There are 10 types of SPAS forms, and 5 types of SPES forms. The following chart indicates the different types of SPAS and SPES forms; hart 3: Types of SPAS Forms ategory Form Number Form Title haracteristics New Projects I-1 SPAS for NEW Technical Promotion For new Technical Promotion projects Assessment for PIP Format I-1 Project Proposal for Technical Promotion Projects The Technical Promotion Project in subject must be directly related to a certain PIP construction project. I-2 SPAS for NEW Feasibility Study and/or asic/detailed Design For projects (or project potentials) that require Feasibility Study, Designing and Social / Environmental Assessment with PIP budget before implementation Assessment for PIP Format I-2 Project Proposal for Feasibility Study and/or asic/detailed Design. Assessment is conducted for both the project idea and the F/S budget IV-6
7 Revival Projects New Projects suitability. I-3 SPAS for NEW onstruction For projects that have completed Feasibility Study, Designing and Social/Environmental Assessment and are requesting to start implementation of a PIP construction project Assessment for PIP Format I-3 Project Proposal for onstruction Projects.. Features economic/financial analysis requirements, I-4 SPAS for Feasibility Study and onstruction For small and short (1 year) projects that conduct F/S, design and construction in the same year. Assessment for PIP Format I-4 Project Proposal for Feasibility Study and onstruction. I-5 SPAS for Revival Projects For projects that have been suspended for more than 2 years after ongoing construction, and require PIP budget to restart construction. Assessment for PIP Format I-5 Project Proposal for Revival Projects Required assumption to revise or modify the original plan that was made before the suspension. I-6 SPAS for New Kum-ban Development Projects under NRP For projects that request PIP budget as Kum-ban Development PIP project through the NRP. Assessment for PIP Format I-6 Project Proposal for Kum-ban Development Projects under NRP Ongoing Projects II-1 II-2 II-3 II-6 SPAS for Ongoing Technical Promotion Projects For ongoing Technical Promotion projects that require PIP budget for further implementation. Assessment for PIP Format II-1 Progress Report for Technical Promotion Projects SPAS for Ongoing Feasibility Study and/or asic/detailed Design For ongoing Feasibility Study and/or Designing that require PIP budget for further studies/designing. Assessment for PIP Format II-2 Progress Report for Feasibility Study and/or asic/detailed Design SPAS for Ongoing onstruction Projects For ongoing construction projects that require PIP budget for further implementation. Assessment for PIP Format II-3 Progress Report for onstruction Projects SPAS for Ongoing Kum-ban Development Projects under NRP For ongoing Kum-ban Development PIP Projects through NRP. Assessment for PIP Format II-6 Progress Report for Kum-ban Development Projects IV-7
8 hart 4: Types of SPES Forms ategory At ompletion After ompletion Form Number III-1 III-2 III-3 III-6 IV Form haracteristics Type SPES for ompletion of Technical Promotion Project For completed Technical Promotion projects. Evaluation for PIP Format III-1 ompletion Report for Technical Promotion Projects SPES for ompletion of Feasibility Study and/or Design For projects that have completed its F/S and Designing stage, and ready for application of construction. Evaluation for PIP Format III-2 ompletion Report for Feasibility Study and/or Design SPES for ompletion of onstruction Project For completed construction projects. Evaluation for PIP Format III-3 ompletion Report for onstruction Projects SPES for ompletion for Kum-ban Development Project For completed Kum-ban Development Projects under NRP. Evaluation for PIP Format III-6 ompletion Report for Kum-an Development Projects under NRP Ex-post For selected PIP projects that have passed a certain amount of years SPES after its completion. Although the contents of the formats are different depending on the types and characteristics of the project, all SPAS and SPES have the same basic structure. They have; Assessment or evaluation questions based on the 5 evaluation criteria. A scoring system for each question, where scores are given depending on the status of the project. A scoring guideline is attached to guide the scores for every question. olumns are provided for each question, so that reasons for the score results can be described. A column for comprehensive comments, for providing recommendation for the project, is also prepared at the bottom of the format. A rating system, of which all projects assessed or evaluated are rated in a range of A to D based on the total of scores, are provided. Projects that are considered unprepared or immature is rated F, and rejected to the Project Owner for further consideration and resubmission of completed reports Absolute Assessment and Evaluation Methods oth SPAS and SPES forms are structured in a questionnaire-type format. There are 10 to 20 assessment questions depending on the type of format. Assessment or Evaluation is done by DPI, MPI-DOE or Planning Department in central government ministries, by analyzing the reports submitted by the Project Owner. Depending on its degree of achievement of the questioned situation, the project is given a certain score based on the scoring sheet attached to the form. IV-8
9 (1) SPAS Form Structure and its omponents The following shows the structure and components of a standard SPAS form. Figure 3: SPAS format and its ontents ˆÂ ¾ Á ³ º À ó  ¾ Á ê ñ»ñ ì ñ  ¾ ÀÎ óãï ÈŒN SPAS ( Ⱥ ¾ Ò É¾ ) Evaluation riteria ì ½¹ ñ  ¾ : ñ ê ó ½À ó : ˆ º ² ½ ñ ¾ ½À ó : ½À² 1) ¾ º Ⱥ Project Name and other basic information etc. «¾ À¹ ñ ½Á ½ ö º  ¾ Ä É«õ ö Ä ÉµÈ¾ ½Á É Áì ½À¹ ¾½ ö Ò? Evaluation ategory ¾ ö ¼ ñ ¾ øéä É»ñ ö ½Â¹ ¾ À¹ ¾½ ö º ¾ ñ Àì õº ê ª  ¾ û ¾ Ⱥ Ⱥ ê ¾ ɾ À «½ ò ñ ¾ª ö š º Ä É º À«ò ½ ö  ¾ µè¾ «õ ª ɺ Í õ Ò? À í¾¹ ¾ ì  ¾ º Ⱥ ñ ½ ö º  ¾ Ò? À í¾¹ ¾ ì  ¾ º Ⱥ ñ Á ² ñ ê ½ ¾À «½ ò ñ ö º ¾ Á ì ½ º Ä Éì ½ À«ò ñ ¾ øéä É»ñ ö ½Â¹ à ½ ö  ¾ Áì ½À í¾ï ¾ ì Í õ Ò? ½Î ¾ º  ¾ º Ⱥ ñ ¾ ª ɺ ¾ º ñ ¾ øéä É»ñ ö ½Â¹ Í õ Ò? Evaluation Questions omments ½ ö  ¾ º Ⱥ ñ ¾ ª ɺ ¾ º ñ ¾ øéä É»ñ ö ½Â¹ Í õ Ò? ½«¾ ê ª  ¾ Ä É«õ ñ Àì õº µè¾ À¹ ¾½ ö ê ¾ ɾ ê ½ ¾ Í õ Ò? ñ ¾ Ó ø ê À ñ À : È¾Ã É È¾, ö ª º Áê Áì ½ ¾ «¾ ¾ ì ö ê ô ª Ⱦ Å ó ¾ Î É¾À ô «õáì ½Ä É Ã É òê ó ¾ ê ÀÏ ¾½ ö ò Äì È Ò( òê óº ñ ª ¾ È Í Ð, ðì ½ ó ó/ Ò óâ ¾, òê ó ø È¾ ½ ñ ê ò)? º ñ ª ¾ È / ù È È¾1 (Í õnpv >0, Í õirr > r) Ò? Scores 2) ¾ À ñ Ä Ä É º ½ ò ê ò ö ½ ö º  ¾ Áì ½Ï ¾ ö 3) ¾ À ñ Ä Ä Éê ¾ ɾ ½ ò ê ò² ¾ Ï ¾ ö º  ¾ Ä É«ô ö Ä ÉµÈ¾ ² ¼ ² ðí õ ÒÀ² ô À»ñ ù É ñ ì ½ ö  ¾? ø Ⱦ ¾ ¾ ¾ ½À ø Ⱦ ¾ ó ¾ À¹ ¾½ ö Áì ½¹ ɾÀ ô «õä É Ò? Á ¾ ½ÀÎ óº ½ ñ ö ½ ¾ µøèã ö ½ ¾ ½ ñ Í õ Ò? ª ¾ª ½ì ¾ À ì ¾ ª ¾ª ½ì ¾ À ì ¾ º ¾ À ó ò ½ ö À¹ ö ö Í õ Ò? ñ ¾Á» ¾ / ñ ½ /º ½ º /Àª ñ  Âì óê ª ɺ ¾ ì ñ ¼ ¾ Á È ² ½² ¾ º ¼ ¾ ¼ ² ðí õ ÒÀ² ô À»ñ ù É ñ ì Ï ¾ ö º  ¾? 4) ö ½ê ö ö ½ê ɺ ɾ ì ö ê ¾ ɾ ñ ö Áì ½ ö ½ê ɺ ɾ ì ö ê ¾ ɾ ñ ö Áì ½ Á ì ɺ»É¾ Á» Í õ Ò? Á ì ɺ ¾ ª ½ ¾ Á ÉÄ ½À ñ ê ¾ ɾ ñ ö Áì Á ì ɺ 5) ¾ õ ö ¾ õ ö ê ¾ ɾ ¾ À ò ¾ õ ö ê ¾ ɾ Àª ñ ò ¾ õ ö ê ¾ ɾ ¾ ñ ª óá ¾ ª ½ ¾ Á ÉÄ ½À ñ ê ¾ ɾ ñ ö Áì ½ Á ì ɺ Í õ Ò? Ä É óá à ÉÁì ½ øì ½ ½»ñ ¾ê ì ½ À«ò ¾ õ ö ê ¾ ɾ ¾ À ò µè¾ ñ À Ò? Ä É óá à ÉÁì ½ øì ½ ½»ñ ¾ê ì ½ À«ò ¾ õ ö ê ¾ ɾ Àª ñ ò Áì ½ ñ «Ò ɾ µè¾ ñ À Ò? Ä É óá à ÉÁì ½ øì ½ ½»ñ ¾ ê ì ½ À«ò º ö ¾ Áì ½/Í õ È»ñ ò º µè¾ ñ À Ò? ½Á ì ½À² À¹ ñ Á ½ The assessor scores each evaluation question, which is related to some evaluation criterion and category. After providing the score, the assessor describes its reasons in the comment column. After covering all questions, scores are added up. The total score is then compared with the rating chart, and appropriate rating is found. After analyzing the total rating and assessment results of each question, the assessor then provides comprehensive recommendations to the Project Owner, for further improvement of the project contents. IV-9
10 (2) Scoring The scoring sheet shows the supposed situations of project for each evaluation question. The scores are higher when the situation of the project is ideal. The following shows the selection chart and scoring recommendation for a certain evaluation question. 1. ¾ º Ⱥ Evaluation riteria Evaluation ategory and Question ¾ ö ¼ 1: ½ ö º  ¾ Ä É«õ ö Ä ÉµÈ¾ ½Á É Áì ½À¹ ¾½ ö Ò? ½Á ÒÄ É ö ½ ö º  ¾ Ä É Probable situation of the project. 0 ½ ö º  ¾, À í¾ï ¾ Áì ½ ñ ¾ª ö š ñ Áª ÈÄì ½À ì ¾ º  ¾ Ò ñ À. Ä É ö ½ ö º  ¾, À í¾ï ¾ Áì ½ ñ ¾ª ö š ñ Áª ÈÄì ½À ì ¾ º  ¾ Áª È ñ À ñ ¾ ¾ ê ½ ñ ì à À ì ¾Â ¾. ½ ö º  ¾, À í¾ï ¾ Áì ½ ñ ¾ª ö š ñ Ä É«õ ö µè¾ ñ À Áì ½À¹ ¾½ ö Scoring Range Figure 4: Structure of SPAS Scoring Standard There is usually a range of scores for each probable situation of the project. ased on the actual situation, the assessor has to give an appropriate score within the range. If the actual situation is considered close to the more improved probable situation, the score is given higher within its described range. If the actual situation is considered close to the worse probable situation, the score is given lower within its described range. After considering the score, the score result is described in the SPAS form. Then, comments on the reasoning of the score are described. The scoring results and comments should therefore be related. ˆ º ² ½ ñ ¾ ½À ó : Scoring result as analyzed in the scoring sheet. 1) ¾ º Ⱥ ½À² «¾ À¹ ñ ½Á ½ ö º  ¾ Ä É«õ ö Ä ÉµÈ¾ ½Á É Áì ½À¹ ¾½ ö Ò? ¾ ö ¼ ñ ¾ª ö š º Ä É º Reasons À«ò ½ of ö  why ¾ µè¾ «õ the ª ɺ Íactual õ Ò? situation is as scored. À í¾¹ ¾ ì  ¾ º Ⱥ ñ ½ ö º  ¾ Ò? Figure 5: Scores and omments on SPAS À í¾¹ ¾ ì  ¾ º Ⱥ ñ Á ² ñ ê ½ ¾À «½ ò ñ ö º ¾ Á ì ½ º (3) Rating A rating system is introduced as a conclusion guideline in SPAS and SPES. The rating ranges from A to D, and an F included for projects that are immature, therefore rejected. IV-10
11 After all questions are scored and reasons commented, all scores are added up 2, and the total score is described in the Total Score column. The total score is then compared with the rating chart in the scoring sheet. The following shows the rating chart for SPAS before onstruction. Figure 6: SPAS Rating hart Á ì Áì ¾ ñ À² ê ɾ ( à À ì ¾ À ó N SPAS ê ɾ Ⱥ ¾ ñ ª ª Rating ò ñ ) Result. Í ¾ Ⱦ205 ½Á ì ½¹ Ⱦ 169 ¹ ¾204 ½Á º ñ ñ A ì ½¹ Ⱦ 133 ½Á ¹ ¾168 ½Á ª ú Ⱦ132 ½Á Total score levels. D ¾ «¾ à ¹ ɾ Àê ò ó Á Áª È0 ¹ ¾3 ( Á à ¾ «¾ ¾ À ñ Ä Ä Éê ¾ ɾ ò ê ò ö ª Ô È¾15 ) F The projects are rated closer to A when the scores are higher. In SPAS for New onstruction Projects (PIP format I-3), if the scores add up to over 205, the project is rated A. If the added scores are between 169 and 204, the project is rated, and so on. If the project is scored 0, or scored in the second lowest range (in case when 10 is the highest, it is scored 0-4), in ANY OF THE QUESTIONS, the rate becomes F nonetheless the total score reaches certain rates. It must be noted that the maximum total score of SPAS and SPES is different, therefore the scoring and rating relations varies depending on the form. The following chart shows the rates, required scoring percentage (full scores considered 100%), and the probable situation and analysis of the projects. hart 5: SPAS Rate--/Score Standard Rating Result Scoring percentage A Over 85% 71% to 85% 56% To 70% General Situation The project (or project potential) is in good condition. It can be implemented effectively and efficiently if PIP budget is approved. The project (or project potential) is in fair condition, although some improvements are recommended before implementation in some aspects. The project may face minor difficulties if implemented as it is reported. The project (or project potential) is not in good condition, and improvements are recommended before implementation in many aspects. The project may have major difficulties if implemented as it is. Analysis and Follow-up Try improving minor points. Ensure that the project implementation follows the current plan. Try improving the project with the priority on improvement based on the recommendations made in the assessment. When implemented, take special attention to points that are considered weak. Recommend to improve plans before implementation. If there are many difficult points, it is recommended to redesign the project from the basic concept. 2 Depending on the format used, there are some exceptions. I-2 SPAS before F/S and II-2 SPAS during F/S require different ways. See chapter xx (page xx) for details. IV-11
12 D Under 55% F - The project (or project potential) is in a very bad condition overall, and reconsideration of the project is highly recommended. The project is definitely ineffective and inefficient if implemented as it is. The project is facing a critical defect in at least one aspect of the project. The project is rejected to the Project Owner to reconsider the critical point. It must be re-assessed. If the defect cannot be amended, the project is rejected altogether. Strongly recommend to redesign the project altogether. Must redesign defected points, and receive re-assessment before proceeding. There is an ideal probability of all PIP project being rated A. However, in reality not all project can be rated A, moreover, the more assessment or evaluation becomes mature; the results tend to become strict. Ultimately, all projects fitted within the range of A and may be an ideal target for PIP project improvement. (4) Improvement of Projects in the Assessment Process One of the objectives of Absolute Assessment is to find out whether further improvements are necessary for the project. Therefore, in the process of assessment, discussions are made among MPI, DPI and PO to seek countermeasures to the project issues in relation to the SPAS results and recommendations. If improvement is possible through these countermeasures, the project is reassessed and provided improved results. If the rating improves, the improved rate is considered as the updated rate of the project. Attempt for improvement may be continued until the submission of SPAS results to the Decision Maker. (5) riteria Weight and Score-Rate Relations by Form As mentioned in hapter 3.2, there are various types of SPAS and SPES forms to match the assessment / evaluation objective of each project and its stage. The 5 evaluation is used in all forms, but depending on the project type and stage, the importance of each criterion is different, therefore the weight of importance affecting the total score has been adjusted. Generally, the following definitions are used as a guideline for criteria weight. hart 6: Guideline of riteria Weight by Project Stage Project Stage New projects (before implementation) Revival project (after suspension) Important Points (higher weight) Verification of relevance and necessity of project. onfirm feasibility of effectiveness. Existence of an Operation & Maintenance idea (especially the organization in charge) in the planning stages (sustainability). Relevance of the project based on the updated development goal and plan. Expectations of effectiveness and efficiency of the project ASED ON a REVISED PLAN. IV-12
13 Any social and/or environmental negative impact caused during suspension, Ongoing (during implementation) or expected upon revival. Efficiency (schedule, cost, quality of work) of the project. Effectiveness, or whether the Project Purpose would be achieved. Any social and/or environmental negative impact caused during implementation. ompletion Achievement of the Project Purpose (effectiveness). Existence of a detailed Operation & Maintenance Plan (sustainability). Any social and/or environmental negative impact caused during implementation, or may arise during operation stages. Operation Results of operation and progress of maintenance (sustainability). Achievement of the Overall Goal (relevance). Any social and/or environmental negative impact caused during operation. hart 7: Guideline of riteria Weight by Project Type Project Type Important Points Technical Promotion Sustainability of the technical progress. (Social and environmental negative impacts generally does not affect technical Feasibility Studies and Designing promotion projects, therefore no assessment questions.) Assessment of the project (expected construction) and the F/S itself is separately done. Therefore an independent criterion F/S & Designing is set up. oth the project and F/S need to reach the scoring requirements. Rating is based on the lower rate of achievement (i.e. if the project score reaches A and F/S reaches, the total rating is applied.) For Ongoing F/S & Design SPAS (II-2), only relevance is asked for the project itself. Other questions are related to the F/S & Design. onstruction Relevance of the project from the development plan, beneficiaries and regional viewpoints. ost efficiency, schedule and quality of works. Social and environmental negative impacts always monitored. Existence O&M organizations and plans. riteria for each SPAS forms are developed by combining the abovementioned 2 definitions. The following chart indicates the actual scores and weight by criterion for SPAS forms (I-1 to II-3). Note that for SPAS related to Feasibility Study, a criterion F/S & Designing is set up. IV-13
14 hart 8: Scores and Weight omparison for SPAS Forms (New Projects;I-1 to I-6) I-1(T/P) I-2(f/s&D) I-3 (onst.) I-4(for F/S & const.) I-5 (revival.) I-6 (Kum-ban) points ratio% points ratio% points ratio% points ratio% points ratio% points ratio% 1. Relevance % % % % % % 2. Effectiveness % % % % % % 3. Efficiency % % % % % % 4. Impact 0 0.0% 0 0.0% % % % % 5. Sustainability % 0 0.0% % % % % TOTAL % % % % % % 6. F/S & Designing The following indicates the actual score and weight by criterion for SPES forms. Note that for ex-post evaluation, effectiveness and efficiency of the finished project is not deeply analyzed, therefore not scored. hart 9: Scores and Weight omparison for SPAS Forma (Ongoing Projects: II--1 to II-5) II-1(ongoing TP) II-2(ongoing F/S) II-3 (ongoing constr.) II-6 (Kum-ban devt.) points ratio% points ratio% points ratio% points ratio% 1. Relevance % % % % 2. Effectiveness % 0 0.0% % % 3. Efficiency % 0 0.0% % % 4. Impact 0 0.0% 0 0.0% % % 5. Sustainability 0 0.0% 0 0.0% % % TOTAL % % % % 6. F/S & Designing Evaluation of a ompleted Project Evaluation for PIP projects are done at time of project completion, after the Project Owner has submitted the ompletion Report. This evaluation session is called Terminal Evaluation. Terminal Evaluation is obligatory for PIP projects for the following reasons; As mentioned earlier, Terminal Evaluation is conducted after the ompletion Report is submitted from the Project Owner, it must be conducted as soon as possible, ideally within the financial year that the project is physically conducted. For PIP projects that have not completed its payment within the physical completion and require budget for the following years to complete its payment, the Terminal Evaluation results must be submitted along with the ompletion Report as attachments every time the PIP budget is requested. IV-14
15 4. omparative Assessment This chapter explains the concept and methods of omparative Assessment (ompass). Explanation concentrates of the method of ompass Workshop, where PIP projects are discussed and compared, among participants related to the subject Definition of omparative Assessment (ompass) ompass is conducted by comparing PIP projects within a certain sector or region. Its objective is to recommend the best choice of PIP projects, or best allocation of PIP budget within projects that are applied in the same sector or region. Due to many limitations, not all projects that has been requested from PO s are approved and implemented. The typical reason is from the limitation of PIP budget. Through ompass, it would be capable of selecting the most important projects within these limitations. With the same reason as mentioned above, ODA projects are not exceptions. Although the ODA/PIP projects are basically given high priority among all the PIP projects, the PIP budget limitation does not allow all the ODA/PIP projects to be allocated with national contribution budget,. Therefore, both ODA/PIP and domestic PIP projects must be placed one table to be assessed with ompass. ompass can be done when there is a common subject in the projects that are being assessed. Subjects can either be a sector, sub-sector or region. If there are projects with more than one related subject, it is recommendable to conduct ompass in separate sessions. In case of provinces, it would be realistic if ompass Workshops are done by each sector department (or sub-department). When the issue is on rural development, it can be done within districts, kum-bans or villages. onclusion and outputs of ompass vary depending on the situation of the subject and availability of information. Assessment results can be simple recommendations of more important or less important projects, or may proceed to allocation of budget of each project. Levels of suggestion also vary; while in some cases suggestions limit to simple recommendations, others may extend to orders. It is therefore important that before commencing the omparative Assessment, MPI/DPI and the organization in charge of the subject agree on the expected conclusion and outputs, including its level of effect omparative Assessment Workshop ompass deals with information from various sources, therefore difficult to come to conclusion by individuals. Therefore, ompass is generally done through a workshop session with participation of stakeholders from the projects, sector and region.. Using a ompass hart as the base of discussion, workshop participants discuss and analyze the importance of PIP projects related to the development criteria in the sector or region. IV-15
16 The following subchapters explain the requirements, methods and tools for a ompass Workshop. (1) Pre-Selection of the projects Some PIP projects with certain conditions can get special priority to be applied by Pre-Selection procedure, without ompass. The process and criteria of the Pre-Selection are mentioned below; 1) In case of ODA/PIP a) Pre-Selection by financial priority; According to financial analysis based on Mid-Term PIP Expenditure Outlook, some particular ODA/PIP projects such as heavily indebted projects should be given urgent priority and be applied on the draft PIP list without ompass due to their financial urgency. b) Pre-Selection by political priority; According to official decision with democratic transparency in the political level, some particular ODA/PIP projects such as emergency relief projects should also be given special priority and be applied on the draft PIP list without ompass. c) Listing ODA/PIP for ompass; The remaining ODA/PIP projects that have not been given either financial priority or political priority through the Pre-Selection process in above a) and b) should be discussed in the ompass procedure. d) Giving result A to remaining ODA/PIP projects; Due to the assumption that all ODA/PIP projects are properly managed and monitored along with the development partner, as long as the project information and the national contribution budget is clear, all ODA/PIP projects should be given rate A as absolute assessment results before ompass. Figure 7: PIP udget Report Flow toward ompass 2) In case of domestic PIP a) Absolute Assessment by SPAS; It is essential for all the domestic PIP projects to be assessed by Absolute Assessment with SPAS. Methodology of Absolute Assessment with SPAS is IV-16
17 explained in hapter 3 Absolute assessment. b) Pre-Selection by financial priority; According to financial analysis based on Mid-Term PIP Expenditure Outlook, some particular domestic PIP projects such as heavily indebted projects should be given urgent priority and be applied on the draft PIP list without ompass due to their financial urgency. Pre-Selection by political priority; According to official decision with democratic transparency in the political level, some particular domestic PIP projects such as emergency relief project should be given special priority and be applied on the draft PIP list without ompass, too. c) Listing Domestic PIP projects for ompass; With the same manner as in the case of ODA/PIP projects, the domestic PIP that have not given either financial priority or political priority through the Pre-Selections in above b) and c) should be discussed in the ompass procedure. (2) General onditions and Procedure of the Workshop As mentioned above, the ompass Workshop is held with a certain subject or theme, which means that PIP projects that are supposedly related to the subject or theme would be compared through the workshop. It is ideal that no more than 10 projects are on the table for comparison. It would be difficult for the participants to follow the contents of each project, which is required when providing comparative scores. If there are more that 10 projects in the same subject, sector or region, it is recommended to further categorize the projects into sub-sectors or smaller regions. The following shows an ideal situation of a ompass Workshop; hart 10: ompass Workshop ase < Subject: ompass Workshop for Health Sector in xxx Province > < Assessment Objective and Number of Projects > PIP budget for provincial health department is very limited. udget has room for only 3 to 5 projects in full scale. There are 5 ongoing projects and 3 new projects to be compared. 4 of them are Technical Promotion Projects (2 ongoing and 2 new) 1 of the 5 projects is an ODA/PIP projects. Others are domestically funded. All domestically funded PIP projects have completed SPAS.. The ODA/PIP project has completed SPIS and has been automatically given SPAS result A. < Duration; 2 to 3 days > 1/2 day pre-meeting session (selection of projects and request for information) 1-2 day workshop session 1/2 day presentation to authority IV-17
18 < Participants; 10 to 14 participants > 3 DPI staff, including 1 staff acting as facilitator. PO for all 8 projects (some staff may be in charge of more than 1 project). 2 staff from Provincial Department of Health, Planning Division. 1 staff from MPI, and 1 staff from Ministry of Public Health as observers (if possible). Ideal procedure of the workshop is as follows hart 11: ompass Procedure Topic title Procedure 0. Orientation Explanation of the workshop objective, method used, expected outputs and schedule by the facilitator. 1. Presentation of Development Plans and Projects riefing of the health situation in the province by staff from Planning Division from Provincial Department in subject. riefing of PIP projects by PO to other participants. Discuss and confirm the PIP projects that should be compared. 2. Select ompass riteria Discuss and decide the important criteria for comparison within the sector. Provide coefficient rate to criteria depending on its importance. 3. ompare and score projects ompare the influence of each project to the criteria. Score depending on its influence. 4. oefficient score and total Multiply each score with the coefficient rate. Sum up the total for each project. 5. Rating Rate the project depending on the order of total. The facilitator has an important role of guiding the workshop. The facilitator needs to know both the ompass method and the subject of discussion. It is also required to smoothly direct the discussions to its conclusion. (3) ompass hart The ompass hart is the tool used in the workshop. It is a matrix to merge the projects to the criterion. y using this tool, discussions become focused and conclusions come out relatively smoother. hart 12: ompass hart x2,5 x1 A I- qo] txtmko II - qo] txtmko ^ f zjvo7; k, m5d9qo ^ 5f zjvodko 4k xjkgi af wi j (2) 5 (2) 2 (2) 5 (3) 3 IIIzt^ yf g, af r ao (2) 5 (1) 1 x1 gr uj, zt^ yf 8t zqog0qhjk (2) 2 (2) 2 (3) 3 x2 zts ^ yf gxao l yo7hk (1) 2 (2) 4 (2) 4 Rating results 11 A Score by (3) Projects to compare are seen by IV] hp l af (2) 2 (1) 1 (2) 4 14,5 A A Projects 7,5 Total score rows, and the assessment criteria (1) criterion A Vzt^ yf l t[ P vks ko (2) 2 (2) 2 (3) 6 12,5 A 2,5 are seen by columns. Total scores and rating is seen on the columns on the right side. When workshops are conducted, this chart can easily be written on large papers or on a whiteboard. IV-18
19 (4) Preparation Preparation is important for ompass. Accuracy of the workshop may change depending on the preparation. The following points must be prepared and cleared before conducting the workshop. hart 13: Preparation for ompass Workshop Preparation Topics Specific Points to Follow-up 1. Selection of subject ased on the PIP Project List and application by sector or region, MPI/DPI decides which subject the ompass session should be commenced. MPI/DPI coordinates with the organizations related to the subject (including PO) on the commencement of the workshop. It is important that the specific definition and borderline of the subject is clarified beforehand. 2. Selection of projects to be compared ased on the PIP Project List and application, MPI/DPI coordinates with related organizations on which PIP project should be compared, and which should be left out. A pre-meeting session is recommended. 3. Agreement on the workshop outputs onclusions and outputs of the workshop vary depending on the situation faced and information obtainable. Therefore, it is necessary to discuss on what conclusions and outputs can be expected through the workshop. 3. Selection of workshop participants ased on the subject and projects to be compared, workshop participants are decided and announced. It is important to provide the workshop objective along with the announcement. 4. Preparation of Develop- MPI/DPI requests preparation of SEDP for the sector / region to related ment Plans, Programs organizations. If PIP Program Workshops 3 are already held, its analysis results are very useful. These documents are compiled and distributed to all workshop participants in prior of the workshop, along with a request to read through beforehand. 5. Preparation of Project MPI/DPI requests the PO for information and documents of all projects that Information has been selected for comparison. These information are compiled and distributed to all workshop participants in prior of the workshop, along with a request to read through beforehand. Abovementioned topics 1 to 3 can be done through a pre-meeting session among MPI/DPI and organization in charge of the subject. Also, as stated in the chart, prior studies of development plans, PIP programs and project for comparison is needed by the participants. This ensures smooth discussions at the workshop, without spending time for repeated explanations. (5) Orientation The workshop is opened with an orientation session by the facilitator. Recommendable topics to 3 PIP Program Workshop : PAP recommends that a PIP Program Workshop is held before entering omparative Assessment workshop. A PIP Program Workshop is held jointly by PI/DPI and organizations related to the subject. See Manual for PIP Program Management for more details. IV-19
20 be covered at this session are as follows; hart 14: Orientation Topics in the ompass Workshop Orientation Topics Explanation of the workshop objective and expected outputs Presentation of the workshop method and schedule Introduction of facilitator and participants Specific Points to Follow-up It is important to clarify the workshop objective, including its expected outputs first. Although participants may already have a rough idea, the aim is to create a consensus of its goal. Explain the workshop method in brief. This is needed beforehand so that the participants can understand where they can reveal their topics. Specific procedures can be explained during the actual workshop. Explain the schedule of the workshop. Schedule is one of participants topic of interest. Moreover, it helps in the sense of time management. Have an introduction session. Introduction of the facilitator is very important. Good introduction may lead to good impression by the participants, which would make the facilitation easier. It may be good if an introduction sheet is made on large paper showing the participant s name, organization and projects in charge (for PO). (6) Presentation of Development Plans, PIP Program After orientation, a brief presentation of development plans and PIP program of the subject, preferably by Planning Division staff from the department or region in subject is done. The presentation would be closely related to selecting the ompass riteria as described in (6). (7) ODA/PIP Projects ODA/PIP projects and domestic PIP should be on the same table of the comparison. Each project should be evaluated by Absolute Assessment and compared by ompass for better achievement of the SEDP. If some listed ODA/PIP projects for the next year have any of special priority in financial or political aspects beyond any comparison, they should be applied after the Pre-Selection before and without ompass. However, it is often impossible by limitation of national contribution budget that all the listed ODA/PIP projects can be applied. Therefore, consequently, some or many ODA/PIP projects must remain on the table for ompass. In this case, the remaining ODA/PIP projects should have no more priority on the ompass, except that the Absolute Assessment results are automatically given A rate to the ODA/PIP projects. If the stakeholders are not satisfied with the point that some ODA/PIP projects do not get higher rate through ompass, the ODA/PIP projects should have been selected on the Pre-Selection by political and/or financial reasons. (8) ompass riteria While the 5 evaluation criteria are used as a standard in Absolute Assessment, the evaluation criteria of ompass are not fixed beforehand, rather decided depending on the background of the assessment. It is decided before or during the ompass Workshop. IV-20
21 Generally, SEDP priorities and target of the respective sector and/or group are selected as criteria. However, depending on the characteristic of the session, the criteria is discussed and decided within the participants. In such cases, the facilitator must clarify the definition of the criteria so that each participant has the same view and consensus when conducting the comparison of projects. Indicators may be used to clarify the criteria. Ideally, 4 to 5 criteria are reasonable from the workshop viewpoint. However, it depends on the priorities and target of the subject. Following are example of assessment criteria in sectors and region. hart 15: Examples of ompass riteria in Sector and Province Assessment riteria for Education in xxx Province onstruction of primary schools (target of completing 20 more schools by 2010). Provision of textbooks for all children (target of 100% distribution by 2010) Improvement of primary school entrance for ethnic children (target entrance ratio of 75% by 2010). Improvement of curriculum in secondary schools (curriculum and textbooks renewed to match new standards, to be completed by 2008) Improvement of teacher quality (teaching capability for all teachers reach national standard level by 2010) Assessment riteria for xxx District, yyy Province Electrification to all villages by 2010 (outreach to 5 villages). Improvement of access to 10 villages by 2010 (road completion 40km, repair 4 bridges). Improvement of agriculture and handicraft production and sales outside the district (cash-crop production, handicraft production training, market- ing, access etc.) onstruction of District Hospital (includes instalment of facilities, hiring nurses etc.) Education for ethnic groups in 5 villages (construction of 2 schools by 2010) Improvement of secondary schools. Within the selected criteria, some may be more important than others. Therefore, a coefficient weight within the assessment criteria is decided at the workshop. Participants discuss which criteria are comparatively important that others, and provide coefficient score to every criterion based on its importance. The least important criterion (or criteria) is scored 1.0, and further weight is provided based on its importance by 0.1. Obviously, the most important criterion (or criteria) is scored with the highest weight. Following is an example of coefficient scores for assessment criteria. hart 16: oefficient Scores and Assessment riteria Assessment riteria Score onstruction of primary schools (target of completing 20 more schools by 2010). 2.0 Provision of textbooks for all children (target of 100% distribution by 2010) 1.5 Improvement of primary school entrance for ethnic children (target entrance ratio of 75% by 2010). Improvement of curriculum in secondary schools (curriculum and textbooks renewed to match new standards, to be completed by 2008) Improvement of teacher quality (teaching capability for all teachers reach national standard level by 2010) IV-21
22 When conducting ompass, selection and weight of the assessment criteria becomes the crucial point that may decide the assessment results. Following are examples of assessment criteria and coefficient rate described in the ompass Sheet Figure 8: Assessment riteria and oefficient Rates in the ompass hart oefficient Rate SPAS Results Assessment riteria Project Name onstruction of Primary schools Provision of textbooks for all children Primary school entrance for ethnic children Improvement in curriculum in secondary schools Improvement of teacher quality Total Score omparative Rating omprehensive Rating I oefficient Rate Assessment riteria II (9) Presentation of Projects in omparison After selecting and weighing the assessment criteria, the PO provides brief presentations of each project that are to be compared. The presentation for one project should not be more than 10 minutes. At least the following information should be presented. hart 17: Information of the Project Provided by PO Project Name Location of Project Status of Project (new, ongoing or eneficiaries PIP budget request amount revival), and expected completion year. for the following year. Project Purpose, Overall Goal SPAS rating and results. During the presentation, the facilitator describes the project names and SPAS rating results on the ompass hart. In this process, ODA/PIP projects among the listed projects should be given the rate A automatically (refer to the previous section (1) ). Following is a described example; IV-22
23 hart 18: Project Name and SPAS Rating on omparative Assessment Sheet oefficient Rate SPAS Results Assessment riteria onstruction of Primary schools Provision of textbooks for all children Primary school entrance for ethnic children Improvement in curriculum in secondary schools Improvement of teacher quality Total Score omparative Rating omprehensive Rating Project Name < New Technical Promotion > Project for improving teaching skills, curriculum and textbooks for secondary schools. <New onstruction> onstruction of Ethnic Primary School in xx Village. Project Name A < Ongoing Technical Promotion> Project for improving technical skills, curriculum and textbooks for primary schools. SPAS Rating < Revival onstruction > onstruction of Ethnic Primary School in yyy Village. ODA/PIP project is < Ongoing onstruction > Expansion of lassroom uilding and Teacher Office in District apital school No.3 automatically given rate A (10) omparing and Scoring After the presentation, projects are compared using each assessment criterion. The basic question is to find out the projects that contribute most/least to the criterion. Figure 9: Scoring in omparative Assessment (1) SPAS Results Project Name oefficient Rate Assessment riteria onstruction of Primary schools Provision of textbooks for all children Each project is scored in a range of 1 to 3, with 3 being the project that contributes most to the criterion. The number of projects scored 3 2 and 1 is fixed depending on the number of projects that are compared. The following chart indicates the number of projects scored respectively; A < New Technical Promotion > Project for improving teaching skills, curriculum and textbooks for secondary schools. < New onstruction > onstruction of Ethnic Primary School in xx Village. < Ongoing Technical Promotion > Project for improving technical skills, curriculum and textbooks for primary schools < Revival onstruction > onstruction of Ethnic Primary School in yy Village hart 19: Scoring Standard ased on Total Number of Projects < Ongoing onstruction > Expansion of lassroom uilding and Teacher Office in District apital School No Total number of projects Projects scored Projects scored 3 points 2 points % 40% 30% Projects scored 1 point IV-23
24 The balance of score basically follows a 30% - 40% - 30% ratio, although it does not apply to total number of projects that are not divisible. The ratio is used so that both important and unimportant projects can be more highlighted, since it is relatively difficult to make judgements when all 3 scores have the same or similar ratio. All Assessment riteria are covered with the same method. Figure 10: Scoring in omparative Assessment (2) oefficient Rate SPAS Results Project Name Assessment riteria onstruction of Primary schools Provision of textbooks for all children Primary school entrance for ethnic children Improvement in curriculum in secondary schools Improvement of teacher quality A < New Technical Promotion > Project for improving teaching skills, curriculum and textbooks for secondary schools. < New onstruction > onstruction of Ethnic Primary School in xx Village. < Ongoing Technical Promotion > Project for improving technical skills, curriculum and textbooks for primary schools < Revival onstruction > onstruction of Ethnic Primary School in yy Village. < Ongoing onstruction > Expansion of lassroom uilding and Teacher Office in District apital School No Topics and opinions that relate directly to the reasoning of assessment results normally come out through the course of discussion. Therefore, the facilitator or staff supporting the facilitator must take notes. (11) oefficient Scores and Total After scoring is completed, coefficient rates for each assessment criteria are multiplied to the score in each cell. This can be done without discussion. Final scores after multiplied are added up as the total score. Figure 11: Scores with oefficient Rate and Total oefficient Rate SPAS Results A Assessment riteria Project Name < New Technical Promotion > Project for improving teaching skills, curriculum and textbooks for secondary schools. < New onstruction > onstruction of Ethnic Primary School in xx Village. < Ongoing Technical Promotion > Project for improving technical skills, curriculum and textbooks for primary schools < Revival onstruction > onstruction of Ethnic Primary School in yy Village. < Ongoing onstruction > Expansion of lassroom uilding and Teacher Office in District apital School No.3 onstruction of Primary schools Provision of textbooks for all children Primary school entrance for Improvement in curriculum in secondary Improvement of teacher Total Score ethnic quality multiplied schools children IV-24
25 (12) Rating Rating is done with a range from A to. Overall, A can be considered important while unimportant. However, ompass rating results and its reasons depend on the contents of discussion; therefore the rating definition varies depending on the workshop. The following indicates the probable reasons; hart 20: Rating results and their Probable Reasons Rate A Probable Reasons The projects deal with the most important areas in the sector or region. ompletion of the projects create positive effect to many important areas in the sector or region. The project(s) have comprehensive results. The projects deal with important areas in the sector or region, but are focused on a very limited area in the sector or region. ompletion of the projects create positive effect to certain areas in the sector or region, but the effectiveness for each area is comparatively lower than that of A rated projects. ompletion of the projects create positive effect to many areas in the sector or region, but does not effect the most important areas. The projects deal with the areas that are considered relatively unimportant for the sector or region. The projects results focus on a very limited area in the sector or region. The rating method is quite similar with the scoring method, whereas the number of projects scored A and is fixed depending on the number of projects that are compared. The following chart indicates the number of projects scored respectively; hart 21: Rating Standard ased on Total Number of Projects Total number of projects Projects rated A Projects rated Projects rated % 40% 30% IV-25
26 The balance of score basically follows a 30% - 40% - 30% ratio, although it does not apply to total number of projects that are not divisible. The reasons that ratio is used is similar to that of scoring. oth important (A) and unimportant() projects can be more highlighted, since it is relatively difficult to make judgements when all 3 rates have the same or similar ratio. The following indicates the rating results in the ompass hart ; Figure 12: Rating Results in the omparative Assessment hart oefficient Rate SPAS Results A Project Name Assessment riteria < New Technical Promotion > Project for improving teaching skills, curriculum and textbooks for secondary schools. < New onstruction > onstruction of Ethnic Primary School in xx Village. < Ongoing Technical Promotion > Project for improving technical skills, curriculum and textbooks for primary schools < Revival onstruction > onstruction of Ethnic Primary School in yy Village. < Ongoing onstruction > Expansion of lassroom uilding and Teacher Office in District apital School No.3 onstruction of Primary schools Provision of textbooks for all children Primary school entrance for ethnic children Improvement in curriculum in secondary schools Improvement of teacher quality Total Score omparative Rating A In cases where the score adds up to the same total amount, discussion is made specifically with the projects that have the same total. The chart is inspected, focusing on these projects. The conclusion of the discussion is respected for the final rating. 5. omprehensive Results and Recommendations This chapter explains the method of combining the Absolute and omparative assessment results (omprehensive Rating Results), and provide specific recommendations to the Decision Makers. 5.1 omprehensive Rating Results After both Absolute Assessment and ompass are conducted, each project has assessment results from different aspects. omprehensive Rating Results provide the status of each project by combining these 2 assessment results. It helps not only the staff involved in PIP management, but also the Decision Makers who makes concise decisions based on the recommendations by its staff. IV-26
27 The method is simple; just combine the 2 assessment results in order of ompass and Absolute Assessment. If the ompass Results is and Absolute Assessment Results is A, the project has a omprehensive Rate of A. Figure 13: omparative Assessment Results xtgr f dkoxtg, uo c[ [ l q, mp[ xtgr f dkoxtg, uoc[ [ 0kf 8q; (SPAS) + A = A As mentioned earlier, ompass explains the importance of a project, while Absolute Assessment explains its potential performance. y having the two rates combined, it not only explains the status of the project, but also provides recommendation of its further improvement. As explained in the previous sections, although the ODA/PIP projects have no priority than the domestic PIP in the ompass, they are automatically given an absolute assessment result A in the Pre-Selection before the workshop. Therefore, if an ODA/PIP project is given ompass Results through the workshop and given Absolute Assessment Results A automatically in the Pre-Selection, the project has a omprehensive Rate of A consequently. (1) Recommendation to Decision Makers With the omprehensive Rating and results of specific analyses, MPI/DPI as well as the PO would be able to provide information to Decision Maker. There are two ways of providing the assessment results and recommendation to the Decision Maker; from the viewpoint of the ompass Results, or as a sector/region; and from the viewpoint of each project. (2) Recommendations within the sector/region (ompass Viewpoint) This way of recommendation gives an idea of how the PIP should be managed in a comprehensive manner. Normally PIP budget expenditure has certain policy or direction by the Decision Maker, therefore the recommendations should be based on this policy direction. For example, if the Decision Maker announces that the PIP budget should be spent for projects with the priority of importance than its performance, the recommendations of projects have more emphasis on the ompass. If the Decision Maker has a policy or direction that the performance of PIP projects itself is prioritised, the recommendation of projects may change. If the budget ceiling for the sector/region in subject is already decided, allocation by project may also be considered. However, this is upon necessity, since it is a very sensitive issue. IV-27
28 The following reporting format is recommendable when reporting to the Decision Maker; hart 22: Recommendation by Sector / Region Format < asic Direction of Recommendation > Recommendation Order Project Name ompre- hensive Rate udget Allocation (Proposal) Reasons and Further Improvements Required The following is an example of a completed recommendation form; hart 23: Recommendation Form for Education Sector in xxx Province. < asic Direction of Recommendation > The Governor instructed the important projects are given priority, with a condition that they would be completed effectively and efficiently. DPI interpreted the above instruction that ompass results would be prioritized, but for projects that have Absolute Assessment results over rating. Absolute Assessment results under rating would be less prioritized nonetheless the importance. Vice Minister from the Ministry of Education has commented on the promotion of stabilizing the quality of secondary schools in general, when he came to the province. The Governor stresses the situation within ethnic groups, and commented that PIP should be allocated more to xx ethnic group in for the next coming years. DPI has already interpreted that before the ompass, and confirmed that the results have reflected the comments. udget Project Reasons and Further Order Rate Allocation Name Improvements Required (Proposal) 1. <Ongoing Technical Promotion> Project for improving technical skills, curriculum and textbooks for primary schools 2. < New Technical Promotion > Project for improving technical skills, curriculum and textbooks for secondary schools. 3. < Ongoing onstruction > Expansion of classroom building and teacher office in xxx District apital School No.3 AA 100Mil.K (request: 100Mil.K) 100Mil.K (request; 100Mil.K 300Mil.K (request; 500Mil.K oth importance and performance of the project is worth providing full amount of budget request. earing the importance of the project, as seen through the Vice Minister s comments, although the project needs more thoughts on sustainability, the budget request is fully approved. Although the priority of importance is relatively low, bearing to the high needs and fair performance of the project, it is recommended to continue with construction, although not the full IV-28
29 4. < New onstruction > onstruction of Ethnic Primary School in xx ethnic village 5 < Revival onstruction > onstruction of Ethnic Primary School in yy ethnic village 100Mil.K (request: 500Mil.K) 0Kip (request: 300Mil.K) amount requested. The plan of the project is immature; therefore although the importance is relatively high, due to the direction of the Governor, priority is low. Xx ethnic group is considered as priority of support, so with the condition of improving the project plan, the budget is allocated with a small amount. Due to the immature revival plan, although the importance is relatively high, the prority is low. Limitation of budget does not allow the allocation to this project. Need to improve the revival plan. It is known that the final decision of the project priorities or the budget allocation does not be matched with the above suggestion made by the analysis made by the staff level. However, it can help the Decision Maker to understand the actual situation of PIP. (3) omprehensive recommendations for each project In addition to analysis of the sector/region, it is also recommendable to compile analysis by project. This format can be turned out as a cover sheet when assessment results of each project are requested. It can also be shared between DPI/MPI and the PO, for further follow-up. The following reporting format is recommendable; hart 24: Recommendation by Project Name of Project : Status of Project (new/ongoing) : 1. omprehensive Rate : < Overview of the omprehensive Results > Project ode: 2. ompass Results : < Assessment Analysis > < Improvements necessary > 3. Absolute Assessment Results : < Assessment Analysis > < Improvements necessary or required > The following is an example of a completed recommendation form; IV-29
30 hart 25: Recommendation Form for Ethnic Primacy School in XX Village Name of Project : onstruction of an Ethnic Primary School in XX Village Project ode: xxxxxxxxxxxxxx Status of Project (new/ongoing) : New Project (after F/S, before construction) 1. omprehensive Rate : < Overview of the omprehensive Results > Although the importance is relatively high (rated, scored 2 nd out of 5), due to the direction of the Governor instructing efficient and effective projects are needed, priority is low because the project plan is immature. Education of ethnic children still stand as an important development policy in this province, and xx ethnic group is considered as priority of support, so with the condition of improving the project plan, the budget is allocated with a small amount. (100Mil.Kip approved out of 500Mi.Kip request) 2. ompass Results : < Assessment Analysis > The project is considered important, due to emphasis to education to ethnic schools. Since the construction < Improvements necessary > In the context of importance, the project has all basic components that is required. involves a classroom for secondary children, it improves the situation for secondary schoolchildren in the area. 3. Absolute Assessment Results : < Assessment Analysis > < Improvements necessary or required > In prior to construction plan, social analysis on the Reflection of social analysis results is needed. beneficiary ethnic group has been made, but the context of the beneficiaries has not reflected to the school design. School location does not match the beneficiaries onsideration of changing the location, or further discussions with the group is needed to satisfy their requirements. needs. The cost and schedule of school construction remains the same as it is made in the city area. earing to the fact that the school is made in a mountainous area where it may be isolated in heavy rain, reconsideration in Scheduling of the construction, with the consideration of the rainy season operation is needed. In such case, consideration of costs with no (or minimum) increase must be made. schedule and cost is necessary. IV-30
ÕÙ ØÝ ÌÖ Ò Ý ÁÒ Ø ØÙØ ÓÒ Ð ÁÒÚ ØÓÖ ÌÓ ÖÓ ÓÖ ÆÓØ ØÓ ÖÓ Ì Ó Ø ÆÓÖÛ Ò È ØÖÓÐ ÙÑ ÙÒ º Ê Ò Æ ÆÓÖ Ò ÖÒØ ÖÒ Ö ÆÓÖ Ò Ò ÆÓÖÛ Ò Ë ÓÓÐ Ó Å Ò Ñ ÒØ ½ Â ÒÙ ÖÝ ¾¼¼¼ ØÖ Ø Ì Ó Ø ØÓ Ò Ø ØÙØ ÓÒ Ð ÒÚ ØÓÖ Ó ØÖ Ò ÕÙ ØÝ Ö Ó
Ê ÔÓÒ Ú Ì ÒÛ Ö Î Ù Ð Þ Ø ÓÒ Ó Ä Ö Ó Ö Ô Ø Ø Ý Ã ÒÒ Ø Ò ÖØ Ø ÓÒ Ù Ñ ØØ Ò Ô ÖØ Ð ÙÐ ÐÐÑ ÒØ Ó Ø Ö ÕÙ Ö Ñ ÒØ ÓÖ Ø Ö Ó ÓØÓÖ Ó È ÐÓ ÓÔ Ý Ô ÖØÑ ÒØ Ó ÓÑÔÙØ Ö Ë Ò Æ Û ÓÖ ÍÒ Ú Ö ØÝ Ë ÔØ Ñ Ö ¾¼¼¾ ÔÔÖÓÚ Ô Ã ÒÒ Ø Ò
Universitat Autònoma de Barcelona
Universitat Autònoma de Barcelona ÙÐØ Ø Ò Ë Ó ³ Ò ÒÝ Ö ÁÒ ÓÖÑ Ø ÇÒ Ø Ò Ò ÓÒ ØÖÙØ ÓÒ Ó ÒØ¹Ñ Ø ÁÒ Ø ØÙØ ÓÒ Å Ñ ÓÖ ÔÖ ÒØ Ô Ö Ò ÂÙ Ò ÒØÓÒ Ó ÊÓ Ö Ù Þ Ù Ð Ö Ô Ö ÓÔØ Ö Ð Ö Ù ÓØÓÖ Ò ÒÝ Ö Ò ÁÒ ÓÖÑ Ø ÐÐ Ø ÖÖ Å ¾¼¼½
Ì ÍÆÁÎ ÊËÁÌ ÌÁË ÆÁ ˵ Ë Öº Ð º Ò Ö º ÚÓк ½ ÆÓº ½ ÔÖ Ð ¾¼¼¾ ½ ¹ ½ ÐÓ Ò Ò ÅÙÐØ ¹ ÀÞ ÒÚ ÖÓÒÑ ÒØ ÎÓ Ò º Ç ÐÓ Þ ÁÒÚ Ø È Ô Ö ØÖ Ø Ò ÓÚ ÖÚ Û Ó ÐÓ Ò Ò Ò Ó ÐÓ ØÓÖ Ð Ñ ÒØ ÔÖ ÒØ º ËÝ Ø Ñ Ø Ò Ó Ô¹ ÓÔ ÜÔÐ Ò Û ÐÐ Ø
Bud row 1. Chips row 2. Coors. Bud. row 3 Milk. Chips. Cheesies. Coors row 4 Cheesies. Diapers. Milk. Diapers
Ð ØÖ ØÝ ÜØ ÖÒ Ð Ë Ñ Ð Ö ØÝ Ó Ø ÓÖ Ð ØØÖ ÙØ Ö ØÓÔ Ö Êº È ÐÑ Ö ½ Ò Ö ØÓ ÐÓÙØ Ó ¾ ¾ ½ Î Ú ÑÓ ÁÒº ¾ ÛÓÓ ÐÚ È ØØ ÙÖ È Ô ÐÑ ÖÚ Ú ÑÓºÓÑ ÓÑÔÙØ Ö Ë Ò Ô ÖØÑ ÒØ ÖÒ Å ÐÐÓÒ ÍÒ Ú Ö ØÝ ¼¼¼ ÓÖ Ú È ØØ ÙÖ È Ö ØÓ ºÑÙº Ù
Ë ÓÒ Ð ØÝ Ò Ö ÙÐØÙÖ Ð ÓÑÑÓ ØÝ ÙØÙÖ Ö Ø Ò Ë Ö Ò Ò Ô ÖØÑ ÒØ Ó Ò Ò ÓÔ Ò Ò Ù Ò Ë ÓÓÐ ÊÓ Ò ÖÒ ÐÐ ½ ù½ ¼ Ö Ö Ö ÒÑ Ö Ì Ä ½ ½ ½ ¼¼ ¹Ñ Ð Óº º Ñ Ö ½ Ì ÙØ ÓÖ Ø Ò ÓÖ ÐÔ ÙÐ Ø Ò ÖÓÑ Â Ô Ö ĐÙÐÓÛ Ò ÓÑÑ ÒØ Ò Ù Ø ÓÒ ÖÓÑ
ÆÏ ÈÈÊÇÀ ÌÇ Ëµ ÁÆÎÆÌÇÊ ËËÌÅË ÂÒ¹ÉÒ ÀÙ ÅÒÙØÙÖÒ ÒÒÖÒ Ó ØÓÒ ÍÒÚÖ ØÝ ËÓÖ ÆÒÒÙÙÐ Ý Ò Ï¹Ó ÓÒ Ý ÐØÖÐ Ò ÓÑÔÙØÖ ÒÒÖÒ ÍÒÚÖ ØÝ Ó Å Ù ØØ ÑÖ Ø ÖÙÖÝ ØÖØ ÁÒ Ø ÔÔÖ Û ÓÒ Ö ÔÖÓ ÖÚÛ Ëµ ÒÚÒØÓÖÝ Ý ØÑ ÛØ ÒÔÒÒØ Ò ÒØÐÐÝ ØÖÙØ
ÓÒØÜØ¹ ÔÔÖÓ ÓÖ ÅÓÐ ÔÔÐØÓÒ ÚÐÓÔÑÒØ ÄÙØÓ ÆÙÖÓÓ ÁÖº ŵ ź˺ ÂÑ ÓÓµ Ì ÙÑØØ Ò ÙÐ ÐÐÑÒØ Ó Ø ÖÕÙÖÑÒØ ÓÖ Ø Ö Ó ÓØÓÖ Ó ÈÐÓ ÓÔÝ ËÓÓÐ Ó ÓÑÔÙØÖ ËÒ Ò ËÓØÛÖ ÒÒÖÒ ÅÓÒ ÍÒÚÖ ØÝ ÅÖ ¾¼¼½ ÐÖØÓÒ Ì Ø ÓÒØÒ ÒÓ ÑØÖÐ ØØ Ò ÔØ ÓÖ
ÓÑÔ Ö Ø Ú ËØÙ Ý Ó ÌÛÓ ØÖÓÒÓÑ Ð ËÓ ØÛ Ö È Ò Ì Ø Ù Ñ ØØ Ò Ô ÖØ Ð ÙÐ ÐÐÑ ÒØ Ó Ø Ö ÕÙ Ö Ñ ÒØ ÓÖ Ø Ö Ó Å Ø Ö Ó Ë Ò Ò ÓÑÔÙØ Ö Ë Ò Ì ÍÒ Ú Ö ØÝ Ó Ù Ð Ò ½ ÌÓ ÅÙÑ Ò Ò ØÖ Ø Ì Ø ÓÑÔ Ö Ø Ú ØÙ Ý Ó ØÛÓ ÓÔ Ò ÓÙÖ ØÖÓÒÓÑ
ÉÙ ÖÝ Ò Ë Ñ ØÖÙØÙÖ Ø ÇÒ Ë Ñ Å Ø Ò Á Ë Ë Ê Ì Ì Á Ç Æ ÞÙÖ ÖÐ Ò ÙÒ Ñ Ò Ö ÓØÓÖ Ö ÖÙÑ Ò ØÙÖ Ð ÙÑ Öº Ö Öº Ò Øºµ Ñ ÁÒ ÓÖÑ Ø Ò Ö Ø Ò Ö Å Ø Ñ Ø ¹Æ ØÙÖÛ Ò ØÐ Ò ÙÐØĐ Ø ÁÁ ÀÙÑ ÓРعÍÒ Ú Ö ØĐ Ø ÞÙ ÖÐ Ò ÚÓÒ À ÖÖ Ôк¹ÁÒ
ÓÒØÖÓÐ ËÝ Ø Ñ Ò Ò Ö Ò ÖÓÙÔ Ò ÙØÓÑ Ø ÓÒ Ì ÒÓÐÓ Ý ÖÓÙÔ Ö¹ÁÒ Ò Ö ÂÓ Ñ ÔÙØÝµ Ø ØÖ ½ ¼ ¼ À Ò È ÓÒ ¼¾ ½¹ ¹½½¼¼ Ü ¼¾ ½¹ ¹ ¹Å Ð Ò Ö Ó Ñ ÖÒÙÒ ¹ Ò Ñ Ø «È ÓÒ ÓÒØÖÓÐ ËÝ Ø Ñ Ò Ò Ö Ò ÖÓÙÔ ÔйÁÒ Ò Ö Ó«Ö¹ÁÒ ÍÐÖ ÓÖ ÓÐØ
ÔØ Ö ½ ÊÇÍÌÁÆ ÁÆ ÅÇ ÁÄ ÀÇ Æ ÌÏÇÊÃË Å Ãº Å Ö Ò Ò Ë Ñ Ö Êº Ô ÖØÑ ÒØ Ó ÓÑÔÙØ Ö Ë Ò ËØ Ø ÍÒ Ú Ö ØÝ Ó Æ Û ÓÖ Ø ËØÓÒÝ ÖÓÓ ËØÓÒÝ ÖÓÓ Æ ½½ ¹ ¼¼ ØÖ Ø Æ ÒØ ÝÒ Ñ ÖÓÙØ Ò ÓÒ Ó Ø Ý ÐÐ Ò Ò ÑÓ Ð Ó Ò ØÛÓÖ º ÁÒ Ø Ö ÒØ Ô
ÔØ Ö Ê Ö ÓÐÓ Ý ÁÒ Ø ÔØ Ö Ø Ö Ñ Ò ÛÓÖ Ø Ø ÓÒ Ú ÐÓÔ ÔÖ ÒØ º Ì ÛÓÖ Ø ¹ Ø ÓÒ ÓÑÔÙØ Ö Ø ÒÓ Ø ÑÓ ÙÐ Û Ö Ø ÓÖÓÒ ÖÝ ØÖ ÑÓ Ð ÐÐ ÔÐ Ý Ò ÑÔÓÖØ ÒØ ÖÓÐ Û Ò Ó Ò ÙØÓÑ Ø Ú Ð Ò ÐÝ Û Ø ÓÖÓÒ ÖÝ Ò Ó Ö ¹ Ô Ý Ñ º Ì ÔØ Ö Ò Û
ÍÒ Ö Ø Ò Ò Ø ÒØ ÖÔÖ ÁÒ ÓÖÑ Ø ÓÒ ËÝ Ø Ñ Ì ÒÓÐÓ Ý Ó Ò ÊÈ Ö Ï Ò Ö ØØÔ»»ÛÛÛº Ò º Ù¹ ÖÐ Òº» Û Ò Ö ÁÒ Ø ØÙØ ĐÙÖ ÁÒ ÓÖÑ Ø Ö ÍÒ Ú Ö ØĐ Ø ÖÐ Ò Ì Ù ØÖº ¹½ ½ ÖÐ Ò ÖÑ ÒÝ ¹Ñ Ð Û Ò º Ù¹ ÖÐ Òº ÔÖ Ð ¾ ¾¼¼¼ ÌÙØÓÖ Ð Ø Ø
Ê ½µ ¼»¼»¼½ ÓÑÔÙØÖ ËÒ»ÅØÑØ ½ Ô Ê Ö ÊÔÓÖØ Ì ÈÊËÍË ËÝ ØÑ ÖØØÙÖ ÖØ ÈØÞÑÒÒ ½ ÂÑ ÊÓÖÒ ¾ Ö ØÒ ËØÐ ½ ÅÐ ÏÒÖ ¾ ÖÒ ÏÖ ½ ÍÒÚÖ ØØ ËÖÐÒ ÁÑ ËØØÛÐ ¹½¾ ËÖÖÒ ÖÑÒÝ ßÔØÞÑÒÒ ØÙÐÐ ºÙÒ¹ º ¾ ÁÅ ÙÖ Ê Ö ÄÓÖØÓÖÝ ËÙÑÖ ØÖ À¹¼ Ê
Ø Ö ØÒ ÓÑÔ Ð Â Ú ÈÖÓ º ÓÒÒ Ø ÔÖÓÚ º Ø Þº µ ÔÖ Ð ¾ ¾¼¼½ ØÖ Ø ÓÖ ÕÙ Ø ÓÑ Ø Ñ ÒÓÛ Ñ Ö Ó Â Ú Î ÖØÙ Ð Å Ò ÂÎÅ µ ÓÒ Â٠عÁÒ¹Ì Ñ ÂÁ̵ Ò ¹Ç ¹Ì Ñ Ç̵ ÓÑÔ Ð Ö Û Ø Óҹع Ý ÓÔØ Ñ Þ Ø ÓÒ Ú Ò ÙÒØ Ò Ø Ö ÈÖÓ ÙØ ÖÙÒÒ Ò
FRAME. ... Data Slot S. Data Slot 1 Data Slot 2 C T S R T S. No. of Simultaneous Users. User 1 User 2 User 3. User U. No.
ÂÓÙÖÒ ÐÓ ÁÒØ ÖÓÒÒ Ø ÓÒÆ ØÛÓÖ ÎÓк¾ ÆÓº½ ¾¼¼½µ ¹ ÏÓÖÐ Ë ÒØ ÈÙ Ð Ò ÓÑÔ ÒÝ È Ê ÇÊÅ Æ Î ÄÍ ÌÁÇÆÇ Ê ÉÍ ËÌ¹Ì Å» Å ÈÊÇÌÇ ÇÄ ÇÊÏÁÊ Ä ËËÆ ÌÏÇÊÃË ÒØ Ö ÓÖÊ Ö ÒÏ Ö Ð ÅÓ Ð ØÝ Ò Æ ØÛÓÖ Ò Ê ÏŠƵ Ô ÖØÑ ÒØÓ ÓÑÔÙØ ÖË Ò
ÆÓØ Ä ØÙÖ Ð Ñ Ø ØÖÙ ÙØ ÓÒ ØÓ Á ¼ ØÙ ÒØ ÓÖ ÐÐ ÓØ Ö Ö Ø Ö ÖÚ Á ¼ ÈÊÇ Í ÌÁÇÆ ÈÄ ÆÆÁÆ Æ ÇÆÌÊÇÄ Ê Æ Ô ÖØÑ ÒØ Ó ÁÒ Ù ØÖ Ð Ò Ò Ö Ò ÍÒ Ú Ö ØÝ Ø Ù«ÐÓ ¹ ËØ Ø ÍÒ Ú Ö ØÝ Ó Æ Û ÓÖ Ò Ù«ÐÓº Ù Á ¼ ÈÊÇ Í ÌÁÇÆ ÈÄ ÆÆÁÆ Æ
Ò Ñ Ö Ð ÓÙÒ Ø ÓÒ ÓÖ ÙØÓÑ Ø Ï ÁÒØ Ö Ú ÐÙ Ø ÓÒ Ý Å ÐÓ Ý Ú ØØ ÁÚÓÖÝ ºËº ÈÙÖ Ù ÍÒ Ú Ö ØÝµ ½ ź˺ ÍÒ Ú Ö ØÝ Ó Ð ÓÖÒ Ø Ö Ð Ýµ ½ ÖØ Ø ÓÒ Ù Ñ ØØ Ò ÖØ Ð Ø Ø ÓÒ Ó Ø Ö ÕÙ Ö Ñ ÒØ ÓÖ Ø Ö Ó ÓØÓÖ Ó È ÐÓ Ó Ý Ò ÓÑÙØ Ö
Æ ÒØ Ò Ö Ø ÓÒ Ó ÊÓØ Ø Ò ÏÓÖ ÓÖ Ë ÙÐ ÆÝ Ö Ø ÅÙ Ð ÂÓ ÒÒ Đ ÖØÒ Ö Ò ÏÓÐ Ò ËÐ ÒÝ ØÖ غ Ò Ö Ø Ò ¹ÕÙ Ð ØÝ ÙÐ ÓÖ ÖÓØ Ø Ò ÛÓÖ ÓÖ Ö Ø Ð Ø Ò ÐÐ ØÙ Ø ÓÒ Û Ö ÖØ Ò Ø ÆÒ Ð Ú Ð ÑÙ Ø Ù Ö¹ ÒØ Ù Ò Ò Ù ØÖ Ð ÔÐ ÒØ Ó Ô Ø Ð
ËØØ ØÐ ÒÐÝ Ó ÒÓÑÔÐØ Ø ØÓÖÝ ØÒÕÙ Ò ÓØÛÖ º ź ÆÓÖÓÚ ÈÖ Ò ÙÔÔÐÑÒØ ØÓ Ø ÊÙ Ò ØÓÒ Ó ÄØØÐ ʺºº ÊÙÒ ºº ËØØ ØÐ ÒÐÝ ÏØ Å Ò Øº ÅÓ ÓÛ ÒÒ Ý ËØØ Ø ÔÔº ¹ ¾¹ ¾ ½½µ Ò ÊÙ Òµ ÈÖ ËØØ ØÐ ÒÐÝ ÛØ Ñ Ò Ø ÔÖÓÐÑ ÒÓÛÒ ØÓ ÐÑÓ Ø
(a) Original Images. (b) Stitched Image
ÁÅ Ê ÁËÌÊ ÌÁÇÆ ÁÆ ÌÀ Å ÌÄ ÆÎÁÊÇÆÅ ÆÌ ź ÅÙ ÖÓÚ º ÈÖÓ Þ ÁÒ Ø ØÙØ Ó Ñ Ð Ì ÒÓÐÓ Ý Ô ÖØÑ ÒØ Ó ÓÑÔÙØ Ò Ò ÓÒØÖÓÐ Ò Ò Ö Ò ØÖ Ø Ì Ô Ô Ö ÚÓØ ØÓ ÔÓ Ð Ø Ó ÓÑ ØÖ ÑÓ Ø ÓÒ Ó Ñ ØÓ Ò Ð ÓÒÒ Ø ÓÒ Ó Ô Ö Ø Ò Ó ÓÚ ÖÐ Ý Ò Ñ
ÅÁÌ ½ º ÌÓÔ Ò Ì Ë ÁÒØ ÖÒ Ø Ê Ö ÈÖÓ Ð Ñ ËÔÖ Ò ¾¼¼¾ Ä ØÙÖ ½ ÖÙ ÖÝ ¾¼¼¾ Ä ØÙÖ Ö ÌÓÑ Ä ØÓÒ ËÖ ÇÑ Ö Ø ÑÓÒ Ï Ð ½º½ ÁÒØÖÓ ÙØ ÓÒ Ì Ð Û ÐÐ Ù Ú Ö Ð Ö Ö ÔÖÓ Ð Ñ Ø Ø Ö Ö Ð Ø ØÓ Ø ÁÒØ ÖÒ Øº Ð ØÙÖ Û ÐÐ Ù ÀÓÛ Ô ÖØ ÙÐ
(a) Hidden Terminal Problem. (b) Direct Interference. (c) Self Interference
ØÖ ÙØ ÝÒ Ñ ÒÒ Ð Ë ÙÐ Ò ÓÖ ÀÓ Æ ØÛÓÖ ½ Ä ÙÒ Ó Ò ÂºÂº Ö ¹ÄÙÒ ¹ Ú Ë ÓÓÐ Ó Ò Ò Ö Ò ÍÒ Ú Ö ØÝ Ó Ð ÓÖÒ Ë ÒØ ÖÙÞ ¼ ¹Ñ Ð ÓÐ Ó ºÙ º Ù Î Ö ÓÒ ¼»½»¾¼¼¾ Ì Ö ØÝÔ Ó ÓÐÐ ÓÒ¹ Ö ÒÒ Ð ÔÖÓØÓÓÐ ÓÖ Ó Ò ØÛÓÖ Ö ÔÖ ÒØ º Ì ÔÖÓØÓÓÐ
application require ment? reliability read/write caching disk
Í Ò Ê ÑÓØ Å ÑÓÖÝ ØÓ ËØ Ð Ø Æ ÒØÐÝ ÓÒ Ò Ì¾ Ä ÒÙÜ Ð ËÝ Ø Ñ Ö Ò Ó Ö Ð ÖÓ Ï Ð Ö Ó ÖÒ Ö È Ó Ò Ì Ø Ò ËØ Ò ÍÒ Ú Ö Ö Ð È Ö ÓÓÖ Ò Ó È Ó ¹ Ö Ù Ó Ñ ÁÒ ÓÖÑ Ø Úº ÔÖ Ó Î ÐÓ Ó»Ò Ó ÓÓÒ Ó ½¼ ¹ ¼ ÑÔ Ò Ö Ò È Ö Þ Ð Ì Ð µ
ÀÖÖÐ ÈÐÑÒØ Ò ÆØÛÓÖ Ò ÈÖÓÐÑ ËÙÔØÓ Ù Ñ ÅÝÖ ÓÒ Ý ÃÑ ÅÙÒÐ Þ ÅÝ ½ ¾¼¼¼ ØÖØ ÁÒ Ø ÔÔÖ Û Ú Ø Ö Ø ÓÒ ØÒعÔÔÖÓÜÑØÓÒ ÓÖ ÒÙÑÖ Ó ÐÝÖ ÒØÛÓÖ Ò ÔÖÓÐÑ º Ï Ò Ý ÑÓÐÒ ÖÖÐ Ò ÛÖ Ö ÔÐ Ò ÐÝÖ Ò ÐÝÖ Ø Ü ÔÖÒØ Ó Ø ÑÒ ÓÙÒ Ñ ÖØ µº
ÔÔÖ Ò ÂÓÙÖÒÐ Ó ÓÑÔÙØÖ Ò ËÝ ØÑ ËÒ ÎÓк ½ ÆÓº ¾¼¼¼ ÔÔº ¾ß º ÈÖÐÑÒÖÝ ÚÖ ÓÒ Û Ò ÚÒ Ò ÖÝÔØÓÐÓÝ ß ÖÝÔØÓ ÈÖÓÒ ÄØÙÖ ÆÓØ Ò ÓÑÔÙØÖ ËÒ ÎÓк º ÑØ º ËÔÖÒÖ¹ÎÖÐ ½º Ì ËÙÖØÝ Ó Ø ÔÖ ÐÓ ÒÒ Å ÙØÒØØÓÒ Ó ÅÖ ÐÐÖ ÂÓ ÃÐÒ Ý ÈÐÐÔ
ÔØÖ ÄÒÖ Ç ÐÐØÓÖ ß ÇÒ Ö Ó ÖÓÑ º½ ÇÚÖÚÛ ÏØ Ó Ø ØÒ Ú Ò ÓÑÑÓÒ ÔÖÓÔØÓÒ Ó Ñ ÛÚ ÒÖØ Ý ÖØ¹ ÕÙ ÖÑÓØ ØØÓÒ Ó ÓÑÔÐÜ ÑÓÐÙÐ Ú ÒÖÖ ÔØÖ Ø ÐØÖ Ò ÑÒØ Ð Ò ÑÖÓÛÚ ÚØÝ Ò ÖÒØÖ ÐÓ Ì ÙÑÐ ÑÔÐ ÖÑÓÒ Ó Ð¹ ÐØÓÖ ÔÐÝ ØÖÖÒ ÖÓÐ Ò Ø ÙÒÖ
ÌÖ Ò Ø ÓÒ¹ Ö Ò Ò ÅÒ ÑÓ ÝÒ È Ö¹ØÓ¹È Ö ËØ ÒÓ Ö Ô ËØÓÖ ËÝ Ø Ñ Ì ÑÓØ Ý ÊÓ Ó ½ Ò ËØ Ú Ò À Ò ¾ ¾ ½ ËÔÖ ÒØ Ú Ò Ì ÒÓÐÓ Ý Ä ÓÖ ØÓÖÝ ÙÖÐ Ò Ñ ¼½¼ ÍË ÍÒ Ú Ö ØÝ Ó Ñ Ö ÓÑÔÙØ Ö Ä ÓÖ ØÓÖÝ Ñ Ö ¼ Íà ØÖ غ ÅÒ ÑÓ ÝÒ Ô Ö¹ØÓ¹Ô
HowPros and Cons of Owning a Home-Based Business
ÄØ Ø ÊÚ ÓÒ ÅÖ ¾½ ¾¼¼½ ÓÑÑÒØ ÏÐÓÑ ÖÑ Ò ÅÒÖÐ ÁÒÒØÚ ØÓ ÅÒÔÙÐØ Ø ÌÑÒ Ó ÈÖÓØ Ê ÓÐÙØÓÒ Ú ÀÖ ÐÖ ÌÖÙÒ ÓÖ ËÓÒÝÓÒ ÄÑ Ï ØÒ º ÕÙØ º ÖÓÚØ ˺ ÒÒ Åº ÖÒÒÒ Àº Ó º ÓÛÖÝ ÈºÙÐÖ Êº ÀÒРº ÀÖ ÐÖ º ÄÑÒÒ Åº ÅØÐÐ ÁºÈÒ ºÊ ÑÙ Ò
Ì ÈÒÒ ÝÐÚÒ ËØØ ÍÒÚÖ ØÝ Ì ÖÙØ ËÓÓÐ ÔÖØÑÒØ ÓËØØ Ø ËÌÊÌÁË ÇÊ Ì ÆÄËÁË ÏÁÌÀ ÌÏÇ ÌÈË Ç ÅÁËËÁÆ ÎÄÍË Ì Ò ËØØ Ø Ý ÇÖ ÀÖÐ ¾¼¼ ÇÖ ÀÖÐ ËÙÑØØ Ò ÈÖØÐ ÙÐ ÐÐÑÒØ Ó Ø ÊÕÙÖÑÒØ ÓÖ Ø Ö Ó ÓØÓÖ Ó ÈÐÓ ÓÔÝ ÙÙ Ø ¾¼¼ Ì Ø Ó ÇÖ ÀÖÐ
Client URL. List of object servers that contain object
ÄÓ Ø Ò ÓÔ Ó Ç Ø Í Ò Ø ÓÑ Ò Æ Ñ ËÝ Ø Ñ ÂÙ Ã Ò Ö Ù Ã Ø Ïº ÊÓ ÁÒ Ø ØÙØ ÙÖ ÓÑ ËÓÔ ÒØ ÔÓÐ Ö Ò Ò ÖÓ ÙÖ ÓѺ Ö Â Ñ Ïº ÊÓ ÖØ Ö Ò Ì Ð ÓÑ ß Æ Ì Á Ý Ð ÅÓÙÐ Ò ÙÜ Ö Ò ØÖ Ø ½ ÁÒØÖÓ ÙØ ÓÒ ÁÒ ÓÖ Ö ØÓ Ö Ù Ú Ö Ð Ý Ò Ò ¹
Ì ÈÖ Ò Ó ËØÖ ÔÔ ÅÓÖØ ¹ Ë ÙÖ Ø Â Ó ÓÙ ÓÙ Å ØØ Û Ê Ö ÓÒ Ê Ö ËØ ÒØÓÒ Ò ÊÓ ÖØ º Ï Ø Ð Û Â ÒÙ ÖÝ ½ ØÖ Ø ÁÒØ Ö Ø ÓÒÐÝ Áǵ Ò ÔÖ Ò Ô Ð ÓÒÐÝ Èǵ ØÖ ÔÔ ÑÓÖØ ¹ ÙÖ Ø Å Ëµ Ö Ö Ú Ø Ú ÙÖ Ø Û Ô Ý ÓÙØ ÓÒÐÝ Ø ÒØ Ö Ø ÓÑÔÓÒ
Ø Ú ÉÙ Ù Å Ò Ñ ÒØ ÓÒ Ø Ú Æ ØÛÓÖ ¹ ÍÒ Ø ÓÒ Ø ÓÒ ÓÒØÖÓÐ ÈÖÓØÓÓÐ Ê Ö ØÖ Ë Ö Ã Ö Ñ Ñ Ñ Æ ØÛÓÖ Ò Ê Ö ÖÓÙÔ Ë ÓÓÐ Ó ÓÑÔÙØ Ò ÍÒ Ú Ö ØÝ Ó Ä Ä Ä˾ ÂÌ ÍÒ Ø Ã Ò ÓÑ ßÖ Ö Ö ÑÐÓÑԺРº ºÙ ØØÔ»»ÛÛÛºÓÑԺРº ºÙ» ØÑ¹ÑÑ ØÖ
Author manuscript, published in "1st International IBM Cloud Academy Conference - ICA CON 2012 (2012)" hal-00684866, version 1-20 Apr 2012
Author manuscript, published in "1st International IBM Cloud Academy Conference - ICA CON 2012 (2012)" Á ÇÆ ¾¼½¾ ÌÓÛ Ö Ë Ð Ð Ø Å Ò Ñ ÒØ ÓÖ Å Ô¹Ê Ù ¹ Ø ¹ÁÒØ Ò Ú ÔÔÐ Ø ÓÒ ÓÒ ÐÓÙ Ò ÀÝ Ö ÁÒ Ö ØÖÙØÙÖ Ö Ð ÒØÓÒ
autocorrelation analysis
ÌÓÛÖ ËÔ¹ÒÖØ ÖÝÔØÓÖÔ ÃÝ ÓÒ Ê ÓÙÖ ÓÒ ØÖÒ Ú ÜØÒ ØÖص Ò ÅÓÒÖÓ ÅРú ÊØÖ Ý É Ä ÒРȺ ÄÓÔÖ Ø ÐÒ Ë ØÖØ ÈÖÓÖÑÑÐ ÑÓÐ ÔÓÒ Ò ÔÖ ÓÒÐ ØÐ ØÒØ È µ ÛØ ÑÖÓÔÓÒ ÔÖÑØ ÚÓ¹ ÖÚÒ Ù Ö ÒØÖ Ò Û Ù Ö ÔÖÓÚ Ò¹ ÔÙØ Ý ÔÒº ÁÒ Ø ÔÔÖ Û ÓÛÓÛØÓܹ
Ä ØÙÖ ËÐ ÁÒÚ ØÑ ÒØ Ò ÐÝ ½ ÌÖ Ò Ò ÁÒØÖÓ ØÓ ÌË Ó Ð ØÖ Ò Ø ÖÑ ÒÓÐÓ Ý ÜÔÐ Ò Ä Û Ó ÇÒ ÈÖ Ò Ö ØÖ ÐÙÐ Ø Ö ÔÐ Ø Ò ÔÓÖØ ÓÐ Ó Ó ÓÒ ÜÔÐ Ò Ö Ð Ø ÓÒ Ô Ö ØÖ Ò Ê ÔÐ Ø ÓÒ ËÔÓØ Ê Ø ÓÖÛ Ö Ê Ø Ä ØÙÖ ËÐ ÁÒÚ ØÑ ÒØ Ò ÐÝ ¾ ÇÖ
ÁÒØÖÔÖØØÓÒ Ó Î ÙÐÐÝ ËÒ ÍÖÒ ÒÚÖÓÒÑÒØ ÓÖ ËйÖÚÒ Ö ÖØØÓÒ ÞÙÖ ÖÐÒÙÒ Ö ÓØÓÖ¹ÁÒÒÙÖ Ò Ö ÙÐØØ ÐØÖÓØÒ Ö ÊÙÖ¹ÍÒÚÖ ØØ ÓÙÑ ÖÒ ÈØÞÓÐ ËØÙØØÖØ»ÓÙÑ ËÔØÑÖ ¾¼¼¼ ÊÖÒØÒ ÈÖÓº Öº¹ÁÒº ÏÖÒÖ ÚÓÒ ËÐÒ ÁÒ ØØÙØ Ö ÆÙÖÓÒÓÖÑØ ÄÖ ØÙÐ
ÓÑÔ Ö Ø Ú Ê Ú Û Ó ÊÓ ÓØ ÈÖÓ Ö ÑÑ Ò Ä Ò Ù ÁÞÞ Ø È Ñ Ö ÓÖÝ À Ö Ù Ù Ø ½ ¾¼¼½ ØÖ Ø ÁÒ Ø Ô Ô Ö Û Ñ ÓÑÔ Ö Ø Ú Ö Ú Û Ó Ú Ö ØÝ Ó ÒØ ÖÑ Ø ¹Ð Ú Ð ÖÓ ÓØ Ð Ò Ù Ø Ø Ú Ñ Ö Ò Ö ÒØ Ý Ö º Ï Ð Ó Ö ÖÓ ÓØ ÔÖÓ Ö ÑÑ Ò Ð Ò Ù
universe nonself self detection system false negatives false positives
Ö Ø ØÙÖ ÓÖ Ò ÖØ Ð ÁÑÑÙÒ ËÝ Ø Ñ ËØ Ú Ò º ÀÓ Ñ ÝÖ ½ Ò Ëº ÓÖÖ Ø ½ ¾ ½ Ô ÖØÑ ÒØ Ó ÓÑÔÙØ Ö Ë Ò ÍÆÅ Ð ÙÕÙ ÖÕÙ ÆÅ ½ ½ ¾ Ë ÒØ ÁÒ Ø ØÙØ ½ ÀÝ È Ö ÊÓ Ë ÒØ ÆÅ ¼½ ØÖ Ø Ò ÖØ Ð ÑÑÙÒ Ý Ø Ñ ÊÌÁ˵ Ö Û ÒÓÖÔÓÖ Ø Ñ ÒÝ ÔÖÓÔ
Downloaded from SPIE Digital Library on 29 Aug 2011 to 128.196.210.138. Terms of Use: http://spiedl.org/terms
ÔØ Ú ÓÒ ÖÝ Ñ ÖÖÓÖ ÓÖ Ø Ä Ö ÒÓÙÐ Ö Ì Ð ÓÔ º Ê Ö º Ö٠Ⱥ Ë Ð Ò Ö º ÐÐ Ò Êº ź Ò Ö ØØÓÒ ÀºÅº Å ÖØ Ò Ç ÖÚ ØÓÖ Ó ØÖÓ Ó Ö ØÖ Ä Ö Ó º ÖÑ ¼½¾ Ö ÒÞ ÁØ ÐÝ Ë ÁÒØ ÖÒ Ø ÓÒ Ð ºÖºÐº ÓÖ Ó ÈÖÓÑ ËÔÓ ¾» ¾¾¼ Ä Ó ÁØ ÐÝ Å ÖÓ
Applications. Decode/ Encode ... Meta- Data. Data. Shares. Multi-read/ Multi-write. Intermediary Software ... Storage Nodes
ËÐØÒ Ø ÊØ Ø ØÖÙØÓÒ ËÑ ÓÖ ËÙÖÚÚÐ ËØÓÖ ËÝ ØÑ ÂÝ Âº ÏÝÐ ÅÑØ ÐÓÐÙ ÎÝ ÈÒÙÖÒÒ ÅРϺ Ö ËÑ ÇÙÞ ÃÒ ÌÛ ÓÖÝ ÏÐÐÑ ÖÓÖÝ Êº ÒÖ ÈÖÔ Ãº ÃÓ Ð ÅÝ ¾¼¼½ Å͹˹¼½¹½¾¼ ËÓÓÐ Ó ÓÑÔÙØÖ ËÒ ÖÒ ÅÐÐÓÒ ÍÒÚÖ ØÝ ÈØØ ÙÖ È ½¾½ ØÖØ ËÙÖÚÚÐ
Primitives. Ad Hoc Network. (a) User Applications Distributed Primitives. Routing Protocol. Ad Hoc Network. (b)
Ï Ö Ð Æ ØÛÓÖ ¼ ¾¼¼½µ ß ½ ÅÙØÙ Ð ÜÐÙ ÓÒ Ð ÓÖ Ø Ñ ÓÖ ÀÓ ÅÓ Ð Æ ØÛÓÖ Â ÒÒ Ö º Ï ÐØ Ö Â ÒÒ Ö Äº Ï Ð Æ Ø Ò Àº Î Ý Ô ÖØÑ ÒØ Ó ÓÑÔÙØ Ö Ë Ò Ì Ü ²Å ÍÒ Ú Ö ØÝ ÓÐÐ ËØ Ø ÓÒ Ì ¹ ½½¾ ¹Ñ Ð ÒÒÝÛ ºØ ÑÙº Ù Û Ð ºØ ÑÙº Ù
ÆØÛÓÖ ÏÓÖÒ ÖÓÙÔ ÁÒØÖÒØ ÖØ ÜÔÖØÓÒ Ø ÙÙ Ø ¾¼¼¾ º ÓÖÐØØ ÉÇË ÁÒº ÁÖÚÒ ºÁº ÈÙÐÐÒ ÐÓÖÒ ÁÒ ØØÙØ Ó ÌÒÓÐÓÝ Ëº ËÖÓÓ ÆÓÖØÐ ÆØÛÓÖ ÍÃ ËØØ Ø Ó ÇÒ¹ÏÝ ÁÒØÖÒØ ÈØ ÐÝ ÖØ¹ÓÖÐØØ¹ËØØ Ø ¹Ó¹ÔعÐÝ ¹¼¼ºØÜØ ½ ËØØÙ Ó Ø ÅÑÓ Ì ÓÙÑÒØ
ÁÆÎÆÌÇÊ ÇÆÌÊÇÄ ÍÆÊÌÁÆ ÅÆµ ÑÒ ÙÒÖØÒ Ø ÊÒ ÎÖÒ Ó ÊÒ Ø Á ¼ ÈÊÇÍÌÁÇÆ ÈÄÆÆÁÆ Æ ÇÆÌÊÇÄ ÁÒÚÒØÓÖÝ ÓÒØÖÓÐ ÙÒÖØÒ ÑÒµ ÁÒØÖÓÙØÓÒ ÊÒÓÑ ÚÖØÓÒ ÑÔÓÖØÒØ ÔÖØÐ ÚÖØÓÒ ÈÖÓÐÑ ØÖÙØÙÖ ÑÔÐ ØÓ ÖÔÖ ÒØ ÖÒÓÑÒ Ò Ø ÑÓÐ Ê Æ ÍÒÚÖ ØÝ Ø
PROCESSOR IS OCCUPIED BY T i
ËÙÐÒ ÐÓÖØÑ ÓÖ ÅÙÐØÔÖÓÖÑÑÒ Ò ÀÖ¹ÊйÌÑ ÒÚÖÓÒÑÒØ º ĺ ÄÙ ÈÖÓØ Å Å Ù ØØ ÁÒ ØØÙØ Ó ÌÒÓÐÓÝ ÂÑ Ïº ÄÝÐÒ ÂØ ÈÖÓÔÙÐ ÓÒ ÄÓÖØÓÖÝ ÐÓÖÒ ÁÒ ØØÙØ Ó ÌÒÓÐÓÝ ØÖØ Ì ÔÖÓÐÑ Ó ÑÙÐØÔÖÓÖÑ ÙÐÒ ÓÒ ÒÐ ÔÖÓ ÓÖ ØÙ ÖÓÑ Ø ÚÛÔÓÒØ Ó Ø ÖØÖ
TheHow and Why of Having a Successful Home Office System
ÊÇÄ ¹ Ë ËË ÇÆÌÊÇÄ ÇÆ ÌÀ Ï ÍËÁÆ Ä È ÂÓÓÒ Ëº È Ö ÁÒ ÓÖÑ Ø ÓÒ Ò ËÓ ØÛ Ö Ò Ò Ö Ò Ô ÖØÑ ÒØ ÓÖ Å ÓÒ ÍÒ Ú Ö ØÝ Ô Ö Ø ºÒÖÐºÒ ÚÝºÑ Ð Ð¹ÂÓÓÒ Ò ÓÐÐ Ó ÁÒ ÓÖÑ Ø ÓÒ Ì ÒÓÐÓ Ý ÍÒ Ú Ö ØÝ Ó ÆÓÖØ ÖÓÐ Ò Ø ÖÐÓØØ ÒÙÒº Ù Ê Ú
Sliding Window ... Basic Window S[0] S[k 1] S[k] Digests Digests Digests
ËØØËØÖÑ ËØØ ØÐ ÅÓØÓÖ Ó ÌÓÙ Ó Ø ËØÖÑ ÊÐ ÌÑ ÙÝÙ Ù Ë ÓÙÖØ Á ØØÙØ Ó ÅØÑØÐ Ë ÔÖØÑØ Ó ÓÑÔÙØÖ Ë ÆÛ ÓÖ ÍÚÖ ØÝ ÝÙÝÙ ºÝÙºÙ ØÖØ Ó Ö Ø ÔÖÓÐÑ Ó ÑÓØÓÖ Ø Ó ØÓÙ Ó ØÑ Ö Ø ØÖÑ ÓÐ Ó Ñ Ó Ó ØÑº Á ØÓ ØÓ Ð ØÖÑ ØØ ¹ Ø Ù ÚÖ ØÖ
Application. handle layer. access layer. reference layer. transport layer. ServerImplementation. Stub. Skeleton. ClientReference.
ÜÔÐÓ Ø Ò Ç Ø ÄÓ Ð ØÝ Ò Â Ú È ÖØÝ ØÖ ÙØ ÓÑÔÙØ Ò ÒÚ ÖÓÒÑ ÒØ ÓÖ ÏÓÖ Ø Ø ÓÒ ÐÙ Ø Ö ÖÒ Ö À ÙÑ Ö Ò Å Ð È Ð ÔÔ Ò ÍÒ Ú Ö ØÝ Ó Ã ÖÐ ÖÙ ÖÑ ÒÝ ÙÑ Ö ºÙ º Ò Ô Ð ÔÔ Ö ºÙ º ØØÔ»»ÛÛÛ Ô º Ö ºÙ º»Â Ú È ÖØÝ» ØÖ غ ÁÒ ØÖ
In Proceedings of the 1999 USENIX Symposium on Internet Technologies and Systems (USITS 99) Boulder, Colorado, October 1999
In Proceedings of the 999 USENIX Symposium on Internet Technologies and Systems (USITS 99) Boulder, Colorado, October 999 ÓÒÒ Ø ÓÒ Ë ÙÐ Ò Ò Ï Ë ÖÚ Ö Å Ö º ÖÓÚ ÐÐ ÊÓ ÖØ Ö Ò Ó Ó Ô ÖØÑ ÒØ Ó ÓÑÔÙØ Ö Ë Ò Ó
} diff. } make. fetch. diff. (a) Standard LRC. (c) Home-based LRC. (b) AURC. Node 0 Node 1 Node 2 (home) Node 0 Node 1 Node 2 (home) Compute
ÈÙÐ Ò Ø ÈÖÓÒ Ó Ø ¾Ò ËÝÑÔÓ ÙÑ Ó ÇÔÖØÒ ËÝ ØÑ Ò Ò ÁÑÔÐÑÒØØÓÒ ÇËÁ³µ ÈÖÓÖÑÒ ÚÐÙØÓÒ Ó ÌÛÓ ÀÓѹ ÄÞÝ ÊÐ ÓÒ ØÒÝ ÈÖÓØÓÓÐ ÓÖ ËÖ ÎÖØÙÐ ÅÑÓÖÝ ËÝ ØÑ ÙÒÝÙÒ ÓÙ ÄÚÙ ÁØÓ Ò Ã Ä ÔÖØÑÒØ Ó ÓÑÔÙØÖ ËÒ ÈÖÒØÓÒ ÍÒÚÖ ØÝ ÈÖÒØÓÒ ÆÂ
hospital physician(2)... disease(4) treat(2) W305(2) leukemia(3) leukemia(2) cancer
Ë ÙÖ ÅÄ ÈÙ Ð Ò Û Ø ÓÙØ ÁÒ ÓÖÑ Ø ÓÒ Ä Ò Ø ÈÖ Ò Ó Ø ÁÒ Ö Ò Ó ÙÒ Ò Ô ÖØÑ ÒØ Ó ÓÑÔÙØ Ö Ë Ò ÆÓÖØ Ø ÖÒ ÍÒ Ú Ö ØÝ Ä ÓÒ Ò ½½¼¼¼ Ò Ý Ò ÜÑ ÐºÒ Ùº ÙºÒ Ò Ä Ý Ë ÓÓÐ Ó ÁÒ ÓÖÑ Ø ÓÒ Ò ÓÑÔÙØ Ö Ë Ò ÍÒ Ú Ö ØÝ Ó Ð ÓÖÒ ÁÖÚ
ÌÀ ÀÁ¹ÇÅÈÊÇÅÁË ÎÄÍ ÇÊ ÆÇÆßÌÊÆËÊÄ ÍÌÁÄÁÌ ÅË Ý Ù ØÚÓ ÖÒØÒÓ Ò ÂÓÖ Å Ó Ý ÏºÈº ͹Á º¼¼ ÖÙÖÝ ¾¼¼¼ ØÖØ Ï ÒØÖÓÙ Ò ØÙÝ ÓÑÔÖÓÑ ÚÐÙ ÓÖ ÒÓÒ¹ØÖÒ ÖÐ ÙØÐØÝ Ñ Ø ¹ÓÑÔÖÓÑ ÚÐÙº ÁØ ÐÓ ÐÝ ÖÐØ ØÓ Ø ÓÑÔÖÓ¹ Ñ ÚÐÙ ÒØÖÓÙ Ý ÓÖÑ
ÇÔ Ò ÈÖÓ Ð Ñ Ò Ø ¹Ë Ö Ò È Ö¹ØÓ¹È Ö ËÝ Ø Ñ Æ Ð Û Ò À ØÓÖ Ö ¹ÅÓÐ Ò Ò Ú ÖÐÝ Ò ËØ Ò ÓÖ ÍÒ Ú Ö ØÝ ËØ Ò ÓÖ ¼ ÍË Û Ò ØÓÖ Ý Ò º Ø Ò ÓÖ º Ù ØØÔ»»ÛÛÛ¹ º Ø Ò ÓÖ º Ù ØÖ غ ÁÒ È Ö¹ÌÓ¹È Ö È¾Èµ Ý Ø Ñ ÙØÓÒÓÑÓÙ ÓÑÔÙØ Ö
ÅÓÖÐ ÀÞÖ ÅÖØ ÈÓÛÖ Ò ËÓÒ Ø ÀÐØ ÁÒ ÙÖÒ Ý ÖØÓРͺ ÏÖ ÔÖØÑÒØ Ó ÓÒÓÑ ÍÒÚÖ ØÝ Ó ÅÒÒÑ ¹½ ½ ÅÒÒÑ ÛÖÓÒºÙÒ¹ÑÒÒѺ Ò ÅÖÙ ÒÐÙ ÎÖÒØ ÃÖÒÒÚÖ ÖÙÒ ¹½ ÅĐÙÒÒ ÑÖÙ ºÒÐÙÚÖÒØºº ÙÙ Ø ¾¼¼½ ØÖØ ÁÒÚÙÐ ÑÓÖÐ ÞÖ ÒÒÖ Ý ÐØ Ò ÙÖÒ Ò ÑÓÒÓÔÓÐ
Real Business Cycles with Disequilibrium in the Labor Market: A Comparison of the U.S. and German Economies
Working Paper No. 5 Real Business Cycles with Disequilibrium in the Labor Market: A Comparison of the U.S. and German Economies by Gang Gong and Willi Semmler University of Bielefeld Department of Economics
ÓÒ ÖØÓÒ ÓÒ ÙÖÓÔÒ ÓÒÚÖ ÓÒ ÔÖÐÐÐ ÓÖÔÙ ÅÐÒ ÅÒ ÍÒÚÖ Ø ÈÚ ÑÐÒÑÒÑкÓÑ ÖÒ ¾ ÂÒÙÖÝ ¾¼¼ ½ ½µ ¾µ Ñ Ó Ø ÛÓÖ Ì ÛÓÖ ÒÐÝ Ø ÇÆÎÊË ÓÖ ÓÒÚÖÐ ÓÒ ØÖÙØÓÒ µ Û Ò ÓÙÒ Ò Ø Ç ÓÖÔÙ Ñ ÙÔ Ó Ø ØÖÒ ÐØÓÒ Ó ÚÒ ÔØÖ Ó Ó³ ÒÓÚÐ ÁÐ ÒÓÑ ÐÐ
PROTOCOLS FOR SECURE REMOTE DATABASE ACCESS WITH APPROXIMATE MATCHING
CERIAS Tech Report 2001-02 PROTOCOLS FOR SECURE REMOTE DATABASE ACCESS WITH APPROXIMATE MATCHING Wenliang Du, Mikhail J. Atallah Center for Education and Research in Information Assurance and Security
Chen Ding Yutao Zhong Computer Science Department University of Rochester Rochester, New York U.S.A. cding,ytzhong @cs.rochester.
ÓÑÔ Ð Ö¹ Ö Ø ÊÙÒ¹Ì Ñ ÅÓÒ ØÓÖ Ò Ó ÈÖÓ Ö Ñ Ø Chen Ding Yutao Zhong Computer Science Department University of Rochester Rochester, New York U.S.A. cding,ytzhong @cs.rochester.edu ABSTRACT ÙÖ Ø ÖÙÒ¹Ø Ñ Ò ÐÝ
Archiving Scientific Data
Archiving Scientific Data Peter Buneman Sanjeev Khanna Ý Keishi Tajima Þ Wang-Chiew Tan Ü ABSTRACT Ï ÔÖ ÒØ Ò Ö Ú Ò Ø Ò ÕÙ ÓÖ Ö Ö Ð Ø Û Ø Ý ØÖÙØÙÖ º ÇÙÖ ÔÔÖÓ ÓÒ Ø ÒÓ¹ Ø ÓÒ Ó Ø Ñ Ø ÑÔ Û Ö Ý Ò Ð Ñ ÒØ ÔÔ Ö
Push-communities. Pull-communities. Wrapped Services ... ... processors hardwarecircuits peripherals PCshopping
ÓÑÔÓ Ò Ò ÅÒØÒÒ Ï¹ ÎÖØÙÐ ÒØÖÔÖ ÓÙÐÑ ÒØÐÐ ½ Ò ÖÑ Å ¾ Ò ØÑÒ ÓÙÙØØÝ ¾ Ò Ñ ÐÑÖÑ Ò ÂÑ Ö ½ ËÓÓÐ Ó ÓÑÔÙØÖ ËÒ Ò ÒÒÖÒ ÍÒÚÖ ØÝ Ó ÆÛ ËÓÙØ ÏÐ Ù ØÖÐ ÓÙÐÑ ºÙÒ ÛºÙºÙ ÔÖØÑÒØ Ó ÓÑÔÙØÖ ËÒ ÈÙÖÙ ÍÒÚÖ ØÝ ÍË ºÔÙÖÙºÙ ¾ ÔÖØÑÒØ
drop probability maxp
ÓÑÔÖ ÓÒ Ó ÌÐ ÖÓÔ Ò ØÚ ÉÙÙ ÅÒÑÒØ ÈÖÓÖÑÒ ÓÖ ÙÐ¹Ø Ò Ï¹Ð ÁÒØÖÒØ ÌÖ ÒÐÙ ÁÒÒÓÒ Ö ØÓ ÖÒÙÖ ÌÓÑ ÐÖ Ö ØÓÔ ÓØ ËÖ ÅÖØÒ ÅÝ ËÔÖÒØ ÌÄ ÙÖÐÒÑ ÍË ßÓØ ÒÐÙÐ ÔÖÒØÐ ºÓÑ ËÐÞÙÖ Ê Ö Ù ØÖ ßÖ ØÓºÖÒÙÖ ÌÓÑ ºÐÖÐ ÐÞÙÖÖ ÖºØ ½ ÍÒÚÖ Ø
Section VI Formats for PIP Project Management
Manual for PIP Project Management (Version3.0, August 2010) Section VI Formats for PIP Project Management Manual for PIP Project Management Section VI Formats for PIP Project Management Section VI provides
P1 P2 P3. Home (p) 1. Diff (p) 2. Invalidation (p) 3. Page Request (p) 4. Page Response (p)
ËÓØÛÖ ØÖÙØ ËÖ ÅÑÓÖÝ ÓÚÖ ÎÖØÙÐ ÁÒØÖ ÖØØÙÖ ÁÑÔÐÑÒØØÓÒ Ò ÈÖÓÖÑÒ ÅÙÖÐÖÒ ÊÒÖÒ Ò ÄÚÙ ÁØÓ ÔÖØÑÒØ Ó ÓÑÔÙØÖ ËÒ ÊÙØÖ ÍÒÚÖ ØÝ È ØÛÝ ÆÂ ¼¹¼½ ÑÙÖÐÖ ØÓ ºÖÙØÖ ºÙ ØÖØ ÁÒ Ø ÔÔÖ Û Ö Ò ÑÔÐÑÒØØÓÒ Ó ÓØÛÖ ØÖÙØ ËÖ ÅÑÓÖÝ Ëŵ
The ASCII Character Set
The ASCII Character Set The American Standard Code for Information Interchange or ASCII assigns values between 0 and 255 for upper and lower case letters, numeric digits, punctuation marks and other symbols.
Optimal Crawling Strategies for Web Search Engines
Optimal Crawling Strategies for Web Search Engines J.L. Wolf, M.S. Squillante, P.S. Yu IBM Watson Research Center ÐÛÓÐ Ñ Ô ÝÙÙ ºÑºÓÑ J. Sethuraman IEOR Department Columbia University [email protected]
Resource Management for Scalable Disconnected Access to Web Services
Resource Management for Scalable Disconnected Access to Web Services Bharat Chandra, Mike Dahlin, Lei Gao, Amjad-Ali Khoja Amol Nayate, Asim Razzaq, Anil Sewani Department of Computer Sciences The University
Best Place to Find Information For a Wedding?
ÔÔ Ö Ò ÈÖÓ Ò Ó Ø Ø ÁÒØ ÖÒ Ø ÓÒ Ð ÓÒ Ö Ò ÓÒ Ö Ø ØÙÖ Ð ËÙÔÔÓÖØ ÓÖ ÈÖÓ Ö ÑÑ Ò Ä Ò Ù Ò ÇÔ Ö Ø Ò ËÝ Ø Ñ ¾¼¼¼ Designing Computer Systems with MEMS-based Storage Steven W. Schlosser, John Linwood Griffin, David
desired behaviour (global constraints) composite system putative behaviour: putative agents, actions, etc.
Ó Ð¹ Ö Ú Ò ÔÔÖÓ ØÓ Ö ÕÙ Ö Ñ ÒØ Ò Ò Ö Ò ËØ Û ÖØ Ö Ò Ø Û ÖØº Ö ÒÙÛ º ºÙ Ô ÖØÑ ÒØ Ó ÓÑÔÙØ Ò ÍÒ Ú Ö ØÝ Ó Ø Ï Ø Ó Ò Ð Ò Ö ØÓÐ ØÖ Ø ÒÙÑ Ö Ó Ö ÒØ ÔÔÖÓ ØÓ Ö ÕÙ Ö Ñ ÒØ Ò Ò Ö Ò Ù Ø Ø Ø Ò Û Ô Ö Ñ Ñ Ö Ò ÙÔÓÒ Ø Ù Ó
é é ä ä é ö é é ò é ó é Ü ä Ü ä ä
é é é ä èé èé ö ß é éé é é é ß ß ßß ß é é é é ä ä ä ä ä é ä ä éé é ä é é ä ä é ä ö é é ò é é ó é Üä Üää à ò éè ì é è èè è ó üü èè è ü è è é é ä éé óé ä ìò ì é ä é ó ó é é ó é éé é é Ü é é ò ò ò ä ää é
Open Programmable Architecture for Java-enabled Network Devices
Open Programmable Architecture for Java-enabled Network Devices Tal Lavian, Nortel Networks - Advanced Technology Center Robert F. Jaeger, University of Maryland Jeffrey K. Hollingsworth, University of
How to create OpenDocument URL s with SAP BusinessObjects BI 4.0
How to create OpenDocument URL s with SAP BusinessObjects BI 4.0 Creator: Twitter: Blog: Pieter Verstraeten http://www.twitter.com/pverstraeten http://www.pieterverstraeten.com/blog Hi, Thanks for downloading
An Investigation of Geographic Mapping Techniques for Internet Hosts
An Investigation of Geographic Mapping Techniques for Internet Hosts Venkata N. Padmanabhan Microsoft Research Lakshminarayanan Subramanian Ý University of California at Berkeley ABSTRACT ÁÒ Ø Ô Ô Ö Û
Finding Near Rank Deficiency in Matrix Products
TR-CS-98-13 Finding Near Rank Deficiency in Matrix Products Michael Stewart December 1998 Joint Computer Science Technical Report Series Department of Computer Science Faculty of Engineering and Information
HTML Codes - Characters and symbols
ASCII Codes HTML Codes Conversion References Control Characters English version Versión español Click here to add this link to your favorites. HTML Codes - Characters and symbols Standard ASCII set, HTML
To: Enclosed is a packet of distribution election forms for your review and signature:
Date: To: Re: Distribution from SPS Retirement Plan Enclosed is a packet of distribution election forms for your review and signature: Withdrawal Request & Benefit Selection Form Special Tax Notice Regarding
Finding ASCII Codes for Special Fonts and Characters
Finding Codes for Special Fonts and Characters WingDings and WebDIngs are special fonts that are pre-installed on all copies of Microsoft Windows XP. These fonts do not contain letters and numbers as we
Working Paper Simulating Tail Probabilities in GI/GI.1 Queues and Insurance Risk Processes with Subexponentail Distributions
econstor www.econstor.eu Der Open-Access-Publikationsserver der ZBW Leibniz-Informationszentrum Wirtschaft The Open Access Publication Server of the ZBW Leibniz Information Centre for Economics Boots,
ASCII control characters (character code 0-31)
ASCII control characters (character code 0-31) DEC HEX 0 00 NUL Null char 1 01 SOH Start of Heading 2 02 STX Start of Text 3 03 ETX End of Text 4 04 EOT End of Transmission
