INTERNATIONAL ORGANIZATION FOR STANDARDIZATION

ORGANISATION INTERNATIONAL NORMALIZATION

ISO/IEC JTC1/SC29/WG11

CODING OF MOVING PICTURES AND ASSOCIATED AUDIO

 

 

ISO/IEC JTC1/SC29/WG11/N2463

MPEG 98/ Atlantic City

 

 

 

 

 

 

Title: MPEG-7 Evaluation Process Document

Source: Requirements

Status: Approved

Date: October 1998

 

TABLE OF CONTENT

1. INTRODUCTION *

1.1. General Context *

1.2. Intended Use of this Document *

1.3 Detailed Time Schedule and Contacts *

1.4 What is Called For *

2. EVALUATION OF DESCRIPTORS *

2.1 Evaluation Criteria *

2.2 Evaluation Procedures *

2.3 Form for Proposing Descriptor(s) *

2.4 Summary of Proposal's Content for Descriptors *

3. EVALUATION OF DESCRIPTION SCHEMES *

3.1 Evaluation Criteria *

3.2 Evaluation Procedures *

3.3 Form for Proposing Description Scheme(s) *

3.4 Summary of Proposal's Content for Description Schemes *

4. EVALUATION OF DDL *

4.1 Evaluation Criteria *

4.2 Evaluation Procedures *

4.3 Form for Proposing DDL *

4.4 Summary of Proposal's Content for DDL *

5. EVALUATION OF CODING SCHEMES FOR DESCRIPTIONS *

5.1 Evaluation Criteria *

5.2 Evaluation Procedures *

5.3 Content Data *

5.4 Form for Proposing Coding Schemes for Descriptions *

5.5 Summary of Proposal's Content for Coding Schemes for Descriptions *

6. DEVELOPING THE MPEG-7 STANDARD AFTER THE CFP *

6.1 The Evaluation of the Proposals *

6.1.1. The February meeting *

6.1.2. The March meeting *

6.2 The eXperimentation Model *

6.3 The Core Experiment (CE) Process *

7. MPEG-7 Content set *

Annex A: Distribution of MPEG-7 content set *

Annex B: MPEG-7 evaluation pre-registration form *

Annex C: MPEG-7 proposal format *

Annex D: Cover Page for each proposal *

Annex E: Form to fill out for Descriptors *

Annex F: Form to fill out for Description Schemes *

Annex G: Form to fill out for Coding Schemes *

Annex H: MPEG-7 content set *

1 MPEG-7 Audio Content Set *

2 MPEG-7 Still images Content Set *

3 MPEG-7 Video Content Set *

Detailed list of items: *

Annex I: Licensing Agreement for the MPEG-7 Content Set *

MPEG-7 Audio Content Set *

MPEG-7 Still images Content Set *

MPEG-7 Video Content Set *

Annex J1: Descriptor evaluation output sheet: Feature evaluation *

Annex J2: Descriptor evaluation output sheet: Proposal evaluation *

Annex J3: Descriptor evaluation output sheet: Evaluation conclusion *

Annex K1: Description Schemes proposal evaluation sheet *

Annex K2: Description Schemes evaluation conclusions *

Annex L: DDL proposals evaluation output sheet *

General Information *

PART 1: Evaluation of the DDL proposal against the requirements *

PART 2: Ability to define other Ds, DSs, Descriptions *

PART 3: Summary of the evaluation *

Annex M1: Coding Schemes proposal evaluation sheet *

ANNEX M2: CODING SCHEMES EVALUATION CONCLUSION *

ANNEX N: UML a graphical notation for DS & D proposals. *

 

 

1. INTRODUCTION

 

1.1. General Context

The scope of the future MPEG-7 standard is described in the documents "MPEG-7 Context and Objectives", "MPEG-7 Requirements", and "MPEG-7 Applications." All documents are publicly available from the MPEG web site: http://drogo.cselt.stet.it/mpeg/.

As with all MPEG standards, the MPEG-7 Standard will be drafted in a collaborative process, following a Call For Proposals (CFP), merging and improving the relevant proposals. The proposal evaluation and development of the standard which will follow is conducted by MPEG experts.

Evaluation starts with assessing the scope and technical merits of each proposal. Then, an initial Experimentation Model (XM) is built from a first selection of parts of all proposals. This starts the collaborative phase of the standardization. During the collaborative phase, the XM is updated and improved in an iterative fashion till MPEG-7 reaches the Working Draft (WD) stage. Improvements to the XM are made through Core Experiments (CEs), which are prompted by the contribution of new elements for the standard. CEs will be defined to test the contributed tools within the framework of the XM according to well-defined test conditions and criteria. The goal is to develop the best possible XM.

Finally, those parts of the XM (or of the Working Draft) that correspond to the normative elements of MPEG-7 will be standardized.

 

1.2. Intended Use of this Document

Evaluation takes place at all stages of the MPEG-7 process, in order to select appropriate items for inclusion in the standard. This document however, focuses only on specifying the test and evaluation process for responses to the CFP. , The test and evaluation process will take place at the evaluation meeting in February 1999, before the collaborative phase starts. This document also presents the time schedule for the whole process, the content of the proposals, and other guidelines important to proposers.

This document is closely related to the Proposal Package Description (PPD) document, which introduces the general context of the call. The two documents together (PPD and Evaluation document) constitute what we call the Call For Proposal (CFP). People intending to answer the MPEG-7 Call for Proposal should find all the information they need regarding the mechanics of proposal evaluation in these two documents.

In addition, proposers should also read the "MPEG-7 Context and Objectives", "MPEG-7 Requirements," "MPEG-7 Applications" documents, and other documents listed in the Call for Proposals in order to gain a better understanding of the expected use and content of MPEG-7.

In case of any remaining questions about the current document, you can directly contact its editor (see 1.3).


1.3 Detailed Time Schedule and Contacts

As stated in the introduction, this document aims to specify the evaluation process that will be applied to proposals. The evaluation itself will be mainly conducted in February 1999. However, the entirety of evaluation-related issues will be conducted in a broader time frame. For sake of clarity, the time frame of the whole process is detailed as follows:

 

 

These achievements conclude the initial evaluation phase covered in this document.

 

 

Contacts:

All the relevant contacts for proposers are listed in the following:

(1) For any questions related to this document, you may contact the editor of this document:

Sylvie Jeannin

Philips Research USA

Tel: +1 914 945 64 46

sjn@philabs.research.philips.com

 

(2) Regarding the MPEG-7 content set, the contact person is:

Seungyup Paek

Columbia University New Media Technology Center - USA

syp@ctr.columbia.edu.

(until November 1, 1998—afterwards, please consult http://www.cselt.it/mpeg/ for details.)

Note that the content set is described in Annex H, its distribution process is detailed in Annex A as well as in W2468, and its conditions of use are detailed in Annex I as well as in W2466.

 

(3) For any question about evaluation pre-registration, please contact Michael F. Vetter (See below). Note that the pre-registration form, presented in Annex B, must be filled out and sent by December 1st to:

 

Michael F. Vetter

TASC

55 Walkers Brook Drive

Reading, MA 01867-3297

USA

Tel: +1 781-942-2000

Fax: +1 781-942-9507

Mfvetter@tasc.com

Rob Koenen

KPN Research

PO Box 421

2260 Leidschendam

The Netherlands

Tel: +31 70 332 53 10

Fax: +31 70 332 55 67

r.h.koenen@research.kpn.com

Cc : Leonardo Chiariglione

Convenor WG11

CSELT

Via G. Reiss Romoli, 274

10148 Torino, ITALY

Tel.: +39 11 228 6120

Fax: +39 11 228 6299

leonardo.chiariglione@cselt.it

 

(4) For the registration to the Lancaster meeting, the contact persons are:

Ed Hartley (e.hartley@lancaster.ac.uk), and Barbara Hickson

DMRG

Lancaster University

Lancs, LA1 4YR Tel: (44) 1524 593808

England Fax: (44) 1524 593608

Information about the Lancaster meeting and about how to register can be found in N2465. Note that there are facilities fees included in this registration.

 

(5) For specific matters about the procedure for proposal submission, please contact Michael F. Vetter or Rob Koenen (See above). Each proposal must be submitted by February 1st 1999 in electronic format, preferably Microsoft Word 6.0. Word 97 or Framemaker may be used in case Word 6.0. is inadequate for the document. An ftp site will be set up for these matters. All the needed related information will be provided after pre-registration.

 

1.4 What is Called For

In order to develop the MPEG-7 standard, MPEG calls for the following:

A. For the normative part:

While D, DS, DDL, and coding scheme proposals will be evaluated in February 1999 following the procedures defined in the current document, the proposals addressing MPEG-7 systems tools will not be part of the MPEG-7 Evaluation process in February 1999. These tools will be considered at the Seoul MPEG meeting. The proposers are kindly invited to present and show a demo regarding these tools at the Seoul MPEG meeting, in March 1999.

B. For the development of the standard, to be used in the XM and play an important role in the standardization process:

The proposals from group B are needed for the design and improvement of the XM. They will be discussed during the Lancaster meeting, and in the MPEG-7 Evaluation Ad Hoc Group(s) between Lancaster and Seoul. The conclusion of these discussions will be submitted as input to Seoul for the design of XM1.0. Whether and how these techniques will be formally evaluated will be decided after the XM1.0 is built.

It should be noted that it is not necessary for a proposal to address all of the elements listed above. It is possible for a proposal, for example, to only propose descriptors for some set of features, or description schemes or parts of description schemes. Selected aspects of different proposals will be incorporated into a common model (the XM) during the collaborative phase of the standard with the goal of building the best possible model.

The proposer should provide:

  1. A proposal document according to the format in Annex D. It should highlight all information that will help MPEG in identifying the strengths of the proposal.
  2. The proposer is strongly encouraged to give a presentation/demonstration of the proposal using relevant MPEG-7 test sets at the evaluation meeting in Lancaster. If the MPEG-7 test set is insufficient for the demonstration, another data set may be used. In this case the data set used must be made available to MPEG on the same conditions as the MPEG-7 Test and Evaluation Material (see annex I).

 

2. EVALUATION OF DESCRIPTORS


2.1 Evaluation Criteria

The feature captures important characteristic(s) of the AV material.

Gives better retrieval accuracy (e.g. precision, recall rate) with respect to other descriptors for the same feature.

The Descriptor is applicable to a wide range of application domains.

The Descriptor expresses the given feature(s) precisely, and completely.

The Descriptor represents the features at multiple levels of abstraction.


2.2 Evaluation Procedures

The Descriptors proposals will be evaluated by groups of experts drawn from participants at the evaluation meeting, following the evaluation procedure described as follows:

 

Proposals will be categorized according to the feature they represent.

Groups of MPEG experts will be formed to evaluate proposals by categories.

The evaluation will be based on the following steps:

 

  1. Evaluate feature relevance
  2. Goal: Understand the feature, and assess how it captures important characteristic(s) of the AV material.

    How: paper evaluation, from all paper descriptions of all proposals addressing this feature.

    Who: experts

    Output: feature evaluation sheet (see Annex J1), which includes represented features and their importance, with associated explanations.

     

    Then, for each proposal independently within a given category:

     

  3. Evaluate the paper document.
  4. Goal: The goal of this step is to have an initial assessment of the proposal based on the documentation included (questionnaire, summary and detailed structured description).

    During this step, experts should prepare eventual questions to ask to the proposers at the next step, to clarify some points if needed.

    How: Experts will review/analyze this information against requirements and evaluation criteria, first without being influenced by the participation of the proposers. A limited duration should be given to this step.

    Who: Experts

    Output: the proposal evaluation sheet (Annex J2, proposal summary and criteria table: all relevant criteria), with a short summary of the proposal, and its first evaluation along the analyzed criteria.

  5. Hear presentation / see demonstration.

Goal: To enhance the understanding of the proposal by the experts.

The presentation shall demonstrate the appropriateness of the solution, and disclose the appropriate range of use. The demonstration will provide evidence of (some of) the functionalities claimed.

Who: experts and proposer(s) whose submission is evaluated.

How: Experts will interact with the proposer(s) through a presentation and possibly a demonstration.

Both demonstration and presentation will each have a 10 minutes time limit.

Output: Updated (modified/completed) proposal evaluation sheet (criteria table and eventually summary of the proposal).

Remarks:

 

If the proposed descriptor allows similarity based retrieval:

 

  1. Evaluate results of similarity based retrieval

Goal: To refine the evaluation of the effectiveness and expression efficiency of the Descriptor.

Who: Experts

How: Proposers will provide

The quality of similarity based retrieval using the proposed descriptor will be evaluated using the MPEG-7 test sets sample inputs provided on-site, or samples provided by the proposer. The test will be run as follows:

  1. An input item will be selected from the database elements (the input item will be selected by experts so that it is appropriate for the described feature).
  2. The retrieval program will be run to identify and rank the n database items closest to the input. (Note that the speed of the retrieval program implementation will not be used as criterion at this stage of the evaluation).
  3. The ranking provided by the retrieval program will be judged by experts following their own perception of similarity and dissimilarity based on the feature.

This process will be repeated on several items of the test set, until the evaluators reach a consensus.

Remarks:

Output: Updated proposal evaluation sheet (criteria table, effectiveness and expression efficiency criteria only - similarity measure speed should not be taken into account).

 

For each proposal again:

 

5) Produce a global evaluation conclusion

Goal: To summarize the results of the previous steps to allow the selection of technologies for inclusion in the XM1.0, or in Core Experiments.

Who: Experts.

How: Experts arrive at a consensus.

Output: Evaluation summary of the proposal evaluation sheet (Annex J2) and evaluation conclusions sheet (see Annex J3), the final decision on this being taken as the Seoul meeting.


2.3 Form for Proposing Descriptor(s)

When proposing a Descriptor, each proposer has to fill the form enclosed in Annex E. Besides, proposers have to fill the proposal cover page (Annex D), and then refer to the proposal format guidelines given in Annex C.


2.4 Summary of Proposal's Content for Descriptors

 

The paper proposal, to be sent by February 1st, 1999, must include:

 

At the Lancaster meeting, proposers provide:

 


3. EVALUATION OF DESCRIPTION SCHEMES


3.1 Evaluation Criteria

The DS is applicable for a wide range of applications. "Applicable" means directly usable or usable as a component of a larger DS.

The DS provides an off the shelf solution for a given application domain. For this application domain, it takes into account relevant Descriptors and relevant relations between the Descriptors.

The DS can provide abstractions at multiple levels. An example is a hierarchical scheme where the base layer gives a coarse description and successive layers give more refined descriptions. The type of hierarchy used is appropriate for the purpose of the DS. Descriptors within the DS are amenable to being prioritized.

Part of the DS can be used effectively:

The DS is easily extensible to other DSs (in a way similar to inheritance in Object-Oriented Programming).

A minimal number of Descriptors and possible relationships are used to meet the needs of a particular application domain.


3.2 Evaluation Procedures

The evaluation procedure for Description Schemes is as follows:

 

The proposals are categorized according to the media addressed, applications addressed, functionality etc. Groups of MPEG experts will be formed to evaluate proposals by categories.

 

This phase relies on:

Both are essential to the complete evaluation of DS proposals. They are handled as described in the following.

 

The evaluation will be based on the following steps:

 

1) Evaluate the paper document.

Goal: The goal of this step is to have an initial assessment of the proposal based on the documentation included (questionnaire, summary and detailed structured description).

During this step, experts should prepare eventual questions to ask to the proposers at the next step, to clarify some points if needed.

How: Experts will review/analyze this information against requirements and evaluation criteria, first without being influenced by the participation of the proposers. A limited duration should be given to this step.

Who: Experts

Output: the proposal evaluation sheet (Annex K1, proposal summary and criteria table: all relevant criteria), with a short summary of the proposal, and its first evaluation along the analyzed criteria.

2) Hear presentation / see demonstration.

Goal: To enhance the understanding of the proposal by the experts.

The presentation shall demonstrate the appropriateness of the solution, and disclose the appropriate range of use. The demonstration will provide evidence of (some of) the functionalities claimed, and of how the DS satisfies the evaluation criteria.

Who: experts and proposer(s) whose submission is evaluated.

How: Experts will interact with the proposer(s) through a presentation and possibly a demonstration.

Both demonstration and presentation will each have a 10 minutes time limit.

Output: Updated (modified/completed) proposal evaluation sheet (criteria table and eventually summary of the proposal).

Remark: This step is not mandatory, but strongly encouraged. It gives an opportunity for the proposers to clarify, and provide additional evidence of facts that are not easy to demonstrate on paper.

 

3) Produce a global evaluation conclusion

Goal: To summarize the results of the previous steps. This should allow:

Who: Experts.

How: Experts arrive at a consensus.

Output: Evaluation summary of the proposal evaluation sheet (Annex K1) and evaluation conclusions sheet (see Annex K2), the final decision on this being taken as the Seoul meeting.

 

3.3 Form for Proposing Description Scheme(s)

When proposing a Description Scheme, each proposer has to fill the form enclosed in Annex F. Besides, proposers have to fill the proposal cover page (Annex D), and then refer to the proposal format guidelines given in Annex C.


3.4 Summary of Proposal's Content for Description Schemes

The paper proposal, to be sent by February 1st, 1999, must include:

 

At the Lancaster meeting, proposers provide:

 


4. EVALUATION OF DDL


4.1 Evaluation Criteria

  1. Compositional capabilities: The DDL shall supply the ability to compose a DS from multiple DSs.
  2. Platform independence: The DDL shall be platform and application independent. This is required to make the representation of content as reusable as possible even on grounds of changing technology.
  3. Grammar: The DDL shall follow a grammar that is unambiguous, and allow easy parsing (interpretation) by computers.
  4. Primitive data types: provide a set of primitive data types, e.g. text, integer, real, date, time/time index, version, etc.
  5. Composite datatypes: The DDL must be able to succinctly describe composite datatypes that may arise from the processing of digital signals (e.g., histograms, graphs, and rgb-values).
  6. Multiple media types: The DDL must provide a mechanism to relate Ds to data of multiple media types of inherent structure, particularly audio, video, audio-visual presentations, the interface to textual description, and any combinations of these.
  7. Partial instantiation: The DDL shall provide the capability to allow a DS to be partially instantiated by descriptors.
  8. Mandatory instantiation: The DDL shall provide the capability to allow the mandatory instantiation of descriptors in a DS.
  9. Unique identification: The DDL shall provide mechanisms to uniquely identify DSs and Ds so that they can be referred to unambiguously.
  10. Distinct name spaces: The DDL shall provide support for distinct name-spaces. Note: Different domains use the same descriptor for different features or different purposes.
  11. Transformational capabilities: The DDL shall allow the reuse, extension and inheritance of existing Ds and DSs.
  12. Relationships within a DS and between DSs: The DDL provides the capability to express the following relationships between DSs and among elements of a DS and express the semantics of these relations
  13. a) Spatial relations

    b) Temporal relations

    c) Structural relations

    d) Conceptual relations

  14. Relationship between description and data: The DDL shall supply a rich model for links and/or references between one or several descriptions and the described data.
  15. Intellectual Property Management: The DDL shall provide a mechanism for the expression of Intellectual Property Management and Protection (IPMP) for description schemes and descriptors.
  16. Real time support: The DDL shall desirably provide features to support real time applications (database output like electronic program guides)

 

4.2 Evaluation Procedures

The evaluation procedure for each Description Definition Language is as follows:

 

1) Evaluate the paper document.

Goal: The goal of this step is to have an assessment of the ability of the proposal to meet the requirements for the DDL described in the MPEG-7 requirements document (summarized above in section 4.1). During this step, experts should prepare eventual questions to ask to the proposers at the next step, to clarify some points if needed.

How: Experts will review/analyze this information against requirements and evaluation criteria, first without being influenced by the participation of the proposers. A limited duration should be given to this step.

Who: Experts.

Output: A copy of the form in Annex L part 1 will be completed, assessing how the DDL meets the requirements. .

 

2) Define a test set of Ds and DSs

Goal: The goal of this step is to select a set of DSs and Ds to be used in the following step of the evaluation.

How: A set of DSs and Ds will be identified from the DDL proposals by reviewing the proposal packages. The experts will ensure that the chosen DSs and Ds are of manageable complexity.

Who: Experts.

Output: A set of DSs and Ds to be used in the next step.

 

3) Hear presentation

Goal: To enhance the understanding of the proposal by the experts.

The presentation shall demonstrate the appropriateness of the solution, and the ability to use their DDL beyond the provided examples.

How: Experts will interact with the proposer(s) through a presentation, which will have 10 minutes time limit.

The evaluators will ask the proposer to show that the DDL can be used to generate DSs and Ds and related descriptions contained in the test set, when different from the examples of the proposal.

Who: Experts and proposers

Output: The output will be the completion of the Annex L Part 2, which is designed to provide an assessment of the ability of each DDL to generate DSs, Ds and descriptions.

 

4) Produce a global evaluation conclusion

Goal: To summarize the results of the previous steps to allow the selection of DDLs or parts of DDLs for inclusion in the XM1.0, or in Core Experiments.

How: Experts arrive at a consensus.

Who: Experts.

Output: Evaluation summary of the proposal evaluation sheet (Annex L Parts 1 and 2) and evaluation conclusions sheet (Annex L part 3).

 

Remark: The evaluation procedure for DDL will formally end at the 48th Vancouver meeting. The Lancaster meeting will provide a summary of each DDL’s capability, and an assessment of its suitability for the XM DDL and for participating in the Core Experiments. It will be further discussed on the reflector between Lancaster and Seoul. The conclusion of these discussions will be submitted as input to Seoul. Then, it is recommended that a Core Experiment be conducted until the Vancouver meeting, by utilizing the recommended DDL’s to construct the XM1.0 DSs and Ds. The result of this Core Experiment will conclude the DDL evaluation phase: the collaborative phase starts.

4.3 Form for Proposing DDL

No specific form has to be filled when submitting a DDL. Proposers only have to fill the proposal cover page (Annex D), and then refer to the proposal format guidelines given in Annex C.


4.4 Summary of Proposal's Content for DDL

The paper proposal, to be sent by February 1st, 1999, must include:

 

At the Lancaster meeting, proposers provide:

 

5. EVALUATION OF CODING SCHEMES FOR DESCRIPTIONS

5.1 Evaluation Criteria

Ability to losslessly compress descriptions instantiations.

It is possible to multiplex and stream the coded description.

The coded description shall be robust against transmission errors.

Ability to be applied to a wide range of descriptions

 

5.2 Evaluation Procedures

The main evaluation of proposals will be done based on the documentation included with the proposal (questionnaire, summary and detailed structured description). MPEG experts will review/analyze this information against requirements and evaluation criteria to identify promising techniques for the XM development. If the proposer has performed tests for bitstream verification of his/her coding scheme, either in combination with a D/DS/DDL or separate, a demo is highly encouraged. Furthermore, a short presentation should be given to help the MPEG experts better understand the proposed coding scheme and its possible extensibility and applicability to a wider range of descriptions.

5.3 Content Data

If bitstream verification is performed as part of the proposal, it should be based on encoding descriptions extracted from MPEG-7 content set in the addressed categories (see section 7). If other content is used, it should be justified why it was not possible to use the official material.

5.4 Form for Proposing Coding Schemes for Descriptions

When proposing a Coding Scheme, each proposer has to fill the form enclosed in Annex G. Besides, proposers have to fill the proposal cover page (Annex D), and then refer to the proposal format guidelines given in Annex C.

5.5 Summary of Proposal's Content for Coding Schemes for Descriptions

The paper proposal, to be sent by February 1st, 1999, must include:

 

At the Lancaster meeting, proposers provide:

6. DEVELOPING THE MPEG-7 STANDARD AFTER THE CFP

 

6.1 The Evaluation of the Proposals

The main objective of the Call For Proposals it to lead to the best possible starting of the collaborative phase, which is recognized as one of the strengths of MPEG.

It is reasonable to expect that the competitive process, which culminates at the evaluation meeting, will produce several top performing, possibly competitor, elements. Each of these will have strengths compared to each other, and the goal of the ensuing collaborative process is to produce convergence to a single standard combining the best features of all proposals. During this process, opportunities for further improvements will arise, and these should be verified and incorporated. The method to accomplish this work is called the Core Experiment process and the basis on which it works is the eXperimentation Model (XM).

6.1.1. The February meeting

For each proposal to be evaluated, all information shall be provided following the guidelines of the CFP, by February 1st 1999, so that MPEG members can study them to prepare the evaluation meeting in advance. The evaluation meeting will be held in Lancaster University (UK), February 15-19. During this meeting, MPEG members will:

Proposers will be given the opportunity to present and clarify their proposal.

The quality and the clarity of the proposal are the responsibility of the proposers.

Finally, proposers are encouraged to be present at the February meeting to help with the evaluation.

 

6.1.2. The March meeting

The results of the February meeting, together with the work done on the eXperimentation Model structure and started in October 1998, form the basis to build the first version of the eXperimentation Model. Its first version will be finalized in March 1999, at the MPEG Seoul meeting. This meeting will also specify the first set of Core Experiments to be performed.

 


6.2 The eXperimentation Model

 

The XM is a "common experimentation framework" which is used for further development of the standard. It will have components for evaluating and improving the DDL, DSs and Ds. Moreover, its associated experimentation database(s) will be more complete than the test sets used for evaluation. An XM may have different implementations within the MPEG group.

As a result of both the collaborative work on XM structure started at the Dublin meeting in October 1998 and the evaluation phase, an XM will be built and its first version will be finalized at the MPEG Seoul meeting, March 15-19, 1999. XM will consequently include proposals and/or parts of proposals that were either made before CFP within the Ad Hoc group on XM definition, or given as answers to the CFP.

After the XM is established, new elements can be brought to MPEG-7 following a Core Experiment procedure. They will be then evaluated inside the XM so that in the final standard, if two elements accomplish similar thing(s) in similar conditions, only the best will be chosen.


6.3 The Core Experiment (CE) Process


The Core Experiments (CEs) will provide a more complete evaluation of Ds, DSs, DDLs, combinations of Ds, DSs handling combinations of Ds, etc within XM.

The heart of this process is for multiple, independent, directly comparable experiments to be performed to determine whether or not a proposed algorithmic technique or syntax element has merit. A Core Experiment has to be completely and uniquely defined, so the results are unambiguous. In addition to the specification of the element to be evaluated (syntax, related algorithmic techniques…), a Core Experiment also specifies the parameters to be used so that the results can be compared. Therefore, a Core Experiment may include one or more of the following elements: benchmark databases, associated 'truth', benchmark queries, objective and subjective evaluation.

A Core Experiment is proposed by one or more MPEG experts. It is accepted by consensus, providing that two or more independent experts agree to perform the experiment. Conventionally this is interpreted to mean that the independent experts work for different companies (different divisions or subsidiaries of a company are not considered independent).

Normally, a Core Experiment will be conducted between two successive MPEG meetings. If necessary, and again by consensus, a Core Experiment may be continued until a further MPEG meeting.

The outcome of a Core Experiment is determined by consensus. If the experiment can be evaluated subjectively, then an informal "test" is performed by presenting the results in the MPEG meeting. First, the independent results should evidently be equivalent. Second, they should indicate whether or not the technique has merit. Since this will at times be difficult to determine because there is some merit but it is marginal, the consensus may be that an evaluation of implementation complexity will be used to make the final decision.

In case the technology presented in a Core Experiment is selected for entering the XM, this can only become effective if some MPEG member agrees to donate the source code that fully implements this technology.

 


7. MPEG-7 Content set

 

MPEG-7 is providing a content set, made of still images, audio, and audio-visual data, to be used for evaluation. This content set is described in Annex H. Its distribution process is detailed in Annex A as well as in W2468, and its conditions of use are detailed in Annex I as well as in W2466.

Each proposer planning either to show a demonstration or to participate to the similarity based retrieval evaluation shall use appropriate subsets of this content set. It is the responsibility of the proposer to choose the items of the content set that he/she will use for demonstration or similarity-based retrieval. These elements should be relevant to the evaluated proposal (e.g. color images for a color Descriptor), and in sufficient number to allow evaluators to draw conclusions from what is shown. Moreover, note that the choice of subset must be justified in the proposal/presentation.

In case the MPEG-7 content set is insufficient for demonstrating his/her proposal, the proposer may use his/her own test set and provide justifications for the same. In this case the content set used must be made available to MPEG under the conditions mentioned in Annex I.

The evaluators will judge the rationale of the selection of the content set, and take it into account when drawing conclusions from any evaluation step using it.

Annex A: Distribution of MPEG-7 content set

 

The following is the procedure to obtain the MPEG-7 content set:

 

  1. Each group (company, university, individual etc.) requesting MPEG-7 content must send e-mail to the primary contact (See below) for content set distribution by November 1st, 1998. The content set consists of all image, video and audio content that were reviewed for MPEG-7 evaluation. It will not be possible to obtain a subset of the content. For example, it will not be possible to only obtain the images, without the video or audio material. The MPEG-7 content set will require approximately 20 CDs. Each group must provide the following information. No requests will be honored if any of the following information is missing:

 

  1. Once all the requests have been received by the primary contact by November 1st, 1998, the primary contact will calculate the charge required by group n as follows:
  2. (C/N) + (Shipping charge for group n)

    C is the total cost required for the duplication of the content set.

    N is the total number of groups that have requested the content set by the November 1st deadline.

    Once the charges for each group have been calculated, the primary contact will send an e-mail to all the groups that have requested the content set, and notify each group of their respective payment amounts. This will be done by November 3rd, 1998. At this time, An approximation of the cost for each group is as follows:

    Approximate storage requirement of the content set: 20 CDs;

    Approximate cost to produce a minimum of 250 copies of the content set: $15,640;

    Approximate cost for each group, assuming that 100 groups request the content set: $160.

    These approximations do not include shipping charges or the tax. The cost is subject to change.

     

  3. Once each group receives the payment amount, they must do all of the following:

Conversion Media c/o Jon Taylor

625 North Milwaukee St. Suite 220.

Milwaukee, Wisconsin 53202.

U.S.A.

Phone: +1 888 999 1760

Email: jontaylor@conversionmedia.com

 

  1. The material will be shipped by November 25th, 1998.
  2.  

  3. Groups that did not request the content set by the November 1st, 1998 deadline will be able to obtain the content set by contacting members of MPEG-7 that receive the content set. However, the content set can only be made available subject to the terms and conditions specified in the MPEG-7 content distribution agreement (See N2466 or Annex I of N2463).

 

The primary contact for the content set distribution process described in this document is:

Seungyup Paek

Columbia University New Media Technology Center - USA

syp@ctr.columbia.edu.

 

 

 

Annex B: MPEG-7 evaluation pre-registration form

 

All organizations intending to submit proposals to the MPEG-7 evaluation and development process are requested to complete this pre-registration form and submit it to

 

Michael F. Vetter

TASC

55 Walkers Brook Drive

Reading, MA 01867-3297

USA

Tel: +1 781-942-2000

Fax: +1 781-942-9507

Mfvetter@tasc.com

Rob Koenen

KPN Research

PO Box 421

2260 Leidschendam

The Netherlands

Tel: +31 70 332 53 10

Fax: +31 70 332 55 67

r.h.koenen@research.kpn.com

Cc : Leonardo Chiariglione

Convenor WG11

CSELT

Via G. Reiss Romoli, 274

10148 Torino, ITALY

Tel.: +39 11 228 6120

Fax: +39 11 228 6299

leonardo.chiariglione@cselt.it

 

by the 1st December 1998.

 

Pre-registration of proposals is mandatory. For more information, please see the MPEG-7 Evaluation Document.

A pre-registration form must be submitted for each descriptor, description scheme, Description Definition Language (DDL), or coding scheme for descriptors submitted for evaluation. The same applies for the systems tools as well as for the non-normative tools which are useful for the development of the standard, such as extraction and search methods as well as evaluation and validation techniques.

 

I. Proponent Contact Information

Name:______________________________________________________________________

Company/Institution:__________________________________________________________

Address:____________________________________________________________________

___________________________________________________________________________

___________________________________________________________________________

Phone:______________________________________________________________________

Fax: _______________________________________________________________________

E-Mail:____________________________________________________________________________________

 

II. Proposal Content – Normative Tools

If this pre-registration refers to a normative tool, please circle which item it refers to:

 

Evaluation Item

Descriptor

Still images

 

Video

 

Audio

 

Synthetic visual content

 

Synthetic audio content

 

Others (please specify)

Description Scheme

Still images

 

Video

 

Audio

 

Synthetic visual content

 

Synthetic audio content

 

Others (please specify)

Description Definition Language (DDL)

Coding Scheme for Descriptions

Systems tool

 

Note: Please note that systems tools will not be part of the MPEG-7 Evaluation process in February 1999. These tools will be considered at the Seoul MPEG meeting.

 

III. Proposal Content – Non-Normative Tools

If this pre-registration refers to a non-normative tool useful for the development of the standard, please circle which evaluation item it refers to:

 

Extraction methods

Search Methods

Evaluation and validation techniques

 

IV. Participation in the Evaluation Ad Hoc Group Meeting

  1. Do you intend to be present the Evaluation Ad Hoc Group Meeting in Lancaster, UK and make a presentation regarding this proposal (circle one) ?
  2. Yes No

    Note: If ‘Yes’, don’t forget to register for the meeting.

     

  3. If ‘Yes’, do you intend to show a demo during your presentation (circle one) ?
  4. Yes No

  5. If ‘Yes’, what type of demo ? ________________________________________________
  6.  

     

  7. If ‘Yes’, do you need any audiovisual equipment for the demo, e.g. PAL video recorder ?

 

Note: No guarantees are given in advance regarding the equipment requested. Please contact the Evaluation Ad Hoc Group Meeting organizer for confirmation.

 

V. Additional information (circle one or both)

  1. Does your proposal have any special characteristic that you would like to mention now (circle one) ?

Yes No

If ‘Yes’, please state which:____________________________________________________

___________________________________________________________________________

___________________________________________________________________________

 

 

Signature:___________________________________________________________________

Date: ______________________________________________________________________

 

Annex C: MPEG-7 proposal format

 

1. For each proposal, proposers should provide a detailed paper describing the proposed technology, with emphasis on the components of the proposal that should be evaluated. The paper should:

 

2. For Descriptors, please provide the data needed to perform the similarity evaluation, if applicable, as stated in the current document. If the MPEG-7 test set is augmented with other material or a different test set is used, please state why the MPEG-7 test sets were not suitable. Also the data set used must be made available to MPEG on the same conditions as the MPEG-7 Test and Evaluation Material. (See Annex I).

Please give also:

2.1. Typical search / filtering tasks the proposal applies to.

2.2. Explanation of the similarity measure used, if any.

 

3. For Description Schemes the proposal must include a detailed explanation using English (and pseudo-code if required) and a graphical representation. It is strongly suggested that, to the extent possible, the graphical representation of the proposed description schemes follow the UML notations provided in Annex N. It is strongly encouraged that a presentation and demonstration be given to support the material contained in the written proposal. Provide a written description of any demonstration planned to accompany this proposal and detail the evaluation criteria that will be shown. If your DS demonstration depends on use of Descriptors, please provide a brief summary of these Descriptors (Descriptors proposed separately may simply be referenced by proposal ID). All demonstrations of DSs shall use the MPEG-7 Test Material (or subset consisting of all relevant material) unless the proposer can justify that other test material is necessary to demonstrate the DS.

 

4. For DDL, the submission must also include:

Note that for DDL, the description of how the proposal meets the MPEG-7 requirements should be made by completing a copy of annex L part 1.

 

 

Then, fill out the cover page (given in annex D) and the relevant form(s), given in annexes E, F, G and H, for each element (Descriptor, Description Scheme, Description Definition Language, Coding Scheme) in the proposal that MPEG should evaluate.

Note that a separate form must be completed for each Description Scheme and for each of the associated Descriptors if it is desired to have them separately considered/evaluated by MPEG.

 

Please note that brevity and clarity helps the evaluation.

 

 

 

Annex D: Cover Page for each proposal

 

One cover page is needed for each item (D, DS, DDL or coding scheme) of the proposal that you wish to be evaluated. For example, if you submit a DS that has 2 Ds and wish to have them all (the DS and 2 Ds) evaluated then you need to fill out three forms (one for DS, one for each D). The cover page must be filled out for each of these three forms.

 

 

 

Proposal Id: _____________

(Number obtained from pre-registration)

Name:_____________ _________________________________________

Company/Institution:_____ ______________________________________

 

 

1) Which item is proposed?

a) Descriptor

b) Description Scheme

c) Description Definition Language

d) Coding scheme

e) Others (please specify) ______________

 

  1. Do you have other proposal(s) related to this one? Please list the proposal Ids.

 

Annex E: Form to fill out for Descriptors

 

1. Which media has your descriptor been developed for and/or applied to? (Choose one or more)

(a) Video

(b) Audio

(c) Synthetic video/images

(d) Synthetic Audio

(e) Still Images

(f) Combined Audio-Visual

(g) Cross-Modal

(h) Others (please specify)______________

 

2. Which media do you use to extract your descriptor from? (Choose one or more)

(a) Video

(b) Audio

(c) Synthetic video/images

(d) Synthetic Audio

(e) Still Images

(f) Combined Audio-Visual

(g) Cross-Modal

(h) Others (please specify)______________

 

3. Describe the feature your descriptor is associated to.

 

4. Discuss why the above feature is important to MPEG-7. (For example, tasks it can perform, queries it can answer, the MPEG-7 requirements it satisfies)

5. If your descriptor is content-domain-specific (e.g. biomedical images) please answer the following:

5.1. Which domain does your solution support?

5.2. Is there evidence of support from professional organizations for the technology proposed and/or from applicable current domain practices?

5.3. Is this applicable to other domains? Please list.

 

6. State up to three MPEG-7 applications (you may refer to the Applications Document) to which your proposal applies.

 

7. If you provide a similarity measure please answer the following:

7.1. Is it intended to match human perception of similarity? (yes or no)

7.2. Does it allow ranking? (yes or no)

7.3. Describe the complexity of the matching tool if possible (e.g., O(nlogn)).

 

8. If you provide an extraction method please answer the following:

8.1. Does the method generate a confidence measure? (yes or no)

8.2. Does it require human intervention? (yes or no)

8.3. If your method does not require human intervention please provide the time and memory complexity of the method if possible (e.g., O(n2)).

 

  1. Will you run a (not more than 10 minutes) demonstration of your proposed solution at the evaluation Ad Hoc meeting in February? (yes or no)
  2. If the demonstration will be done using one or more MPEG-7 Test Sets, please Indicate them.
  3. Please indicate the MPEG-7 test sets for which you will be providing the descriptor value and system/program for similarity retrieval.
  4. Does the descriptor apply to global as well as local features? (For example, does it apply to a whole image as well as objects in the image?)

 

Annex F: Form to fill out for Description Schemes

 

1. Which media has your DS been developed for and/or applied to?

(a) Video

(b) Audio

(c) Synthetic video/images

(d) Synthetic Audio

(e) Still Images

(f) Combined Audio-Visual

(g) Cross-Modal

(h) Others (please specify)______________

 

2. What is the main functionality of your DS?

 

3. Is your DS generic to different content domains or is it content domain-specific (e.g. biomedical, etc)? If your description scheme is content-domain-specific please answer the following:

3.1. Which domain and tasks does your solution supports?

3.2. Is there evidence of support from professional organizations for the technology proposed and/or from applicable domain current practices?

3.3. Is this applicable to other domains? Do you have evidence of this?

4. State up to three MPEG-7 applications (for examples you may refer MPEG-7 Applications Document) to which your proposal applies.

 

5. Will you provide a presentation to explain your DS proposal and answer questions about it?

 

6. Will you provide a demonstration to show how your DS meets the evaluation criteria? Are you using the MPEG-7 test sets or are you providing your own test sets?

 

Annex G: Form to fill out for Coding Schemes

 

  1. Which part of the description does your coding scheme apply to?
  2. (a) DDL

    (b) Description scheme(s)

    (c) Descriptor(s)

    (d) Descriptor data

    (e) Others (please specify) ______________

     

  3. Is your coding scheme media specific?

If yes, which media has your coding scheme been developed for and/or applied to?

(a) Video

(b) Audio

(c) Synthetic video/images

(d) Synthetic Audio

(e) Still Images

(f) Combined Audio-Visual

(g) Cross-Modal

(h) Others (please specify)______________

 

3. If your coding scheme is application-domain-specific please answer the following:

3.1. Which domain and tasks does your solution support?

    1. Is this applicable to other domains? Do you have evidence of this?

 

  1. State up to three MPEG-7 applications (for examples you may refer MPEG-7 Applications Document) to which your proposal applies.
  2.  

  3. Was your coding scheme developed for a specific D, DS, DDL? (yes or no)

If so:

5.1. Which?

5.2. Would it be applicable to other Ds/DSs/DDLs? (yes or no)

5.3. If yes, how severe would the modifications be, and would this increase complexity or decrease efficiency?

 

6. Are you providing a demonstration?

Annex H: MPEG-7 content set

 

The MPEG-7 Content set is composed of three main categories: Audio (about 12 hours), Still images (about 7000 images) and Video (about 13 hours). The items in each category are described in the following sections.

1 MPEG-7 Audio Content Set

For supplemental content, please refer to the Video Content Set, especially the music and entertainment category. It contains much Audio-visual material, Also of potential interest are items with monologues and dialogues in diverse languages.

 

Category

Item num.

Short description

Source

Duration (min)

Radio

A1

Radio news broadcast

Radio-France 98

600

Music

A2

"Two Ton Shoe" Rock album

Two Ton Shoe

39

A3

Bruckner's Te Deum, and Mozart's Requiem

A. Lindsay

72

A4

Original composition, a capella. Voice only

S. Lounis

5

Audio

A5

Short sequences of solo instrument and other sounds

SQAM CD

30

A6

Pop song based on an A-A-C motif

Juergen Herre

5

 

2 MPEG-7 Still images Content Set

 

Category

Item

num.

Short description

Source

~number of images

Color photo.

S1

Amateur Photos of landscapes, people.

Tristan Savatier

300

S2

Photographs

Alejandro Jaimes

300

S3

Images of varying scene and color content, many outdoor and indoor images

Department of Water Resources, California

2000

S4

Indoor images of people, under varying lighting conditions and angles

Heinrich Hertz Institute

400

Grey-scale photo.

S5

Photographs

JPEG 2000

250

S6

Collection of print and digitized pictures

National Archives at Maryland

250

S7

Photos from the beginning of the century

Portuguese Photo. Center

260

Trademark logo

S8

Trademark images captured by a scanner (B&W images)

Korean Industrial Property Office

3000

Aerial photo.

S9

Aerial photos, monochrome and colored (5000x5000 pixels)

UC Santa Barbara

30

3D Range image

S10

205 range images and 7 complete 3-D colour models

National Research Council

200

  1. MPEG-7 Video Content Set

 

All items have been assigned to one of the three possible types: "Program", "Sequence" or "Shot". A "Program" is a piece of material of rather long duration which most of the time includes both the beginning and the end of the program. A "sequence" is a clip of a few minutes that has been extracted from a program. Finally, the "shot" type corresponds to either "program" or "sequences" where the list of editing effects (shot cut, fade, etc.) will be available. Finally, note that all video material is encoded in MPEG-1 format, except item V30.

 

Category

Type

Item

Num.

Short description

Source

Duration (min)

News

Shot

V1

Two complete TV news programs

Portuguese TV, RTP & SIC

60

Sequence

V2

Universal newsreel collection. B&W video.

National Archives at Maryland

30

Program

V3

Daily TV news program

Spanish TV, RTVE

29

V4

Weekly TV news program

Spanish TV, RTVE

19

Drama / Movie

Sequence

V5

"Art" movie: Hallo

Christoph Rodatz, GMD

4

V6

Movie: "La sombra de un cipres es alargada"

Spanish TV, RTVE

15

V7

TV Drama series: "Pepa y Pepe"

Spanish TV, RTVE

15

Program

V8

Sitcom (1 and 2)

Portuguese TV, RTP & SIC

60

V9

Filmed theater: The playboy of the western world

INA

120

Documentary

Shot

V10

"Science Eye": Bridge construction

NHK

9

Sequence

V11

5 clips of scientific documentaries

SFRS

25

V12

Documentary about buildings

Lancaster Television

5

Program

V13

Basic Ophthalmic Exam

Univ. of Tennessee

26

V14

Educational video: "A way with waste" and "Animals have young"

Singapore Ministry of Education

25

V15

Documentary about a village: "Santillana del Mar"

Spanish TV, RTVE

30

Sport

Sequence

V16

3 Sport Clips: Soccer, Cycling, Basketball

Spanish TV, RTVE

49

V17

2 Sport clips: Basketball, Golf

Korean Broadcasting Station

20

Program

V18

Soccer sequence

Samsung

45

Commercial

Sequence

V19

14 items of commercials in Korean

Samsung

7

Music video and games

Sequence

V20

Korea's pop singers' live music Show

Korean Broadcasting Station

10

V21

TV quiz program: "Saber y ganar"

Spanish TV, RTVE

15

V22

Music program: "Musica si"

Spanish TV, RTVE

15

Program

V23

Variety Show. First 30 minutes of complete program

Portuguese TV, SIC

30

Cartoon and animation

Sequence

V24

5 Selected pieces of cartoons

D'Ocon Film Productions

12

Program

V25

Harmony

Singapore Ministry of Education

12

V26

Cartoon "Don Quijote de la Mancha"

Spanish TV, RTVE

15

Home video

Shots

V27

Edited home video

LGERCA

34

Surveillance

Shots

V28

3 outdoors scenes

ETRI

9

V29

5 video sequences taken from a bridge over a speedway

UCL

28

Miscellaneous

 

Shots

V30

MPEG4 test sequence (Bream, Children, Fish)

Matsushita

1

 

 

 

 

 

Detailed list of items:

 

Audio Items:

 

ITEM A1

Category: Radio broadcast news.

Source: 98 Radio-France.

MPEG contact: R. Ronfard, INA.

Submission Item: 10

Description: One day from a special program broadcast from June 9th to July 13th, 1998. 17 hours a day, from 8:30 AM to 1:30 AM. Includes:

      1. World cup matches
      2. News bulletins in French
      3. News bulletins in English, German, Spanish, Portuguese
      4. Chronicles
      5. Interviews
      6. Magazines

Associated Data: Transmission control lists have been input in an Excel database.

Size / Duration: 10 hours have been selected from an original material of 17 hours

Format: Audio (mono) compression in MPEG-2, level 2, at 128 kbps, programs are segmented in files of 15 minutes

ITEM A2

Content Category: Music

Source: Two Ton Shoe

MPEG Contact: A. Lindsay, Riverland.

Submission Item: 18

Short Description: "Two Ton Shoe" Rock album.

Associated Data: None.

Size or Duration: 38:18

Format: 44100Hz, 16-bit AIFF

 

ITEM A3

Content Category: Music

Source: A. Lindsay

MPEG contact: A. Lindsay, Riverland.

Submission Item: 19

Short Description: Brussels Choral Society singing Bruckner's Te Deum, and Mozart's Requiem.

Associated Data: None.

Size or Duration: 72:03

Format: 44100Hz, 16-bit AIFF

ITEM A4

Content Category: Music.

Source: S. Lounis

MPEG contact: A. Lindsay, Riverland.

Submission Item: 20

Short Description: Original composition, a capella. Voice only.

Associated Data: French text.

Size or Duration: 5 min

Format: 48000Hz, 16-bit AIFF

 

ITEM A5

Content Category: Audio.

Source: SQAM CD.

MPEG contact: J. Herre, EBU/FhG.

Submission item: 35

Short Description: Short sequences of solo instrument and other sounds

Associated Data: /

Size or Duration: 54 items (some tracks have sub-indices). No sequences are exceeding 640MB in length

ITEM A6

Content Category: Audio.

Source: J. Herre

MPEG contact: J. Herre, EBU/FhG.

Submission item: 52

Short Description: Pop song based on an A-A-C motif

Associated Data: /

Size or Duration: One song (<5 min) in a variety of formats

Format: MIDI, AIFF, MPEG Layer 2 Audio

 

 

 

Still image items:

 

ITEM S1

Content Category: Amateur photographs.

Source: Tristan Savatier.

MPEG contact: R. Ronfard, INA.

Submission item: 24

Short Description: Amateur Photos of landscapes, people. Some nudity.

Associated Data: Keywords May be provided on request for each picture.

Size or Duration: 300 photographs

Format: PhotoCD

ITEM S2

Content Category: Photographs.

Source: Alejandro Jaimes.

MPEG contact: Seungyup Paek, Columbia Univ.

Submission item: 48

Short Description: General.

Associated Data: None

Size or Duration: ~300 images.

Format: JPEG

 

ITEM S3

Content Category: Photographs.

Source: Department of Water Resources, California

MPEG contact: S. Krishnamachari

Submission item: 50

Short Description: Images of varying scene and color content, many outdoor and indoor images. Images of same scene with different viewpoints. Many "similar" images that might come handy during evaluation.

Associated Data: None

Size or Duration: ~ 2000 images

Format: JPEG

ITEM S4

Content Category: Photographs.

Source: Heinrich Hertz Institute

MPEG contact: Sylvie Jeannin, LEP.

Submission item: 53

Short Description: Indoor images of people, under varying lighting conditions and angles

Associated Data: None

 

ITEM S5

Content Category: Photographs.

Source: JPEG 2000.

MPEG contact: M. Zeug, Iterated.

Submission item: 26

Short Description: Test images.

ITEM S6

Content Category: Photographs.

Source: National Archives at Maryland. http://www.nara.gov/nara/menus/audvis.html

MPEG contact: N. Nandhakumar, LG.

Submission item: 28

Short Description: Collection of print and digitized pictures.

ITEM S7

Content Category: Photographs.

Source: Portuguese Photograph Center.

MPEG contact: J. Torres, INESCN.

Submission item: 37

Short Description: Photos by Portuguese photographer from the beginning of the century. One of the pioneers of the cinema in Portugal. Some collections of photos are like a film and describe some sequence.

Size or Duration: 260 images. <300 MB

Format: JPEG

ITEM S8

Content Category: Subset of registered trademark images.

Source: Korean Industrial Property Office.

MPEG contact: M. Kim, ETRI.

Submission number: 34

Short Description: All images are trademarks captured by a scanner.

Size or Duration: 3000 images. 2.6 Mbytes.

Format: TIFF

 

ITEM S9

Content Category: Aerial photographs.

Source:

MPEG contact: B.S. Manjunath, UC Santa Barbara.

Submission item: 25

Short Description: Aerial photos, monochrome and colored.

Associated Data: Associated data available for some of the images

Size or Duration: 30 photos (each 5000x5000 pixels)

Format: tiff

ITEM S10

Content Category: Range images.

Source: National Research Council.

MPEG contact: E. Paquet.

Submission item: 1

Short Description: These images can be used not only as range images but also as generators for 2D images (e.g. same object under different lighting conditions)

Size or Duration : 205 range images and 7 complete 3-D colour models (all scanned)

Format: VRML 1

 

 

 

Video items:

 

ITEM V1

Source: RTP & SIC

MPEG Contact: F. Pereira, IST.

Submission item: 12A

Short Description: Two daily News program

Associated Data: List and timing of shots and semantics.

Size or Duration: 2 * 30 mins

ITEM V2

Source: National Archives at Maryland. http://www.nara.gov/nara/menus/audvis.html

MPEG Contact: N. Nandhakumar, LGERCA

Submission item: 30

Short Description: Universal newsreel collection. B&W video, archive of news.

Size or Duration: 30 mins

ITEM V3

Source: Spanish TV, RTVE.

MPEG contact: P. Salembier, UPC.

Submission item: 39A

Short Description: Clip from daily program "Telediario"

Size or Duration: 29 mins

ITEM V4

Source: Spanish TV, RTVE.

MPEG contact: P. Salembier, UPC.

Submission item: 39B

Short Description: Clip from weekly program "Informe Semanal"

Size or Duration: 19 min

ITEM V5

Source: Christoph Rodatz, GMD.

MPEG Contact: F. Nack.

Submission item: 2B

Short Description: Hallo

Associated Data: German text available, translation possible. Some composition.

Size or Duration: 4 min

ITEM V6

Source: Spanish TV, RTVE.

MPEG contact: P. Salembier, UPC.

Submission item: 41

Short Description: "La sombra de un cipres es alargada".

Associated Data: /

Size or Duration: 15 mins.

ITEM V7

Source: Spanish TV, RTVE.

MPEG Contact: P. Salembier, UPC.

Submission item: 42

Short Description: Drama series titled "Pepa y Pepe".

Size or Duration: 15 mins.

ITEM V8

Source: Portuguese TV, RTP & SIC.

MPEG Contact: F. Pereira, IST.

Submission item: 12B

Short Description: 2 Sitcoms

Associated Data: None.

Size or Duration: 2*30 mins

ITEM V9

Source: INA.

MPEG contact: R. Ronfard, INA.

Submission item: 16

Short Description: The playboy of the western world, filmed for TV. French.

Associated Data: Text available

Size or Duration: Two hours.

ITEM V10

Source: NHK.

MPEG Contact: M. Shibata, NHK.

Submission item: 4

Short Description: "Science Eye" Scientific Video about bridge construction.

Associated Data: Cuts, Japanese text overlays, narration text attached. Text descriptions (English) of cut-content will be attached.

Size or Duration : 9 min

ITEM V11

Source: SFRS.

MPEG Contact: P. Faudemay

Submission item: 5

Short Description: 5 sequences from scientific movies

(1) (1974) Lascaux pictures. French and English; (2) (1990) An island with flamingoes. French, English and Spanish; (3) (1979) Social life of Antelope. French; (4) (1973) Study on the Nyamgatoms tribe. French; (5) (1964) Scientific missions in the Kerguelen islands. French.

Associated Data: 3-page script for each movie.

Size or Duration: 5 * 5 min

ITEM V12

Source: Lancaster Television.

MPEG Contact: E. Hartley, Lancaster Univ.

Submission item: 7

Short Description: Show reel

Associated Data: None.

 

 

ITEM V13

Source: Univ. of Tennessee.

MPEG Contact: S. Jeannin, LEP.

Submission item: 8A

Short Description: Basic Ophthalmic Exam. English.

Associated Data: None.

Size or Duration: 26 min

ITEM V14

Source: Singapore Ministry of Education.

MPEG Contact: V.V. Vinod.

Submission item: 9BC

Short Description: Educational Video. "A way with waste" and "Animals have young"

Size or Duration : 16 min and 9 min

ITEM V15

Source: Spanish TV, RTVE.

MPEG Contact: P. Salembier, UPC.

Submission item: 45

Short Description: Clip of the video titled "Santillana del Mar" included in the TV serial "Los Pueblos. Espana a ras del suelo".

Size or Duration: 30 mins.

ITEM V16

Source: Spanish TV, RTVE.

MPEG Contact: P. Salembier, UPC.

Submission item: 40

Short Description: Three clips of (1) Soccer: Spain vs. Sweden; (2) Cycling: "Vuelta a Espana"; (3) BasketBall: Real Madrid ag. Estudiantes

Size or Duration: (1) 15 mins. (2) 19 min. (3) 15 min

ITEM V17

Source: Korean Broadcating Station (KBS).

MPEG Contact: S. Sull, Korea University.

Submission item: 47A

Short description: (1) Basketball game broadcast, korean comments; (2) Golf tournament broadcast

Size or Duration: (1) 10 minutes; (2) 10 minutes

ITEM V18

Source: Samsung.

MPEG Contact: Y. Choi, Samsung.

Submission item: 15

Short Description: Soccer sequence.

Size or Duration: 45 Min

 

ITEM V19

Source: Samsung.

MPEG Contact: Y. Choi, Samsung.

Submission item: 14

Short Description: 14 items of commercials in Korean

Associated Data: Full transcript may be given. Additional semantics.

Size or Duration: Each 30 sec.

ITEM V20

Source: Korean Broadcasting Station (KBS).

MPEG Contact: S. Sull, Korea University.

Submission item: 47B

Short description: Korea's pop singers' live music Show

Size or Duration: 10 mins

ITEM V21

Source: Spanish TV, RTVE.

MPEG Contact: P. Salembier, UPC.

Submission item: 43

Short Description: TV quiz program called "Saber y ganar"

Size or Duration: 15 mins.

ITEM V22

Source: Spanish TV, RTVE.

MPEG Contact: P. Salembier, UPC.

Submission item: 44

Short Description: "Musica si".

Size or Duration: 15 mins.

Format: MPEG1.

ITEM V23

Source: Portuguese TV, SIC.

MPEG Contact: F. Pereira, IST.

Submission item: 12C

Short Description Variety Show. First 30 minutes of complete program.

Associated Data: None.

Size or Duration: ~30 mins

ITEM V24

Source: D'Ocon Film Productions.

MPEG Contact: P. Salembier, UPC.

Submission item: 32

Short Description: Selected pieces of cartoons. (soundtrack in Spanish and English)

Size or Duration: 5 clips of 1-2 mins: 12 mins.

ITEM V25

Source: Singapore Ministry of Education.

MPEG Contact: V.V. Vinod.

Submission item: 9A

Short Description: Educational Video: "Harmony "

Size or Duration : 12 min

ITEM V26

Source: Spanish TV, RTVE.

MPEG Contact: P. Salembier, UPC.

Submission item: 38

Short Description: Cartoon "Don Quijote de la Mancha"

Size or Duration: 15 mins.

ITEM V27

Source: LGERCA

MPEG Contact: N. Nandhakumar, LGERCA

Submission item: 22

Short Description: Edited home video.

Associated Data: Ground truth information: location of scene cuts, pan, zoom, etc.

Size or Duration: 17 mins each (approx), two files

ITEM V28

Source: ETRI

MPEG Contact: K.W. Lee, ETRI.

Submission item: 33

Short Description: Test material for surveillance. 3 outdoors scenes (different fixed cameras), people appearing / disappearing.

Associated Data: semantic info on what’s happening

Size or Duration: 9 mins

ITEM V29

Source: Universite Catholique de Louvain

MPEG Contact: S. Jeannin, LEP.

Submission item: 49

Short Description: 5 video sequences taken from a bridge over a speedway. The camera watches the vehicles flow. Different luminance conditions are present.

Associated Data: A text file describing the global recording conditions, the description of the events present in the 5 sequences, the translation of pixel coordinates to real world 3D coordinates.

Size or Duration: (1) 0'56"; (2) 8'47"; (3) 9'20"; (4) 4'59"; (5) 4'59"

ITEM V30

Source: Matsushita

MPEG Contact: Taka Senoh, Matsushita.

Submission item: 46

Short Description: Three "video object sequences" used as MPEG-4 test data: (1) Bream; (2) Children; (3) Fish

Associated Data: Each sequence consists of a background still picture, an object with shape information and a caption

Size or Duration : (1) 60MB – 10 secs

(2) 60MB– 10 secs

(3) 30MB. – 10 secs

Format: Raw SIF format

 

 

Annex I: Licensing Agreement for the MPEG-7 Content Set

 

This document establishes the conditions to abide in order to use the MPEG-7 content set.

 

The files here included are part of the MPEG-7 content set. When using or accessing these files, users agree with the following conditions:

  1. The use of MPEG-7 content set is authorized only for the purpose of the development of the MPEG-7 standard, notably for the evaluation, collaboration and verification phases.
  2. MPEG-7 content cannot be used for any commercial purposes, notably any type of broadcasting. Moreover the users are not allowed to make this content generally and freely available by any means, notably through the Internet, since this may contradict clause 1.
  3. All intellectual property rights for the MPEG-7 content set remain the property of the respective owners.
  4. The content owners are under no obligation or liability in respect of the fitness of the content for a particular purpose. The owners do not warrant that the sequences will meet user’s specific requirements.
  5. Copies of this content can only be made by MPEG members for the purposes mentioned in 1. and must always contain the usage conditions here defined by including a "Usage conditions" file present in the same directory of the content files.
  6. MPEG is allowed to make public the results (not the content itself) of tests and experiments made using the MPEG-7 content set.
  7. MPEG members may include only limited parts of the MPEG-7 content set in scientific communications (e.g. conferences, workshops) and publications, provided that no commercial purposes exist. Whenever this is done, the corresponding content owner must be explicitly acknowledged.
  8. Prior to any usage of MPEG-7 content, any person that disagrees with the present conditions, must not make any use of the MPEG-7 content and return the received support (CD-ROM, tape or disk) to its supplier. Moreover any copies that may have been made must be destroyed. In the event such a person has obtained the MPEG-7 content set or any of its parts through Internet access, such material shall be destroyed, and the provider of such material informed
  9. The MPEG-7 content has been made available for MPEG usage under the conditions above specified by the following organizations/individuals:

 

MPEG-7 Audio Content Set

 

Category

Item num.

Short description

Source

Duration (min)

Radio

A1

Radio news broadcast

Radio-France 98

600

Music

A2

"Two Ton Shoe" Rock album

Two Ton Shoe

39

A3

Bruckner's Te Deum, and Mozart's Requiem

A. Lindsay

72

A4

Original composition, a capella. Voice only

S. Lounis

5

Audio

A5

Short sequences of solo instrument and other sounds

SQAM CD

30

A6

Pop song based on an A-A-C motif

Juergen Herre

5

 

 

MPEG-7 Still images Content Set

 

Category

Item

num.

Short description

Source

~number of images

Color photo.

S1

Amateur Photos of landscapes, people.

Tristan Savatier

300

S2

Photographs

Alejandro Jaimes

300

S3

Images of varying scene and color content, many outdoor and indoor images

Department of Water Resources, California

2000

S4

Indoor images of people, under varying lighting conditions and angles

Heinrich Hertz Institute

400

Grey-scale photo.

S5

Photographs

JPEG 2000

250

S6

Collection of print and digitized pictures

National Archives at Maryland

250

S7

Photos from the beginning of the century

Portuguese Photo. Center

260

Trademark logo

S8

Trademark images captured by a scanner (B&W images)

Korean Industrial Property Office

3000

Aerial photo.

S9

Aerial photos, monochrome and colored (5000x5000 pixels)

UC Santa Barbara

30

3D Range image

S10

205 range images and 7 complete 3-D colour models

National Research Council

200

 

 

MPEG-7 Video Content Set

 

Category

Type

Item

Num.

Short description

Source

Duration (min)

News

Shot

V1

Two complete TV news programs

Portuguese TV, RTP & SIC

60

Sequence

V2

Universal newsreel collection. B&W video.

National Archives at Maryland

30

Program

V3

Daily TV news program

Spanish TV, RTVE

29

V4

Weekly TV news program

Spanish TV, RTVE

19

Drama / Movie

Sequence

V5

"Art" movie: Hallo

Christoph Rodatz, GMD

4

V6

Movie: "La sombra de un cipres es alargada"

Spanish TV, RTVE

15

V7

TV Drama series: "Pepa y Pepe"

Spanish TV, RTVE

15

Program

V8

Sitcom (1 and 2)

Portuguese TV, RTP & SIC

60

V9

Filmed theater: The playboy of the western world

INA

120

Documentary

Shot

V10

"Science Eye": Bridge construction

NHK

9

Sequence

V11

5 clips of scientific documentaries

SFRS

25

V12

Documentary about buildings

Lancaster Television

5

Program

V13

Basic Ophthalmic Exam

Univ. of Tennessee

26

V14

Educational video: "A way with waste" and "Animals have young"

Singapore Ministry of Education

25

V15

Documentary about a village: "Santillana del Mar"

Spanish TV, RTVE

30

Sport

Sequence

V16

3 Sport Clips: Soccer, Cycling, Basketball

Spanish TV, RTVE

49

V17

2 Sport clips: Basketball, Golf

Korean Broadcasting Station

20

Program

V18

Soccer sequence

Samsung

45

Commercial

Sequence

V19

14 items of commercials in Korean

Samsung

7

Music video and games

Sequence

V20

Korea's pop singers' live music Show

Korean Broadcasting Station

10

V21

TV quiz program: "Saber y ganar"

Spanish TV, RTVE

15

V22

Music program: "Musica si"

Spanish TV, RTVE

15

Program

V23

Variety Show. First 30 minutes of complete program

Portuguese TV, SIC

30

Cartoon and animation

Sequence

V24

5 Selected pieces of cartoons

D'Ocon Film Productions

12

Program

V25

Harmony

Singapore Ministry of Education

12

V26

Cartoon "Don Quijote de la Mancha"

Spanish TV, RTVE

15

Home video

Shots

V27

Edited home video

LGERCA

34

Surveillance

Shots

V28

3 outdoors scenes

ETRI

9

V29

5 video sequences taken from a bridge over a speedway

UCL

28

Miscellaneous

 

Shots

V30

MPEG4 test sequence (Bream, Children, Fish)

Matsushita

1

 

Annex J1: Descriptor evaluation output sheet: Feature evaluation

 

 

Evaluation group: Group No. and name of responsible person

 

 

Name of the feature:

 

 

Feature description: (few lines explanations)

 

 

 

Feature’s relevance to MPEG-7 Requirements: (few lines summary)

 

 

 

Global appreciation on relevance:

Highly relevant: Include now

Relevant

Low relevance

 

Annex J2: Descriptor evaluation output sheet: Proposal evaluation

 

 

Feature name:

 

Proposal ID:

 

Descriptor name:

 

Proposer name:

 

Evaluation group: Group No. and name of responsible person

 

 

 

Summary of the proposal: (a few lines)

 

 

 

Evaluation along criteria:

 

Evaluation facts

Conclusions

Effectiveness

 

 

Application domain

 

 

Expression efficiency

 

 

Processing efficiency (value calculation)

 

 

Processing efficiency (matching)

 

 

Scalability

 

 

Multi-level representation

 

 

Criteria Table

 

Content of the criteria table cells:

Evaluation facts should mention:

Conclusion should mention:

 

 

Summary of the evaluation:

 

 

 

 

 

 

0: could not be evaluated

1: this proposal is not interesting for MPEG-7

2: this proposal is interesting for MPEG-7 but requires significant amount of further work

3: this proposal is interesting for MPEG-7 but with a few changes

4: this proposal has some very good points and is a good candidate for XM in its category

5: this proposal is superior in its category and very strongly recommended

 

Eventual remarks: (points of importance, not covered above)

 

 

Annex J3: Descriptor evaluation output sheet: Evaluation conclusion

 

Feature name:

 

Evaluation group: Group No. and name of responsible person

 

 

List of evaluated proposals:

Number of proposals

(Proposal Ids, Proposers names)

 

 

 

Feature: name, short description and relevance (0-5)

 

 

Evaluation of proposals conclusions:

List of proposals with overall assessment (0-5), from 5 to 0.

Annex K1: Description Schemes proposal evaluation sheet

 

 

DS name:

 

Media / functionality:

 

Proposal ID:

 

Proposer name:

 

Evaluation group: Group No. and name of responsible person

 

 

 

Summary of the proposal: (a few lines)

 

 

 

Evaluation along criteria:

 

Evaluation facts

Conclusions

1. Effectiveness

 

 

2. Application domain

 

 

3. Comprehensiveness

 

 

4. Abstraction at Multiple Levels

 

 

5. Flexibility

 

 

6. Extensibility

 

 

7. Scalability

 

 

8. Simplicity

 

 

Criteria Table

 

 

Content of the criteria table cells:

Evaluation facts should mention:

Conclusions should mention:

 

 

Summary of the evaluation:

 

 

0: could not be evaluated

1: this proposal is not interesting for MPEG-7

2: this proposal is interesting for MPEG-7 but requires significant amount of further work

3: this proposal is interesting for MPEG-7 but with a few changes

4: this proposal has some very good points and is a good candidate for XM in its category

5: this proposal is superior in its category and very strongly recommended

 

Eventual remarks: (points of importance, not covered above)

 

 

Annex K2: Description Schemes evaluation conclusions

 

Media / functionality:

 

DS name:

 

Evaluation group: Group No. and name of responsible person

 

 

List of evaluated proposals:

Number of proposals

(Proposal Ids, Proposers names)

 

 

 

Evaluation of proposals conclusions:

List of proposals with global appreciation’s (0-5), from 5 to 0.

Annex L: DDL proposals evaluation output sheet

 

 

 

 

General Information

Proposal ID:……………………………….

DDL Name:………………………………..

Proposal Name:……………………………

Proposer:………………………………….

 

 

PART 1: Evaluation of the DDL proposal against the requirements

 

REQUIREMENT

FACTS

CONCLUSIONS

  1. Compositional capabilities: The DDL shall supply the ability to compose a DS from multiple DSs

 

 

  • Platform independence: The DDL shall be platform and application independent. This is required to make the representation of content as reusable as possible even on grounds of changing technology.
  •  

     

  • Grammar: The DDL shall follow a grammar, which is unambiguous, and allow easy parsing (interpretation) by computers.
  •  

     

  • Primitive data types: provide a set of primitive data types, e.g. text, integer, real, date, time/time index, version, etc.
  •  

     

  • Composite datatypes: The DDL must be able to succinctly describe composite datatypes that may arise from the processing of digital signals (e.g., histograms, graphs, rgb-values).
  •  

     

  • Multiple media types: The DDL must provide a mechanism to relate Ds to data of multiple media types of inherent structure, particularly audio, video, audio-visual presentations, the interface to textual description, and any combinations of these.
  •  

     

  • Partial instantiation: The DDL shall provide the capability to allow a DS to be partially instantiated by descriptors.
  •  

     

  • Mandatory instantiation: The DDL shall provide the capability to allow the mandatory instantiation of descriptors in a DS.
  •  

     

  • Unique identification: The DDL shall provide mechanisms to uniquely identify DSs and Ds so that they can be referred to unambiguously.
  •  

     

  • Distinct name spaces: The DDL shall provide support for distinct name-spaces. Note: Different domains use the same descriptor for different features or different purposes.
  •  

     

  • Transformational capabilities: The DDL shall allow the reuse, extension and inheritance of existing Ds and DSs.
  •  

     

  • Relationships within a DS and between DSs: The DDL provides the capability to express the following relationships between DSs and among elements of a DS and express the semantics of these relations
  • a) Spatial relations a) Spatial relations

    b) Temporal relations

    c) Structural relations

    d) Conceptual relations

     

     

  • Relationship between description and data: The DDL shall supply a rich model for links and/or references between one or several descriptions and the described data.
  •  

     

  • Intellectual Property Management: The DDL shall provide a mechanism for the expression of Intellectual Property Management and Protection (IPMP) for description schemes and descriptors.
  •  

     

  • Real time support: The DDL shall desirably provide features to support real time applications (database output like electronic program guides)
  •  

     

     

    PART 2: Ability to define other Ds, DSs, Descriptions

     

    Capability

    Fact

    Conclusion

    Ability to define DSs in the chosen application domain

     

     

    Ability to define Ds in the chosen application domain

     

     

    Ability to define Descriptions in the chosen application domain

     

     

    Ability to define DSs in another application domain

     

     

    Ability to define Ds in another application domain

     

     

    Ability to define Descriptions in another application domain

     

     

    Ability to define DSs in the chosen media type

     

     

    Ability to define Ds in the chosen media type

     

     

    Ability to define Descriptions in the chosen media type

     

     

    Ability to define DSs in another media type

     

     

    Ability to define Ds in another media type

     

     

    Ability to define Descriptions in another media type

     

     

     

     

    PART 3: Summary of the evaluation

     

     

     

     

     

     

     

     

    This proposal will be used in the construction of the DDL for the Experimental Model XM:

    YES, FULL PROPOSAL / YES, PARTS OF THE PROPOSAL / NONE

    This proposal will be chosen to participate in Core Experiments CE: YES/NO

    This proposal is not recommended for either the CE or XM: YES/NO

    Annex M1: Coding Schemes proposal evaluation sheet

     

     

    Coding scheme name:

     

    Proposal ID:

     

    Proposer name:

     

    Evaluation group: Group No. and name of responsible person

     

     

     

    Summary of the proposal: (a few lines)

     

     

     

     

     

    Evaluation facts

    Conclusions

    1. Compression efficiency

     

     

    2. Complexity

     

     

    3. Lossless compression

     

     

    4. Streaming capability

     

     

    5. Error resilience

     

     

    6. Universality

     

     

    Criteria Table

     

     

    Content of the criteria table cells:

    Evaluation facts should mention:

    Conclusions should mention:

     

     

    Summary of the evaluation:

     

     

     

     

     

     

    0: could not be evaluated

    1: this proposal is not interesting for MPEG-7

    2: this proposal is interesting for MPEG-7 but requires significant amount of further work

    3: this proposal is interesting for MPEG-7 but with a few changes

    4: this proposal has some very good points and is a good candidate for XM in its category

    5: this proposal is superior in its category and very strongly recommended

     

    Eventual remarks: (points of importance, not covered above)

     

     

    ANNEX M2: CODING SCHEMES EVALUATION CONCLUSION

     

     

    Coding Scheme name:

     

    Evaluation group: Group No. and name of responsible person

     

     

    List of evaluated proposals:

    Number of proposals

    (Proposal Ids, Proposers names)

     

     

     

    Evaluation of proposals conclusions:

    List of proposals with global appreciation’s (0-5), from 5 to 0.

    ANNEX N: UML a graphical notation for DS & D proposals.

     

    Introduction

    Universal Modelling Language (UML) [Rational1] was developed by Rational Systems Inc. to address the diversity of approaches to modelling object orientated systems. It combines features of several modelling notations that are described in more detail in the following this reference [UML1]. UML was originally intended to become a de-facto standard but has since been adopted by OMG. It has been used in recent CORBA specifications [OMG1] This document is intended to provide a brief overview of UML concepts, provide sufficient graphical notation to allow UML to be used as the graphical notation for responses to the MPEG-7 call for proposals. It’s use is to be restricted to providing a graphical notation for description schemes (DSs) and descriptors (Ds). This document is not intended to be a complete introduction to UML nor is it intended to prevent the use of UML symbols not introduced in this document. A complete introduction to UML can be found either in [UML 1] or [UML 2]. A more extensive introduction to UML can be found in [UML 3]

     

    Approach

     

    UML is a software development tool which builds on the substantial body of work in the software-engineering field addressing the design, development and management of large software projects [SOM 1] and as such provides several components, these are

     

    Views: These show different aspects of the system they are not graphs but abstractions that may consist of a number of diagrams.

    Diagrams: These are the graphs that describe the contents of a view.

    Model elements: The concepts used in the graphs are the model elements.

    General Mechanisms: These provide extra comments and information about a model element or expose more of its semantics.

    Views

    UML provides several views on a system a brief description of each is given below.

     

    Use case view: shows how external actors perceive the functionality of the system

     

    Logical view: Shows the functionality of the internals of the system in terms of its static structure and dynamic behaviour.

     

    Component view: Shows the structure of the component elements of program code.

     

    Concurrency view: A view showing the concurrency within the system. It addresses problems of concurrency and synchronisation.

     

    Deployment view: A view showing the deployment of the system onto physical devices such as computers.

     

     

    Diagrams

    UML provides several diagrams each with its own set diagrammatic conventions to allow a particular view of a system to be described these are:

     

    Use case diagrams: Shows external actors and how they are connected to the use cases a system provides.

     

    Class diagrams: Shows the interrelationships between the classes in the system.

     

    Object- diagrams: Shows the interrelationships between instances of the classes in the system.

     

    Sate diagrams: Show the states that objects of the class may have and events that cause a state change.

     

    Sequence diagrams: Shows dynamic collaboration between a number of objects in particular the messages sent between objects and their temporal interrelationships.

     

    Collaboration diagrams: Provide an alternative method for representing dynamic collaboration between objects that places more emphasis on the objects and their interrelationships.

     

    Activity diagrams: Show sequential flows of activities.

     

    Component diagrams: Shows the implementation modules of a system and their dependencies.

     

    Deployment diagrams: Show the physical deployment of a system.

     

    It is envisaged that only the class diagram and object diagram will be needed for responses to the call for proposals.

     

    Model Elements

     

    UML provides a substantial set of graphic symbols for use in models. It might be expected that the model of a system for MPEG-7 would need all or most of the model elements allowed within UML. In the introduction it is stated that the purpose of this document is to provide sufficient notational elements to allow respondents to the call for proposals to show the structure of their description schemes and the relationships of the description schemes to the descriptors. It is envisaged that the following set of model elements will be adequate for this purpose.

     

     

     

    Class

     

     

    Object


     

    Generalisation

    (Inheritance)

    Aggregation

    (composition)

     

    Composite Aggregation


    Constraint

     

     


    0 or more

     

    At least 1 or more


     

     

     

     

    General Mechanisms

    These provide a method to with which to add notes to a diagram

     

     

     


    Note

     

     

     

    Example: UML

    Representation of the Existing MPEG-7 Terminology

     




    A1.1Class Diagram

     


     



     

     

     


     

     

     

     

     

    A1.2 Object Diagram

     

    Here we are showing the composition of a description through the instantiation of descriptions schemes and descriptors and the relationship to the content.

     

     

     







     

     

     

     

     

     

     

     

     

     

     

     

     

     

     

     

     

    References

    [OMG 1] CORBA telecoms: Telecommunications Domain Specifications ver 1.0 June 1998, telecom-98-07-12

    http://www.omg.org/corba/ctfull.htm

    ftp://ftp.omg.org/pub/docs/formal/98-07-12.pdf

    ftp://ftp.omg.org/pub/docs/formal/98-07-12.ps

     

    [SOM 1] Software Engineering, 5th Edition, Ian Sommerville

     

    [UML 1] UML Toolkit, Hans-Erik Eriksson Magnus Penker, Wiley Computer Publishing 1998, ISBN: 0-471-19161-2

     

    [UML 2] UML Notation Guide version 1.1 September 1997 http://www.rational.com/uml

     

    [UML 3] UML Distilled, Martin Fowler, Addison Wesley Longman, ISBN 0-201-32563-2