INTERNATIONAL ORGANIZATION FOR STANDARDIZATION
ORGANISATION INTERNATIONAL NORMALIZATION
ISO/IEC JTC1/SC29/WG11
CODING OF MOVING PICTURES AND ASSOCIATED AUDIO
ISO/IEC JTC1/SC29/WG11/
MPEG 98/ Atlantic City
Title:
MPEG-7 Evaluation Process DocumentSource: Requirements
Status: Approved
Date: October 1998
TABLE OF CONTENT
1. INTRODUCTION *
1.1. General Context *
1.2. Intended Use of this Document *
1.3 Detailed Time Schedule and Contacts *
1.4 What is Called For *
2. EVALUATION OF DESCRIPTORS *
2.1 Evaluation Criteria *
2.2 Evaluation Procedures *
2.3 Form for Proposing Descriptor(s) *
2.4 Summary of Proposal's Content for Descriptors *
3. EVALUATION OF DESCRIPTION SCHEMES *
3.1 Evaluation Criteria *
3.2 Evaluation Procedures *
3.3 Form for Proposing Description Scheme(s) *
3.4 Summary of Proposal's Content for Description Schemes *
4. EVALUATION OF DDL *
4.1 Evaluation Criteria *
4.2 Evaluation Procedures *
4.3 Form for Proposing DDL *
4.4 Summary of Proposal's Content for DDL *
5. EVALUATION OF CODING SCHEMES FOR DESCRIPTIONS *
5.1 Evaluation Criteria *
5.2 Evaluation Procedures *
5.3 Content Data *
5.4 Form for Proposing Coding Schemes for Descriptions *
5.5 Summary of Proposal's Content for Coding Schemes for Descriptions *
6. DEVELOPING THE MPEG-7 STANDARD AFTER THE CFP *
6.1 The Evaluation of the Proposals *
6.1.1. The February meeting *
6.1.2. The March meeting *
6.2 The eXperimentation Model *
6.3 The Core Experiment (CE) Process *
7. MPEG-7 Content set *
Annex A: Distribution of MPEG-7 content set *
Annex B: MPEG-7 evaluation pre-registration form *
Annex C: MPEG-7 proposal format *
Annex D: Cover Page for each proposal *
Annex E: Form to fill out for Descriptors *
Annex F: Form to fill out for Description Schemes *
Annex G: Form to fill out for Coding Schemes *
Annex H: MPEG-7 content set *
1 MPEG-7 Audio Content Set *
2 MPEG-7 Still images Content Set *
3 MPEG-7 Video Content Set *
Detailed list of items: *
Annex I: Licensing Agreement for the MPEG-7 Content Set *
MPEG-7 Audio Content Set *
MPEG-7 Still images Content Set *
MPEG-7 Video Content Set *
Annex J1: Descriptor evaluation output sheet: Feature evaluation *
Annex J2: Descriptor evaluation output sheet: Proposal evaluation *
Annex J3: Descriptor evaluation output sheet: Evaluation conclusion *
Annex K1: Description Schemes proposal evaluation sheet *
Annex K2: Description Schemes evaluation conclusions *
Annex L: DDL proposals evaluation output sheet *
General Information *
PART 1: Evaluation of the DDL proposal against the requirements *
PART 2: Ability to define other Ds, DSs, Descriptions *
PART 3: Summary of the evaluation *
Annex M1: Coding Schemes proposal evaluation sheet *
ANNEX M2: CODING SCHEMES EVALUATION CONCLUSION *
ANNEX N: UML a graphical notation for DS & D proposals. *
The scope of the future MPEG-7 standard is described in the documents "MPEG-7 Context and Objectives", "MPEG-7 Requirements", and "MPEG-7 Applications." All documents are publicly available from the MPEG web site: http://drogo.cselt.stet.it/mpeg/.
As with all MPEG standards, the MPEG-7 Standard will be drafted in a collaborative process, following a Call For Proposals (CFP), merging and improving the relevant proposals. The proposal evaluation and development of the standard which will follow is conducted by MPEG experts.
Evaluation starts with assessing the scope and technical merits of each proposal. Then, an initial Experimentation Model (XM) is built from a first selection of parts of all proposals. This starts the collaborative phase of the standardization. During the collaborative phase, the XM is updated and improved in an iterative fashion till MPEG-7 reaches the Working Draft (WD) stage. Improvements to the XM are made through Core Experiments (CEs), which are prompted by the contribution of new elements for the standard. CEs will be defined to test the contributed tools within the framework of the XM according to well-defined test conditions and criteria. The goal is to develop the best possible XM.
Finally, those parts of the XM (or of the Working Draft) that correspond to the normative elements of MPEG-7 will be standardized.
1.2. Intended Use of this Document
Evaluation takes place at all stages of the MPEG-7 process, in order to select appropriate items for inclusion in the standard. This document however, focuses only on specifying the test and evaluation process for responses to the CFP. , The test and evaluation process will take place at the evaluation meeting in February 1999, before the collaborative phase starts. This document also presents the time schedule for the whole process, the content of the proposals, and other guidelines important to proposers.
This document is closely related to the Proposal Package Description (PPD) document, which introduces the general context of the call. The two documents together (PPD and Evaluation document) constitute what we call the Call For Proposal (CFP). People intending to answer the MPEG-7 Call for Proposal should find all the information they need regarding the mechanics of proposal evaluation in these two documents.
In addition, proposers should also read the "MPEG-7 Context and Objectives", "MPEG-7 Requirements," "MPEG-7 Applications" documents, and other documents listed in the Call for Proposals in order to gain a better understanding of the expected use and content of MPEG-7.
In case of any remaining questions about the current document, you can directly contact its editor (see 1.3).
1.3 Detailed Time Schedule and Contacts
As stated in the introduction, this document aims to specify the evaluation process that will be applied to proposals. The evaluation itself will be mainly conducted in February 1999. However, the entirety of evaluation-related issues will be conducted in a broader time frame. For sake of clarity, the time frame of the whole process is detailed as follows:
These achievements conclude the initial evaluation phase covered in this document.
Contacts:
All the relevant contacts for proposers are listed in the following:
(1) For any questions related to this document, you may contact the editor of this document:
Sylvie Jeannin
Philips Research USA
Tel: +1 914 945 64 46
sjn@philabs.research.philips.com
(2) Regarding the MPEG-7 content set, the contact person is:
Seungyup Paek
Columbia University New Media Technology Center - USA
(until November 1, 1998—afterwards, please consult http://www.cselt.it/mpeg/ for details.)
Note that the content set is described in Annex H, its distribution process is detailed in Annex A as well as in W2468, and its conditions of use are detailed in Annex I as well as in W2466.
(3) For any question about evaluation pre-registration, please contact Michael F. Vetter (See below). Note that the pre-registration form, presented in Annex B, must be filled out and sent by December 1st to:
Michael F. Vetter TASC 55 Walkers Brook Drive Reading, MA 01867-3297 USA Tel: +1 781-942-2000 Fax: +1 781-942-9507 Mfvetter@tasc.com |
Rob Koenen KPN Research PO Box 421 2260 Leidschendam The Netherlands Tel: +31 70 332 53 10 Fax: +31 70 332 55 67 r.h.koenen@research.kpn.com |
Cc : Leonardo Chiariglione Convenor WG11 CSELT Via G. Reiss Romoli, 274 10148 Torino, ITALY Tel.: +39 11 228 6120 Fax: +39 11 228 6299 leonardo.chiariglione@cselt.it |
(4) For the registration to the Lancaster meeting, the contact persons are:
Ed Hartley (e.hartley@lancaster.ac.uk), and Barbara Hickson
DMRG
Lancaster University
Lancs, LA1 4YR Tel: (44) 1524 593808
England Fax: (44) 1524 593608
Information about the Lancaster meeting and about how to register can be found in N2465. Note that there are facilities fees included in this registration.
(5) For specific matters about the procedure for proposal submission, please contact Michael F. Vetter or Rob Koenen (See above). Each proposal must be submitted by February 1st 1999 in electronic format, preferably Microsoft Word 6.0. Word 97 or Framemaker may be used in case Word 6.0. is inadequate for the document. An ftp site will be set up for these matters. All the needed related information will be provided after pre-registration.
In order to develop the MPEG-7 standard, MPEG calls for the following:
A. For the normative part:
While D, DS, DDL, and coding scheme proposals will be evaluated in February 1999 following the procedures defined in the current document, the proposals addressing MPEG-7 systems tools will not be part of the MPEG-7 Evaluation process in February 1999. These tools will be considered at the Seoul MPEG meeting. The proposers are kindly invited to present and show a demo regarding these tools at the Seoul MPEG meeting, in March 1999.
B. For the development of the standard, to be used in the XM and play an important role in the standardization process:
The proposals from group B are needed for the design and improvement of the XM. They will be discussed during the Lancaster meeting, and in the MPEG-7 Evaluation Ad Hoc Group(s) between Lancaster and Seoul. The conclusion of these discussions will be submitted as input to Seoul for the design of XM1.0. Whether and how these techniques will be formally evaluated will be decided after the XM1.0 is built.
It should be noted that it is not necessary for a proposal to address all of the elements listed above. It is possible for a proposal, for example, to only propose descriptors for some set of features, or description schemes or parts of description schemes. Selected aspects of different proposals will be incorporated into a common model (the XM) during the collaborative phase of the standard with the goal of building the best possible model.
The proposer should provide:
The feature captures important characteristic(s) of the AV material.
Gives better retrieval accuracy (e.g. precision, recall rate) with respect to other descriptors for the same feature.
The Descriptor is applicable to a wide range of application domains.
The Descriptor expresses the given feature(s) precisely, and completely.
The Descriptor represents the features at multiple levels of abstraction.
The Descriptors proposals will be evaluated by groups of experts drawn from participants at the evaluation meeting, following the evaluation procedure described as follows:
Proposals will be categorized according to the feature they represent.
Groups of MPEG experts will be formed to evaluate proposals by categories.
The evaluation will be based on the following steps:
Goal: Understand the feature, and assess how it captures important characteristic(s) of the AV material.
How: paper evaluation, from all paper descriptions of all proposals addressing this feature.
Who: experts
Output: feature evaluation sheet (see Annex J1), which includes represented features and their importance, with associated explanations.
Then, for each proposal independently within a given category:
Goal: The goal of this step is to have an initial assessment of the proposal based on the documentation included (questionnaire, summary and detailed structured description).
During this step, experts should prepare eventual questions to ask to the proposers at the next step, to clarify some points if needed.
How: Experts will review/analyze this information against requirements and evaluation criteria, first without being influenced by the participation of the proposers. A limited duration should be given to this step.
Who: Experts
Output: the proposal evaluation sheet (Annex J2, proposal summary and criteria table: all relevant criteria), with a short summary of the proposal, and its first evaluation along the analyzed criteria.
Goal: To enhance the understanding of the proposal by the experts.
The presentation shall demonstrate the appropriateness of the solution, and disclose the appropriate range of use. The demonstration will provide evidence of (some of) the functionalities claimed.
Who: experts and proposer(s) whose submission is evaluated.
How: Experts will interact with the proposer(s) through a presentation and possibly a demonstration.
Both demonstration and presentation will each have a 10 minutes time limit.
Output: Updated (modified/completed) proposal evaluation sheet (criteria table and eventually summary of the proposal).
Remarks:
If the proposed descriptor allows similarity based retrieval:
Goal: To refine the evaluation of the effectiveness and expression efficiency of the Descriptor.
Who: Experts
How: Proposers will provide
The quality of similarity based retrieval using the proposed descriptor will be evaluated using the MPEG-7 test sets sample inputs provided on-site, or samples provided by the proposer. The test will be run as follows:
This process will be repeated on several items of the test set, until the evaluators reach a consensus.
Remarks:
Output: Updated proposal evaluation sheet (criteria table, effectiveness and expression efficiency criteria only - similarity measure speed should not be taken into account).
For each proposal again:
5) Produce a global evaluation conclusion
Goal: To summarize the results of the previous steps to allow the selection of technologies for inclusion in the XM1.0, or in Core Experiments.
Who: Experts.
How: Experts arrive at a consensus.
Output: Evaluation summary of the proposal evaluation sheet (Annex J2) and evaluation conclusions sheet (see Annex J3), the final decision on this being taken as the Seoul meeting.
2.3 Form for Proposing Descriptor(s)
When proposing a Descriptor, each proposer has to fill the form enclosed in Annex E. Besides, proposers have to fill the proposal cover page (Annex D), and then refer to the proposal format guidelines given in Annex C.
2.4 Summary of Proposal's Content for Descriptors
The paper proposal, to be sent by February 1st, 1999, must include:
At the Lancaster meeting, proposers provide:
3. EVALUATION OF DESCRIPTION SCHEMES
The DS is applicable for a wide range of applications. "Applicable" means directly usable or usable as a component of a larger DS.
The DS provides an off the shelf solution for a given application domain. For this application domain, it takes into account relevant Descriptors and relevant relations between the Descriptors.
The DS can provide abstractions at multiple levels. An example is a hierarchical scheme where the base layer gives a coarse description and successive layers give more refined descriptions. The type of hierarchy used is appropriate for the purpose of the DS. Descriptors within the DS are amenable to being prioritized.
Part of the DS can be used effectively:
The DS is easily extensible to other DSs (in a way similar to inheritance in Object-Oriented Programming).
A minimal number of Descriptors and possible relationships are used to meet the needs of a particular application domain.
The evaluation procedure for Description Schemes is as follows:
The proposals are categorized according to the media addressed, applications addressed, functionality etc. Groups of MPEG experts will be formed to evaluate proposals by categories.
This phase relies on:
Both are essential to the complete evaluation of DS proposals. They are handled as described in the following.
The evaluation will be based on the following steps:
1) Evaluate the paper document.
Goal: The goal of this step is to have an initial assessment of the proposal based on the documentation included (questionnaire, summary and detailed structured description).
During this step, experts should prepare eventual questions to ask to the proposers at the next step, to clarify some points if needed.
How: Experts will review/analyze this information against requirements and evaluation criteria, first without being influenced by the participation of the proposers. A limited duration should be given to this step.
Who: Experts
Output: the proposal evaluation sheet (Annex K1, proposal summary and criteria table: all relevant criteria), with a short summary of the proposal, and its first evaluation along the analyzed criteria.
2) Hear presentation / see demonstration.
Goal: To enhance the understanding of the proposal by the experts.
The presentation shall demonstrate the appropriateness of the solution, and disclose the appropriate range of use. The demonstration will provide evidence of (some of) the functionalities claimed, and of how the DS satisfies the evaluation criteria.
Who: experts and proposer(s) whose submission is evaluated.
How: Experts will interact with the proposer(s) through a presentation and possibly a demonstration.
Both demonstration and presentation will each have a 10 minutes time limit.
Output: Updated (modified/completed) proposal evaluation sheet (criteria table and eventually summary of the proposal).
Remark: This step is not mandatory, but strongly encouraged. It gives an opportunity for the proposers to clarify, and provide additional evidence of facts that are not easy to demonstrate on paper.
3) Produce a global evaluation conclusion
Goal: To summarize the results of the previous steps. This should allow:
Who: Experts.
How: Experts arrive at a consensus.
Output: Evaluation summary of the proposal evaluation sheet (Annex K1) and evaluation conclusions sheet (see Annex K2), the final decision on this being taken as the Seoul meeting.
3.3 Form for Proposing Description Scheme(s)
When proposing a Description Scheme, each proposer has to fill the form enclosed in Annex F. Besides, proposers have to fill the proposal cover page (Annex D), and then refer to the proposal format guidelines given in Annex C.
3.4 Summary of Proposal's Content for Description Schemes
The paper proposal, to be sent by February 1st, 1999, must include:
At the Lancaster meeting, proposers provide:
a) Spatial relations
b) Temporal relations
c) Structural relations
d) Conceptual relations
The evaluation procedure for each Description Definition Language is as follows:
1) Evaluate the paper document.
Goal: The goal of this step is to have an assessment of the ability of the proposal to meet the requirements for the DDL described in the MPEG-7 requirements document (summarized above in section 4.1). During this step, experts should prepare eventual questions to ask to the proposers at the next step, to clarify some points if needed.
How: Experts will review/analyze this information against requirements and evaluation criteria, first without being influenced by the participation of the proposers. A limited duration should be given to this step.
Who: Experts.
Output: A copy of the form in Annex L part 1 will be completed, assessing how the DDL meets the requirements. .
2) Define a test set of Ds and DSs
Goal: The goal of this step is to select a set of DSs and Ds to be used in the following step of the evaluation.
How: A set of DSs and Ds will be identified from the DDL proposals by reviewing the proposal packages. The experts will ensure that the chosen DSs and Ds are of manageable complexity.
Who: Experts.
Output: A set of DSs and Ds to be used in the next step.
Goal: To enhance the understanding of the proposal by the experts.
The presentation shall demonstrate the appropriateness of the solution, and the ability to use their DDL beyond the provided examples.
How: Experts will interact with the proposer(s) through a presentation, which will have 10 minutes time limit.
The evaluators will ask the proposer to show that the DDL can be used to generate DSs and Ds and related descriptions contained in the test set, when different from the examples of the proposal.
Who: Experts and proposers
Output: The output will be the completion of the Annex L Part 2, which is designed to provide an assessment of the ability of each DDL to generate DSs, Ds and descriptions.
4) Produce a global evaluation conclusion
Goal: To summarize the results of the previous steps to allow the selection of DDLs or parts of DDLs for inclusion in the XM1.0, or in Core Experiments.
How: Experts arrive at a consensus.
Who: Experts.
Output: Evaluation summary of the proposal evaluation sheet (Annex L Parts 1 and 2) and evaluation conclusions sheet (Annex L part 3).
Remark: The evaluation procedure for DDL will formally end at the 48th Vancouver meeting. The Lancaster meeting will provide a summary of each DDL’s capability, and an assessment of its suitability for the XM DDL and for participating in the Core Experiments. It will be further discussed on the reflector between Lancaster and Seoul. The conclusion of these discussions will be submitted as input to Seoul. Then, it is recommended that a Core Experiment be conducted until the Vancouver meeting, by utilizing the recommended DDL’s to construct the XM1.0 DSs and Ds. The result of this Core Experiment will conclude the DDL evaluation phase: the collaborative phase starts.
No specific form has to be filled when submitting a DDL. Proposers only have to fill the proposal cover page (Annex D), and then refer to the proposal format guidelines given in Annex C.
4.4 Summary of Proposal's Content for DDL
The paper proposal, to be sent by February 1st, 1999, must include:
At the Lancaster meeting, proposers provide:
5. EVALUATION OF CODING SCHEMES FOR DESCRIPTIONS
Ability to losslessly compress descriptions instantiations.
It is possible to multiplex and stream the coded description.
The coded description shall be robust against transmission errors.
Ability to be applied to a wide range of descriptions
The main evaluation of proposals will be done based on the documentation included with the proposal (questionnaire, summary and detailed structured description). MPEG experts will review/analyze this information against requirements and evaluation criteria to identify promising techniques for the XM development. If the proposer has performed tests for bitstream verification of his/her coding scheme, either in combination with a D/DS/DDL or separate, a demo is highly encouraged. Furthermore, a short presentation should be given to help the MPEG experts better understand the proposed coding scheme and its possible extensibility and applicability to a wider range of descriptions.
If bitstream verification is performed as part of the proposal, it should be based on encoding descriptions extracted from MPEG-7 content set in the addressed categories (see section 7). If other content is used, it should be justified why it was not possible to use the official material.
5.4 Form for Proposing Coding Schemes for Descriptions
When proposing a Coding Scheme, each proposer has to fill the form enclosed in Annex G. Besides, proposers have to fill the proposal cover page (Annex D), and then refer to the proposal format guidelines given in Annex C.
5.5 Summary of Proposal's Content for Coding Schemes for Descriptions
The paper proposal, to be sent by February 1st, 1999, must include:
At the Lancaster meeting, proposers provide:
6. DEVELOPING THE MPEG-7 STANDARD AFTER THE CFP
6.1 The Evaluation of the Proposals
The main objective of the Call For Proposals it to lead to the best possible starting of the collaborative phase, which is recognized as one of the strengths of MPEG.
It is reasonable to expect that the competitive process, which culminates at the evaluation meeting, will produce several top performing, possibly competitor, elements. Each of these will have strengths compared to each other, and the goal of the ensuing collaborative process is to produce convergence to a single standard combining the best features of all proposals. During this process, opportunities for further improvements will arise, and these should be verified and incorporated. The method to accomplish this work is called the Core Experiment process and the basis on which it works is the eXperimentation Model (XM).
For each proposal to be evaluated, all information shall be provided following the guidelines of the CFP, by February 1st 1999, so that MPEG members can study them to prepare the evaluation meeting in advance. The evaluation meeting will be held in Lancaster University (UK), February 15-19. During this meeting, MPEG members will:
Proposers will be given the opportunity to present and clarify their proposal.
The quality and the clarity of the proposal are the responsibility of the proposers.
Finally, proposers are encouraged to be present at the February meeting to help with the evaluation.
The results of the February meeting, together with the work done on the eXperimentation Model structure and started in October 1998, form the basis to build the first version of the eXperimentation Model. Its first version will be finalized in March 1999, at the MPEG Seoul meeting. This meeting will also specify the first set of Core Experiments to be performed.
The XM is a "common experimentation framework" which is used for further development of the standard. It will have components for evaluating and improving the DDL, DSs and Ds. Moreover, its associated experimentation database(s) will be more complete than the test sets used for evaluation. An XM may have different implementations within the MPEG group.
As a result of both the collaborative work on XM structure started at the Dublin meeting in October 1998 and the evaluation phase, an XM will be built and its first version will be finalized at the MPEG Seoul meeting, March 15-19, 1999. XM will consequently include proposals and/or parts of proposals that were either made before CFP within the Ad Hoc group on XM definition, or given as answers to the CFP.
After the XM is established, new elements can be brought to MPEG-7 following a Core Experiment procedure. They will be then evaluated inside the XM so that in the final standard, if two elements accomplish similar thing(s) in similar conditions, only the best will be chosen.
6.3 The Core Experiment (CE) Process
The Core Experiments (CEs) will provide a more complete evaluation of Ds, DSs, DDLs, combinations of Ds, DSs handling combinations of Ds, etc within XM.
The heart of this process is for multiple, independent, directly comparable experiments to be performed to determine whether or not a proposed algorithmic technique or syntax element has merit. A Core Experiment has to be completely and uniquely defined, so the results are unambiguous. In addition to the specification of the element to be evaluated (syntax, related algorithmic techniques…), a Core Experiment also specifies the parameters to be used so that the results can be compared. Therefore, a Core Experiment may include one or more of the following elements: benchmark databases, associated 'truth', benchmark queries, objective and subjective evaluation.
A Core Experiment is proposed by one or more MPEG experts. It is accepted by consensus, providing that two or more independent experts agree to perform the experiment. Conventionally this is interpreted to mean that the independent experts work for different companies (different divisions or subsidiaries of a company are not considered independent).
Normally, a Core Experiment will be conducted between two successive MPEG meetings. If necessary, and again by consensus, a Core Experiment may be continued until a further MPEG meeting.
The outcome of a Core Experiment is determined by consensus. If the experiment can be evaluated subjectively, then an informal "test" is performed by presenting the results in the MPEG meeting. First, the independent results should evidently be equivalent. Second, they should indicate whether or not the technique has merit. Since this will at times be difficult to determine because there is some merit but it is marginal, the consensus may be that an evaluation of implementation complexity will be used to make the final decision.
In case the technology presented in a Core Experiment is selected for entering the XM, this can only become effective if some MPEG member agrees to donate the source code that fully implements this technology.
7. MPEG-7 Content set
MPEG-7 is providing a content set, made of still images, audio, and audio-visual data, to be used for evaluation. This content set is described in Annex H. Its distribution process is detailed in Annex A as well as in W2468, and its conditions of use are detailed in Annex I as well as in W2466.
Each proposer planning either to show a demonstration or to participate to the similarity based retrieval evaluation shall use appropriate subsets of this content set. It is the responsibility of the proposer to choose the items of the content set that he/she will use for demonstration or similarity-based retrieval. These elements should be relevant to the evaluated proposal (e.g. color images for a color Descriptor), and in sufficient number to allow evaluators to draw conclusions from what is shown. Moreover, note that the choice of subset must be justified in the proposal/presentation.
In case the MPEG-7 content set is insufficient for demonstrating his/her proposal, the proposer may use his/her own test set and provide justifications for the same. In this case the content set used must be made available to MPEG under the conditions mentioned in Annex I.
The evaluators will judge the rationale of the selection of the content set, and take it into account when drawing conclusions from any evaluation step using it.
Annex A: Distribution of MPEG-7 content set
The following is the procedure to obtain the MPEG-7 content set:
(C/N) + (Shipping charge for group n)
C is the total cost required for the duplication of the content set.
N is the total number of groups that have requested the content set by the November 1st deadline.
Once the charges for each group have been calculated, the primary contact will send an e-mail to all the groups that have requested the content set, and notify each group of their respective payment amounts. This will be done by November 3rd, 1998. At this time, An approximation of the cost for each group is as follows:
Approximate storage requirement of the content set: 20 CDs;
Approximate cost to produce a minimum of 250 copies of the content set: $15,640;
Approximate cost for each group, assuming that 100 groups request the content set: $160.
These approximations do not include shipping charges or the tax. The cost is subject to change.
Conversion Media c/o Jon Taylor
625 North Milwaukee St. Suite 220.
Milwaukee, Wisconsin 53202.
U.S.A.
Phone: +1 888 999 1760
Email: jontaylor@conversionmedia.com
The primary contact for the content set distribution process described in this document is:
Seungyup Paek
Columbia University New Media Technology Center - USA
syp@ctr.columbia.edu.
Annex B: MPEG-7 evaluation pre-registration form
All organizations intending to submit proposals to the MPEG-7 evaluation and development process are requested to complete this pre-registration form and submit it to
Michael F. Vetter TASC 55 Walkers Brook Drive Reading, MA 01867-3297 USA Tel: +1 781-942-2000 Fax: +1 781-942-9507 Mfvetter@tasc.com |
Rob Koenen KPN Research PO Box 421 2260 Leidschendam The Netherlands Tel: +31 70 332 53 10 Fax: +31 70 332 55 67 r.h.koenen@research.kpn.com |
Cc : Leonardo Chiariglione Convenor WG11 CSELT Via G. Reiss Romoli, 274 10148 Torino, ITALY Tel.: +39 11 228 6120 Fax: +39 11 228 6299 leonardo.chiariglione@cselt.it |
by the 1st December 1998.
Pre-registration of proposals is mandatory. For more information, please see the MPEG-7 Evaluation Document.
A pre-registration form must be submitted for each descriptor, description scheme, Description Definition Language (DDL), or coding scheme for descriptors submitted for evaluation. The same applies for the systems tools as well as for the non-normative tools which are useful for the development of the standard, such as extraction and search methods as well as evaluation and validation techniques.
I. Proponent Contact Information
Name:______________________________________________________________________
Company/Institution:__________________________________________________________
Address:____________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
Phone:______________________________________________________________________
Fax: _______________________________________________________________________
E-Mail:
____________________________________________________________________________________
II. Proposal Content – Normative Tools
If this pre-registration refers to a normative tool, please circle which item it refers to:
Evaluation Item |
|
Descriptor |
Still images |
|
Video |
|
Audio |
|
Synthetic visual content |
|
Synthetic audio content |
|
Others (please specify) |
Description Scheme |
Still images |
|
Video |
|
Audio |
|
Synthetic visual content |
|
Synthetic audio content |
|
Others (please specify) |
Description Definition Language (DDL) |
|
Coding Scheme for Descriptions |
|
Systems tool |
Note: Please note that systems tools will not be part of the MPEG-7 Evaluation process in February 1999. These tools will be considered at the Seoul MPEG meeting.
III. Proposal Content – Non-Normative Tools
If this pre-registration refers to a non-normative tool useful for the development of the standard, please circle which evaluation item it refers to:
Extraction methods |
Search Methods |
Evaluation and validation techniques |
IV. Participation in the Evaluation Ad Hoc Group Meeting
Yes No
Note: If ‘Yes’, don’t forget to register for the meeting.
Yes No
Note: No guarantees are given in advance regarding the equipment requested. Please contact the Evaluation Ad Hoc Group Meeting organizer for confirmation.
V. Additional information (circle one or both)
Yes No
If ‘Yes’, please state which:____________________________________________________
___________________________________________________________________________
___________________________________________________________________________
Signature:___________________________________________________________________
Date: ______________________________________________________________________
Annex C: MPEG-7 proposal format
1. For each proposal, proposers should provide a detailed paper describing the proposed technology, with emphasis on the components of the proposal that should be evaluated. The paper should:
2. For Descriptors, please provide the data needed to perform the similarity evaluation, if applicable, as stated in the current document. If the MPEG-7 test set is augmented with other material or a different test set is used, please state why the MPEG-7 test sets were not suitable. Also the data set used must be made available to MPEG on the same conditions as the MPEG-7 Test and Evaluation Material. (See Annex I).
Please give also:
2.1. Typical search / filtering tasks the proposal applies to.
2.2. Explanation of the similarity measure used, if any.
3. For Description Schemes the proposal must include a detailed explanation using English (and pseudo-code if required) and a graphical representation. It is strongly suggested that, to the extent possible, the graphical representation of the proposed description schemes follow the UML notations provided in Annex N. It is strongly encouraged that a presentation and demonstration be given to support the material contained in the written proposal. Provide a written description of any demonstration planned to accompany this proposal and detail the evaluation criteria that will be shown. If your DS demonstration depends on use of Descriptors, please provide a brief summary of these Descriptors (Descriptors proposed separately may simply be referenced by proposal ID). All demonstrations of DSs shall use the MPEG-7 Test Material (or subset consisting of all relevant material) unless the proposer can justify that other test material is necessary to demonstrate the DS.
4. For DDL, the submission must also include:
Note that for DDL, the description of how the proposal meets the MPEG-7 requirements should be made by completing a copy of annex L part 1.
Then, fill out the cover page (given in annex D) and the relevant form(s), given in annexes E, F, G and H, for each element (Descriptor, Description Scheme, Description Definition Language, Coding Scheme) in the proposal that MPEG should evaluate.
Note that a separate form must be completed for each Description Scheme and for each of the associated Descriptors if it is desired to have them separately considered/evaluated by MPEG.
Please note that brevity and clarity helps the evaluation.
Annex D: Cover Page for each proposal
One cover page is needed for each item (D, DS, DDL or coding scheme) of the proposal that you wish to be evaluated. For example, if you submit a DS that has 2 Ds and wish to have them all (the DS and 2 Ds) evaluated then you need to fill out three forms (one for DS, one for each D). The cover page must be filled out for each of these three forms.
Proposal Id: _____________
(Number obtained from pre-registration)
Name:_____________ _________________________________________
Company/Institution:_____ ______________________________________
1) Which item is proposed?
a) Descriptor
b) Description Scheme
c) Description Definition Language
d) Coding scheme
e) Others (please specify) ______________
Annex E: Form to fill out for Descriptors
1. Which media has your descriptor been developed for and/or applied to? (Choose one or more)
(a) Video |
(b) Audio |
(c) Synthetic video/images |
(d) Synthetic Audio |
(e) Still Images |
(f) Combined Audio-Visual |
(g) Cross-Modal |
(h) Others (please specify)______________ |
2. Which media do you use to extract your descriptor from? (Choose one or more)
(a) Video |
(b) Audio |
(c) Synthetic video/images |
(d) Synthetic Audio |
(e) Still Images |
(f) Combined Audio-Visual |
(g) Cross-Modal |
(h) Others (please specify)______________ |
3. Describe the feature your descriptor is associated to.
4. Discuss why the above feature is important to MPEG-7. (For example, tasks it can perform, queries it can answer, the MPEG-7 requirements it satisfies)
5. If your descriptor is content-domain-specific (e.g. biomedical images) please answer the following:
5.1. Which domain does your solution support?
5.2. Is there evidence of support from professional organizations for the technology proposed and/or from applicable current domain practices?
5.3. Is this applicable to other domains? Please list.
6. State up to three MPEG-7 applications (you may refer to the Applications Document) to which your proposal applies.
7. If you provide a similarity measure please answer the following:
7.1. Is it intended to match human perception of similarity? (yes or no)
7.2. Does it allow ranking? (yes or no)
7.3. Describe the complexity of the matching tool if possible (e.g., O(nlogn)).
8. If you provide an extraction method please answer the following:
8.1. Does the method generate a confidence measure? (yes or no)
8.2. Does it require human intervention? (yes or no)
8.3. If your method does not require human intervention please provide the time and memory complexity of the method if possible (e.g., O(n2)).
Annex F: Form to fill out for Description Schemes
1. Which media has your DS been developed for and/or applied to?
(a) Video |
(b) Audio |
(c) Synthetic video/images |
(d) Synthetic Audio |
(e) Still Images |
(f) Combined Audio-Visual |
(g) Cross-Modal |
(h) Others (please specify)______________ |
2. What is the main functionality of your DS?
3. Is your DS generic to different content domains or is it content domain-specific (e.g. biomedical, etc)? If your description scheme is content-domain-specific please answer the following:
3.1. Which domain and tasks does your solution supports?
3.2. Is there evidence of support from professional organizations for the technology proposed and/or from applicable domain current practices?
3.3. Is this applicable to other domains? Do you have evidence of this?
4. State up to three MPEG-7 applications (for examples you may refer MPEG-7 Applications Document) to which your proposal applies.
5. Will you provide a presentation to explain your DS proposal and answer questions about it?
6. Will you provide a demonstration to show how your DS meets the evaluation criteria? Are you using the MPEG-7 test sets or are you providing your own test sets?
Annex G: Form to fill out for Coding Schemes
(a) DDL |
(b) Description scheme(s) |
(c) Descriptor(s) |
(d) Descriptor data |
(e) Others (please specify) ______________ |
If yes, which media has your coding scheme been developed for and/or applied to?
(a) Video |
(b) Audio |
(c) Synthetic video/images |
(d) Synthetic Audio |
(e) Still Images |
(f) Combined Audio-Visual |
(g) Cross-Modal |
(h) Others (please specify)______________ |
3. If your coding scheme is application-domain-specific please answer the following:
3.1. Which domain and tasks does your solution support?
If so:
5.1. Which?
5.2. Would it be applicable to other Ds/DSs/DDLs? (yes or no)
5.3. If yes, how severe would the modifications be, and would this increase complexity or decrease efficiency?
6. Are you providing a demonstration?
The MPEG-7 Content set is composed of three main categories: Audio (about 12 hours), Still images (about 7000 images) and Video (about 13 hours). The items in each category are described in the following sections.
For supplemental content, please refer to the Video Content Set, especially the music and entertainment category. It contains much Audio-visual material, Also of potential interest are items with monologues and dialogues in diverse languages.
Category |
Item num. |
Short description |
Source |
Duration (min) |
Radio |
A1 |
Radio news broadcast |
Radio-France 98 |
600 |
Music |
A2 |
"Two Ton Shoe" Rock album |
Two Ton Shoe |
39 |
A3 |
Bruckner's Te Deum, and Mozart's Requiem |
A. Lindsay |
72 |
|
A4 |
Original composition, a capella. Voice only |
S. Lounis |
5 |
|
Audio |
A5 |
Short sequences of solo instrument and other sounds |
SQAM CD |
30 |
A6 |
Pop song based on an A-A-C motif |
Juergen Herre |
5 |
2 MPEG-7 Still images Content Set
Category |
Item num. |
Short description |
Source |
~number of images |
Color photo. |
S1 |
Amateur Photos of landscapes, people. |
Tristan Savatier |
300 |
S2 |
Photographs |
Alejandro Jaimes |
300 |
|
S3 |
Images of varying scene and color content, many outdoor and indoor images |
Department of Water Resources, California |
2000 |
|
S4 |
Indoor images of people, under varying lighting conditions and angles |
Heinrich Hertz Institute |
400 |
|
Grey-scale photo. |
S5 |
Photographs |
JPEG 2000 |
250 |
S6 |
Collection of print and digitized pictures |
National Archives at Maryland |
250 |
|
S7 |
Photos from the beginning of the century |
Portuguese Photo. Center |
260 |
|
Trademark logo |
S8 |
Trademark images captured by a scanner (B&W images) |
Korean Industrial Property Office |
3000 |
Aerial photo. |
S9 |
Aerial photos, monochrome and colored (5000x5000 pixels) |
UC Santa Barbara |
30 |
3D Range image |
S10 |
205 range images and 7 complete 3-D colour models |
National Research Council |
200 |
All items have been assigned to one of the three possible types: "Program", "Sequence" or "Shot". A "Program" is a piece of material of rather long duration which most of the time includes both the beginning and the end of the program. A "sequence" is a clip of a few minutes that has been extracted from a program. Finally, the "shot" type corresponds to either "program" or "sequences" where the list of editing effects (shot cut, fade, etc.) will be available. Finally, note that all video material is encoded in MPEG-1 format, except item V30.
Category |
Type |
Item Num. |
Short description |
Source |
Duration (min) |
News |
Shot |
V1 |
Two complete TV news programs |
Portuguese TV, RTP & SIC |
60 |
Sequence |
V2 |
Universal newsreel collection. B&W video. |
National Archives at Maryland |
30 |
|
Program |
V3 |
Daily TV news program |
Spanish TV, RTVE |
29 |
|
V4 |
Weekly TV news program |
Spanish TV, RTVE |
19 |
||
Drama / Movie |
Sequence |
V5 |
"Art" movie: Hallo |
Christoph Rodatz, GMD |
4 |
V6 |
Movie: "La sombra de un cipres es alargada" |
Spanish TV, RTVE |
15 |
||
V7 |
TV Drama series: "Pepa y Pepe" |
Spanish TV, RTVE |
15 |
||
Program |
V8 |
Sitcom (1 and 2) |
Portuguese TV, RTP & SIC |
60 |
|
V9 |
Filmed theater: The playboy of the western world |
INA |
120 |
||
Documentary |
Shot |
V10 |
"Science Eye": Bridge construction |
NHK |
9 |
Sequence |
V11 |
5 clips of scientific documentaries |
SFRS |
25 |
|
V12 |
Documentary about buildings |
Lancaster Television |
5 |
||
Program |
V13 |
Basic Ophthalmic Exam |
Univ. of Tennessee |
26 |
|
V14 |
Educational video: "A way with waste" and "Animals have young" |
Singapore Ministry of Education |
25 |
||
V15 |
Documentary about a village: "Santillana del Mar" |
Spanish TV, RTVE |
30 |
||
Sport |
Sequence |
V16 |
3 Sport Clips: Soccer, Cycling, Basketball |
Spanish TV, RTVE |
49 |
V17 |
2 Sport clips: Basketball, Golf |
Korean Broadcasting Station |
20 |
||
Program |
V18 |
Soccer sequence |
Samsung |
45 |
|
Commercial |
Sequence |
V19 |
14 items of commercials in Korean |
Samsung |
7 |
Music video and games |
Sequence |
V20 |
Korea's pop singers' live music Show |
Korean Broadcasting Station |
10 |
V21 |
TV quiz program: "Saber y ganar" |
Spanish TV, RTVE |
15 |
||
V22 |
Music program: "Musica si" |
Spanish TV, RTVE |
15 |
||
Program |
V23 |
Variety Show. First 30 minutes of complete program |
Portuguese TV, SIC |
30 |
|
Cartoon and animation |
Sequence |
V24 |
5 Selected pieces of cartoons |
D'Ocon Film Productions |
12 |
Program |
V25 |
Harmony |
Singapore Ministry of Education |
12 |
|
V26 |
Cartoon "Don Quijote de la Mancha" |
Spanish TV, RTVE |
15 |
||
Home video |
Shots |
V27 |
Edited home video |
LGERCA |
34 |
Surveillance |
Shots |
V28 |
3 outdoors scenes |
ETRI |
9 |
V29 |
5 video sequences taken from a bridge over a speedway |
UCL |
28 |
||
Miscellaneous
|
Shots |
V30 |
MPEG4 test sequence (Bream, Children, Fish) |
Matsushita |
1 |
Audio Items:
ITEM A1 Category: Radio broadcast news. Source: 98 Radio-France. MPEG contact: R. Ronfard, INA. Submission Item: 10 Description: One day from a special program broadcast from June 9th to July 13th, 1998. 17 hours a day, from 8:30 AM to 1:30 AM. Includes:
Associated Data: Transmission control lists have been input in an Excel database. Size / Duration: 10 hours have been selected from an original material of 17 hours Format: Audio (mono) compression in MPEG-2, level 2, at 128 kbps, programs are segmented in files of 15 minutes |
ITEM A2 Content Category: Music Source: Two Ton Shoe MPEG Contact: A. Lindsay, Riverland. Submission Item: 18 Short Description: "Two Ton Shoe" Rock album. Associated Data: None. Size or Duration: 38:18 Format: 44100Hz, 16-bit AIFF
|
ITEM A3 Content Category: Music Source: A. Lindsay MPEG contact: A. Lindsay, Riverland. Submission Item: 19 Short Description: Brussels Choral Society singing Bruckner's Te Deum, and Mozart's Requiem. Associated Data: None. Size or Duration: 72:03 Format: 44100Hz, 16-bit AIFF |
ITEM A4 Content Category: Music. Source: S. Lounis MPEG contact: A. Lindsay, Riverland. Submission Item: 20 Short Description: Original composition, a capella. Voice only. Associated Data: French text. Size or Duration: 5 min Format: 48000Hz, 16-bit AIFF
|
ITEM A5 Content Category: Audio. Source: SQAM CD. MPEG contact: J. Herre, EBU/FhG. Submission item: 35 Short Description: Short sequences of solo instrument and other sounds Associated Data: / Size or Duration: 54 items (some tracks have sub-indices). No sequences are exceeding 640MB in length |
ITEM A6 Content Category: Audio. Source: J. Herre MPEG contact: J. Herre, EBU/FhG. Submission item: 52 Short Description: Pop song based on an A-A-C motif Associated Data: / Size or Duration: One song (<5 min) in a variety of formats Format: MIDI, AIFF, MPEG Layer 2 Audio
|
Still image items:
ITEM S1 Content Category: Amateur photographs. Source: Tristan Savatier. MPEG contact: R. Ronfard, INA. Submission item: 24 Short Description: Amateur Photos of landscapes, people. Some nudity. Associated Data: Keywords May be provided on request for each picture. Size or Duration: 300 photographs Format: PhotoCD |
ITEM S2 Content Category: Photographs. Source: Alejandro Jaimes. MPEG contact: Seungyup Paek, Columbia Univ. Submission item: 48 Short Description: General. Associated Data: None Size or Duration: ~300 images. Format: JPEG
|
ITEM S3 Content Category: Photographs. Source: Department of Water Resources, California MPEG contact: S. Krishnamachari Submission item: 50 Short Description: Images of varying scene and color content, many outdoor and indoor images. Images of same scene with different viewpoints. Many "similar" images that might come handy during evaluation. Associated Data: None Size or Duration: ~ 2000 images Format: JPEG |
ITEM S4 Content Category: Photographs. Source: Heinrich Hertz Institute MPEG contact: Sylvie Jeannin, LEP. Submission item: 53 Short Description: Indoor images of people, under varying lighting conditions and angles Associated Data: None
|
ITEM S5 Content Category: Photographs. Source: JPEG 2000. MPEG contact: M. Zeug, Iterated. Submission item: 26 Short Description: Test images. |
ITEM S6 Content Category: Photographs. Source: National Archives at Maryland. http://www.nara.gov/nara/menus/audvis.html MPEG contact: N. Nandhakumar, LG. Submission item: 28 Short Description: Collection of print and digitized pictures. |
ITEM S7 Content Category: Photographs. Source: Portuguese Photograph Center. MPEG contact: J. Torres, INESCN. Submission item: 37 Short Description: Photos by Portuguese photographer from the beginning of the century. One of the pioneers of the cinema in Portugal. Some collections of photos are like a film and describe some sequence. Size or Duration: 260 images. <300 MB Format: JPEG |
ITEM S8 Content Category: Subset of registered trademark images. Source: Korean Industrial Property Office. MPEG contact: M. Kim, ETRI. Submission number: 34 Short Description: All images are trademarks captured by a scanner. Size or Duration: 3000 images. 2.6 Mbytes. Format: TIFF
|
ITEM S9 Content Category: Aerial photographs. Source: MPEG contact: B.S. Manjunath, UC Santa Barbara. Submission item: 25 Short Description: Aerial photos, monochrome and colored. Associated Data: Associated data available for some of the images Size or Duration: 30 photos (each 5000x5000 pixels) Format: tiff |
ITEM S10 Content Category: Range images. Source: National Research Council. MPEG contact: E. Paquet. Submission item: 1 Short Description: These images can be used not only as range images but also as generators for 2D images (e.g. same object under different lighting conditions) Size or Duration : 205 range images and 7 complete 3-D colour models (all scanned) Format: VRML 1 |
Video items:
ITEM V1 Source: RTP & SIC MPEG Contact: F. Pereira, IST. Submission item: 12A Short Description: Two daily News program Associated Data: List and timing of shots and semantics. Size or Duration: 2 * 30 mins |
ITEM V2 Source: National Archives at Maryland. http://www.nara.gov/nara/menus/audvis.html MPEG Contact: N. Nandhakumar, LGERCA Submission item: 30 Short Description: Universal newsreel collection. B&W video, archive of news. Size or Duration: 30 mins |
ITEM V3 Source: Spanish TV, RTVE. MPEG contact: P. Salembier, UPC. Submission item: 39A Short Description: Clip from daily program "Telediario" Size or Duration: 29 mins |
ITEM V4 Source: Spanish TV, RTVE. MPEG contact: P. Salembier, UPC. Submission item: 39B Short Description: Clip from weekly program "Informe Semanal" Size or Duration: 19 min |
ITEM V5 Source: Christoph Rodatz, GMD. MPEG Contact: F. Nack. Submission item: 2B Short Description: Hallo Associated Data: German text available, translation possible. Some composition. Size or Duration: 4 min |
ITEM V6 Source: Spanish TV, RTVE. MPEG contact: P. Salembier, UPC. Submission item: 41 Short Description: "La sombra de un cipres es alargada". Associated Data: / Size or Duration: 15 mins. |
ITEM V7 Source: Spanish TV, RTVE. MPEG Contact: P. Salembier, UPC. Submission item: 42 Short Description: Drama series titled "Pepa y Pepe". Size or Duration: 15 mins. |
ITEM V8 Source: Portuguese TV, RTP & SIC. MPEG Contact: F. Pereira, IST. Submission item: 12B Short Description: 2 Sitcoms Associated Data: None. Size or Duration: 2*30 mins |
ITEM V9 Source: INA. MPEG contact: R. Ronfard, INA. Submission item: 16 Short Description: The playboy of the western world, filmed for TV. French. Associated Data: Text available Size or Duration: Two hours. |
ITEM V10 Source: NHK. MPEG Contact: M. Shibata, NHK. Submission item: 4 Short Description: "Science Eye" Scientific Video about bridge construction. Associated Data: Cuts, Japanese text overlays, narration text attached. Text descriptions (English) of cut-content will be attached. Size or Duration : 9 min |
ITEM V11 Source: SFRS. MPEG Contact: P. Faudemay Submission item: 5 Short Description: 5 sequences from scientific movies (1) (1974) Lascaux pictures. French and English; (2) (1990) An island with flamingoes. French, English and Spanish; (3) (1979) Social life of Antelope. French; (4) (1973) Study on the Nyamgatoms tribe. French; (5) (1964) Scientific missions in the Kerguelen islands. French. Associated Data: 3-page script for each movie. Size or Duration: 5 * 5 min |
ITEM V12 Source: Lancaster Television. MPEG Contact: E. Hartley, Lancaster Univ. Submission item: 7 Short Description: Show reel Associated Data: None.
|
ITEM V13 Source: Univ. of Tennessee. MPEG Contact: S. Jeannin, LEP. Submission item: 8A Short Description: Basic Ophthalmic Exam. English. Associated Data: None. Size or Duration: 26 min |
ITEM V14 Source: Singapore Ministry of Education. MPEG Contact: V.V. Vinod. Submission item: 9BC Short Description: Educational Video. "A way with waste" and "Animals have young" Size or Duration : 16 min and 9 min |
ITEM V15 Source: Spanish TV, RTVE. MPEG Contact: P. Salembier, UPC. Submission item: 45 Short Description: Clip of the video titled "Santillana del Mar" included in the TV serial "Los Pueblos. Espana a ras del suelo". Size or Duration: 30 mins. |
ITEM V16 Source: Spanish TV, RTVE. MPEG Contact: P. Salembier, UPC. Submission item: 40 Short Description: Three clips of (1) Soccer: Spain vs. Sweden; (2) Cycling: "Vuelta a Espana"; (3) BasketBall: Real Madrid ag. Estudiantes Size or Duration: (1) 15 mins. (2) 19 min. (3) 15 min |
ITEM V17 Source: Korean Broadcating Station (KBS). MPEG Contact: S. Sull, Korea University. Submission item: 47A Short description: (1) Basketball game broadcast, korean comments; (2) Golf tournament broadcast Size or Duration: (1) 10 minutes; (2) 10 minutes |
ITEM V18 Source: Samsung. MPEG Contact: Y. Choi, Samsung. Submission item: 15 Short Description: Soccer sequence. Size or Duration: 45 Min
|
ITEM V19 Source: Samsung. MPEG Contact: Y. Choi, Samsung. Submission item: 14 Short Description: 14 items of commercials in Korean Associated Data: Full transcript may be given. Additional semantics. Size or Duration: Each 30 sec. |
ITEM V20 Source: Korean Broadcasting Station (KBS). MPEG Contact: S. Sull, Korea University. Submission item: 47B Short description: Korea's pop singers' live music Show Size or Duration: 10 mins |
ITEM V21 Source: Spanish TV, RTVE. MPEG Contact: P. Salembier, UPC. Submission item: 43 Short Description: TV quiz program called "Saber y ganar" Size or Duration: 15 mins. |
ITEM V22 Source: Spanish TV, RTVE. MPEG Contact: P. Salembier, UPC. Submission item: 44 Short Description: "Musica si". Size or Duration: 15 mins. Format: MPEG1. |
ITEM V23 Source: Portuguese TV, SIC. MPEG Contact: F. Pereira, IST. Submission item: 12C Short Description Variety Show. First 30 minutes of complete program. Associated Data: None. Size or Duration: ~30 mins |
ITEM V24 Source: D'Ocon Film Productions. MPEG Contact: P. Salembier, UPC. Submission item: 32 Short Description: Selected pieces of cartoons. (soundtrack in Spanish and English) Size or Duration: 5 clips of 1-2 mins: 12 mins. |
ITEM V25 Source: Singapore Ministry of Education. MPEG Contact: V.V. Vinod. Submission item: 9A Short Description: Educational Video: "Harmony " Size or Duration : 12 min |
ITEM V26 Source: Spanish TV, RTVE. MPEG Contact: P. Salembier, UPC. Submission item: 38 Short Description: Cartoon "Don Quijote de la Mancha" Size or Duration: 15 mins. |
ITEM V27 Source: LGERCA MPEG Contact: N. Nandhakumar, LGERCA Submission item: 22 Short Description: Edited home video. Associated Data: Ground truth information: location of scene cuts, pan, zoom, etc. Size or Duration: 17 mins each (approx), two files |
ITEM V28 Source: ETRI MPEG Contact: K.W. Lee, ETRI. Submission item: 33 Short Description: Test material for surveillance. 3 outdoors scenes (different fixed cameras), people appearing / disappearing. Associated Data: semantic info on what’s happening Size or Duration: 9 mins |
ITEM V29 Source: Universite Catholique de Louvain MPEG Contact: S. Jeannin, LEP. Submission item: 49 Short Description: 5 video sequences taken from a bridge over a speedway. The camera watches the vehicles flow. Different luminance conditions are present. Associated Data: A text file describing the global recording conditions, the description of the events present in the 5 sequences, the translation of pixel coordinates to real world 3D coordinates. Size or Duration: (1) 0'56"; (2) 8'47"; (3) 9'20"; (4) 4'59"; (5) 4'59" |
ITEM V30 Source: Matsushita MPEG Contact: Taka Senoh, Matsushita. Submission item: 46 Short Description: Three "video object sequences" used as MPEG-4 test data: (1) Bream; (2) Children; (3) Fish Associated Data: Each sequence consists of a background still picture, an object with shape information and a caption Size or Duration : (1) 60MB – 10 secs (2) 60MB– 10 secs (3) 30MB. – 10 secs Format: Raw SIF format |
Annex I: Licensing Agreement for the MPEG-7 Content Set
This document establishes the conditions to abide in order to use the MPEG-7 content set.
The files here included are part of the MPEG-7 content set. When using or accessing these files, users agree with the following conditions:
Category |
Item num. |
Short description |
Source |
Duration (min) |
Radio |
A1 |
Radio news broadcast |
Radio-France 98 |
600 |
Music |
A2 |
"Two Ton Shoe" Rock album |
Two Ton Shoe |
39 |
A3 |
Bruckner's Te Deum, and Mozart's Requiem |
A. Lindsay |
72 |
|
A4 |
Original composition, a capella. Voice only |
S. Lounis |
5 |
|
Audio |
A5 |
Short sequences of solo instrument and other sounds |
SQAM CD |
30 |
A6 |
Pop song based on an A-A-C motif |
Juergen Herre |
5 |
MPEG-7 Still images Content Set
Category |
Item num. |
Short description |
Source |
~number of images |
Color photo. |
S1 |
Amateur Photos of landscapes, people. |
Tristan Savatier |
300 |
S2 |
Photographs |
Alejandro Jaimes |
300 |
|
S3 |
Images of varying scene and color content, many outdoor and indoor images |
Department of Water Resources, California |
2000 |
|
S4 |
Indoor images of people, under varying lighting conditions and angles |
Heinrich Hertz Institute |
400 |
|
Grey-scale photo. |
S5 |
Photographs |
JPEG 2000 |
250 |
S6 |
Collection of print and digitized pictures |
National Archives at Maryland |
250 |
|
S7 |
Photos from the beginning of the century |
Portuguese Photo. Center |
260 |
|
Trademark logo |
S8 |
Trademark images captured by a scanner (B&W images) |
Korean Industrial Property Office |
3000 |
Aerial photo. |
S9 |
Aerial photos, monochrome and colored (5000x5000 pixels) |
UC Santa Barbara |
30 |
3D Range image |
S10 |
205 range images and 7 complete 3-D colour models |
National Research Council |
200 |
Category |
Type |
Item Num. |
Short description |
Source |
Duration (min) |
News |
Shot |
V1 |
Two complete TV news programs |
Portuguese TV, RTP & SIC |
60 |
Sequence |
V2 |
Universal newsreel collection. B&W video. |
National Archives at Maryland |
30 |
|
Program |
V3 |
Daily TV news program |
Spanish TV, RTVE |
29 |
|
V4 |
Weekly TV news program |
Spanish TV, RTVE |
19 |
||
Drama / Movie |
Sequence |
V5 |
"Art" movie: Hallo |
Christoph Rodatz, GMD |
4 |
V6 |
Movie: "La sombra de un cipres es alargada" |
Spanish TV, RTVE |
15 |
||
V7 |
TV Drama series: "Pepa y Pepe" |
Spanish TV, RTVE |
15 |
||
Program |
V8 |
Sitcom (1 and 2) |
Portuguese TV, RTP & SIC |
60 |
|
V9 |
Filmed theater: The playboy of the western world |
INA |
120 |
||
Documentary |
Shot |
V10 |
"Science Eye": Bridge construction |
NHK |
9 |
Sequence |
V11 |
5 clips of scientific documentaries |
SFRS |
25 |
|
V12 |
Documentary about buildings |
Lancaster Television |
5 |
||
Program |
V13 |
Basic Ophthalmic Exam |
Univ. of Tennessee |
26 |
|
V14 |
Educational video: "A way with waste" and "Animals have young" |
Singapore Ministry of Education |
25 |
||
V15 |
Documentary about a village: "Santillana del Mar" |
Spanish TV, RTVE |
30 |
||
Sport |
Sequence |
V16 |
3 Sport Clips: Soccer, Cycling, Basketball |
Spanish TV, RTVE |
49 |
V17 |
2 Sport clips: Basketball, Golf |
Korean Broadcasting Station |
20 |
||
Program |
V18 |
Soccer sequence |
Samsung |
45 |
|
Commercial |
Sequence |
V19 |
14 items of commercials in Korean |
Samsung |
7 |
Music video and games |
Sequence |
V20 |
Korea's pop singers' live music Show |
Korean Broadcasting Station |
10 |
V21 |
TV quiz program: "Saber y ganar" |
Spanish TV, RTVE |
15 |
||
V22 |
Music program: "Musica si" |
Spanish TV, RTVE |
15 |
||
Program |
V23 |
Variety Show. First 30 minutes of complete program |
Portuguese TV, SIC |
30 |
|
Cartoon and animation |
Sequence |
V24 |
5 Selected pieces of cartoons |
D'Ocon Film Productions |
12 |
Program |
V25 |
Harmony |
Singapore Ministry of Education |
12 |
|
V26 |
Cartoon "Don Quijote de la Mancha" |
Spanish TV, RTVE |
15 |
||
Home video |
Shots |
V27 |
Edited home video |
LGERCA |
34 |
Surveillance |
Shots |
V28 |
3 outdoors scenes |
ETRI |
9 |
V29 |
5 video sequences taken from a bridge over a speedway |
UCL |
28 |
||
Miscellaneous
|
Shots |
V30 |
MPEG4 test sequence (Bream, Children, Fish) |
Matsushita |
1 |
Annex J1: Descriptor evaluation output sheet: Feature evaluation
Evaluation group: Group No. and name of responsible person
Name of the feature:
Feature description: (few lines explanations)
Feature’s relevance to MPEG-7 Requirements: (few lines summary)
Global appreciation on relevance:
Highly relevant: Include now
Relevant
Low relevance
Annex J2: Descriptor evaluation output sheet: Proposal evaluation
Feature name:
Proposal ID:
Descriptor name:
Proposer name:
Evaluation group: Group No. and name of responsible person
Summary of the proposal: (a few lines)
Evaluation along criteria:
|
Evaluation facts |
Conclusions |
Effectiveness |
|
|
Application domain |
|
|
Expression efficiency |
|
|
Processing efficiency (value calculation) |
|
|
Processing efficiency (matching) |
|
|
Scalability |
|
|
Multi-level representation |
|
|
Criteria Table
Content of the criteria table cells:
Evaluation facts should mention:
Conclusion should mention:
Summary of the evaluation:
0: could not be evaluated
1: this proposal is not interesting for MPEG-7
2: this proposal is interesting for MPEG-7 but requires significant amount of further work
3: this proposal is interesting for MPEG-7 but with a few changes
4: this proposal has some very good points and is a good candidate for XM in its category
5: this proposal is superior in its category and very strongly recommended
Eventual remarks: (points of importance, not covered above)
Annex J3: Descriptor evaluation output sheet: Evaluation conclusion
Feature name:
Evaluation group: Group No. and name of responsible person
List of evaluated proposals:
Number of proposals
(Proposal Ids, Proposers names)
Feature: name, short description and relevance (0-5)
Evaluation of proposals conclusions:
List of proposals with overall assessment (0-5), from 5 to 0.
Annex K1: Description Schemes proposal evaluation sheet
DS name:
Media / functionality:
Proposal ID:
Proposer name:
Evaluation group: Group No. and name of responsible person
Summary of the proposal: (a few lines)
Evaluation along criteria:
|
Evaluation facts |
Conclusions |
1. Effectiveness |
|
|
2. Application domain |
|
|
3. Comprehensiveness |
|
|
4. Abstraction at Multiple Levels |
|
|
5. Flexibility |
|
|
6. Extensibility |
|
|
7. Scalability |
|
|
8. Simplicity |
|
|
Criteria Table
Content of the criteria table cells:
Evaluation facts should mention:
Conclusions should mention:
Summary of the evaluation:
0: could not be evaluated
1: this proposal is not interesting for MPEG-7
2: this proposal is interesting for MPEG-7 but requires significant amount of further work
3: this proposal is interesting for MPEG-7 but with a few changes
4: this proposal has some very good points and is a good candidate for XM in its category
5: this proposal is superior in its category and very strongly recommended
Eventual remarks: (points of importance, not covered above)
Annex K2: Description Schemes evaluation conclusions
Media / functionality:
DS name:
Evaluation group: Group No. and name of responsible person
List of evaluated proposals:
Number of proposals
(Proposal Ids, Proposers names)
Evaluation of proposals conclusions:
List of proposals with global appreciation’s (0-5), from 5 to 0.
Annex L: DDL proposals evaluation output sheet
Proposal ID:……………………………….
DDL Name:………………………………..
Proposal Name:……………………………
Proposer:………………………………….
PART 1: Evaluation of the DDL proposal against the requirements
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
a) Spatial relations a) Spatial relations b) Temporal relations c) Structural relations d) Conceptual relations |
|
|
|
|
|
|
|
|
|
|
|
PART 2: Ability to define other Ds, DSs, Descriptions
Fact |
Conclusion |
|
Ability to define DSs in the chosen application domain |
|
|
Ability to define Ds in the chosen application domain |
|
|
Ability to define Descriptions in the chosen application domain |
|
|
Ability to define DSs in another application domain |
|
|
Ability to define Ds in another application domain |
|
|
Ability to define Descriptions in another application domain |
|
|
Ability to define DSs in the chosen media type |
|
|
Ability to define Ds in the chosen media type |
|
|
Ability to define Descriptions in the chosen media type |
|
|
Ability to define DSs in another media type |
|
|
Ability to define Ds in another media type |
|
|
Ability to define Descriptions in another media type |
|
|
PART 3: Summary of the evaluation
This proposal will be used in the construction of the DDL for the Experimental Model XM:
YES, FULL PROPOSAL / YES, PARTS OF THE PROPOSAL / NONE
This proposal will be chosen to participate in Core Experiments CE: YES/NO
This proposal is not recommended for either the CE or XM: YES/NO
Annex M1: Coding Schemes proposal evaluation sheet
Coding scheme name:
Proposal ID:
Proposer name:
Evaluation group: Group No. and name of responsible person
Summary of the proposal: (a few lines)
|
Evaluation facts |
Conclusions |
1. Compression efficiency |
|
|
2. Complexity |
|
|
3. Lossless compression |
|
|
4. Streaming capability |
|
|
5. Error resilience |
|
|
6. Universality |
|
|
Criteria Table
Content of the criteria table cells:
Evaluation facts should mention:
Conclusions should mention:
Summary of the evaluation:
0: could not be evaluated
1: this proposal is not interesting for MPEG-7
2: this proposal is interesting for MPEG-7 but requires significant amount of further work
3: this proposal is interesting for MPEG-7 but with a few changes
4: this proposal has some very good points and is a good candidate for XM in its category
5: this proposal is superior in its category and very strongly recommended
Eventual remarks: (points of importance, not covered above)
ANNEX M2: CODING SCHEMES EVALUATION CONCLUSION
Coding Scheme name:
Evaluation group: Group No. and name of responsible person
List of evaluated proposals:
Number of proposals
(Proposal Ids, Proposers names)
Evaluation of proposals conclusions:
List of proposals with global appreciation’s (0-5), from 5 to 0.
ANNEX N: UML a graphical notation for DS & D proposals.
Introduction
Universal Modelling Language (UML) [Rational1] was developed by Rational Systems Inc. to address the diversity of approaches to modelling object orientated systems. It combines features of several modelling notations that are described in more detail in the following this reference [UML1]. UML was originally intended to become a de-facto standard but has since been adopted by OMG. It has been used in recent CORBA specifications [OMG1] This document is intended to provide a brief overview of UML concepts, provide sufficient graphical notation to allow UML to be used as the graphical notation for responses to the MPEG-7 call for proposals. It’s use is to be restricted to providing a graphical notation for description schemes (DSs) and descriptors (Ds). This document is not intended to be a complete introduction to UML nor is it intended to prevent the use of UML symbols not introduced in this document. A complete introduction to UML can be found either in [UML 1] or [UML 2]. A more extensive introduction to UML can be found in [UML 3]
Approach
UML is a software development tool which builds on the substantial body of work in the software-engineering field addressing the design, development and management of large software projects [SOM 1] and as such provides several components, these are
Views: These show different aspects of the system they are not graphs but abstractions that may consist of a number of diagrams.
Diagrams: These are the graphs that describe the contents of a view.
Model elements: The concepts used in the graphs are the model elements.
General Mechanisms: These provide extra comments and information about a model element or expose more of its semantics.
Views
UML provides several views on a system a brief description of each is given below.
Use case view
: shows how external actors perceive the functionality of the system
Logical view: Shows the functionality of the internals of the system in terms of its static structure and dynamic behaviour.
Component view: Shows the structure of the component elements of program code.
Concurrency view: A view showing the concurrency within the system. It addresses problems of concurrency and synchronisation.
Deployment view: A view showing the deployment of the system onto physical devices such as computers.
Diagrams
UML provides several diagrams each with its own set diagrammatic conventions to allow a particular view of a system to be described these are:
Use case diagrams
Class diagrams: Shows the interrelationships between the classes in the system.
Object- diagrams: Shows the interrelationships between instances of the classes in the system.
Sate diagrams: Show the states that objects of the class may have and events that cause a state change.
Sequence diagrams: Shows dynamic collaboration between a number of objects in particular the messages sent between objects and their temporal interrelationships.
Collaboration diagrams: Provide an alternative method for representing dynamic collaboration between objects that places more emphasis on the objects and their interrelationships.
Activity diagrams: Show sequential flows of activities.
Component diagrams: Shows the implementation modules of a system and their dependencies.
Deployment diagrams: Show the physical deployment of a system.
It is envisaged that only the class diagram and object diagram will be needed for responses to the call for proposals.
Model Elements
UML provides a substantial set of graphic symbols for use in models. It might be expected that the model of a system for MPEG-7 would need all or most of the model elements allowed within UML. In the introduction it is stated that the purpose of this document is to provide sufficient notational elements to allow respondents to the call for proposals to show the structure of their description schemes and the relationships of the description schemes to the descriptors. It is envisaged that the following set of model elements will be adequate for this purpose.
Class |
|
Object |
|
Generalisation (Inheritance) |
|
Aggregation (composition) |
|
Composite Aggregation |
|
Constraint |
|
0 or more |
|
At least 1 or more |
|
General Mechanisms
These provide a method to with which to add notes to a diagram
Note |
|
Example: UML
Representation of the Existing MPEG-7 Terminology
A1.1Class Diagram
A1.2 Object Diagram
Here we are showing the composition of a description through the instantiation of descriptions schemes and descriptors and the relationship to the content.
References
[OMG 1] CORBA telecoms: Telecommunications Domain Specifications ver 1.0 June 1998, telecom-98-07-12
http://www.omg.org/corba/ctfull.htm
ftp://ftp.omg.org/pub/docs/formal/98-07-12.pdf
ftp://ftp.omg.org/pub/docs/formal/98-07-12.ps
[SOM 1] Software Engineering, 5th Edition, Ian Sommerville
[UML 1] UML Toolkit, Hans-Erik Eriksson Magnus Penker, Wiley Computer Publishing 1998, ISBN: 0-471-19161-2
[UML 2] UML Notation Guide version 1.1 September 1997
http://www.rational.com/uml
[UML 3] UML Distilled, Martin Fowler, Addison Wesley Longman, ISBN 0-201-32563-2