A method of organizing streamed data includes storing streamed data in a file, identifying metadata applicable to a subset of the streamed data, and forming at least one group of one or more samples of the streamed data, each sample in a group having identical metadata content for a metadata type. The file may be in accordance with ISO base media file format. The storing streamed data in a file may include storing in a reception hint track. The at least one group may be indicated in a sample group description box. The metadata type may be indicated by a grouping type and grouping instance data, the grouping type specifying semantics of the grouping instance data and the metadata content. The metadata content may comprise a metadata payload and zero or more metadata payload extensions, where the metadata payload is included in a first structure and the zero or more metadata payload extensions are included in a second structure.
|
1. A method comprising:
causing streamed data to be stored in a file, wherein the file consists of media data and metadata enclosed separately, wherein causing the streamed data to be stored in the file includes storing in a reception hint track;
identifying metadata applicable to two or more samples of the streamed data;
causing at least one timed metadata track to be created based on the identified metadata, the at least one timed metadata track describing a referred media track and the reception hint track, wherein the hint track refers to samples comprising instructions for constructing packets for transmission over an indicated communication protocol, wherein the media track refers to samples formatted according to a media compression format;
forming at least one group from the two or more samples of the streamed data, each sample in a group having identical metadata content for a metadata type;
selecting each sample to group box associated with the reception hint track and the media track; and
finding a sample group description index of a particular reception hint sample or a media sample.
47. An apparatus, comprising:
means for causing streamed data to be stored in a file, wherein the file consists of media data and metadata enclosed separately, wherein causing the streamed data to be stored a file includes storing a reception hint track;
means for identifying metadata applicable to two or more samples of the streamed data;
means for causing at least one timed metadata track to be created based on the identified metadata, the at least one timed metadata track describing a referred media track and the reception hint track, wherein the hint track refers to samples comprising instructions for constructing packets for transmission over an indicated communication protocol, wherein the media track refers to samples formatted according to a media compression format;
means for forming at least one group from the two or more samples of the streamed data, each sample in a group having identical metadata content for a metadata type;
means for selecting each sample to group box associated with the reception hint track and the media track; and
means for finding a sample group description index of a particular reception hint sample or media sample.
32. An apparatus comprising a processor and a memory storing program code, the memory and program code being configured to, with the processor, cause the apparatus to at least:
cause streamed data to be stored in a file, wherein the file consists of media data and metadata enclosed separately, wherein causing the streamed data to be stored in a file includes storing in a reception hint track;
identifying metadata applicable to two or more samples of the streamed data;
cause at least one timed metadata track to be created based on the identified metadata, the at least one timed metadata track describing a referred media track and the reception hint track, wherein the hint track refers to samples comprising instructions for constructing packets for transmission over an indicated communication protocol, wherein the media track refers to samples formatted according to a media compression format;
form at least one group from the two or more samples of the streamed data, each sample in a group having identical metadata content for a metadata type;
select each sample to group box associated with the reception hint track and the media track; and
find a sample group description index of a particular reception hint sample or media sample.
17. A computer program product comprising a non-transitory computer-readable medium storing program code portions therein, the program code portions being configured to, upon execution, cause an apparatus to at least:
cause streamed data to be stored in a file, wherein the file consists of media data and metadata enclosed separately, wherein causing the streamed data to be stored in a file includes storing in a reception hint track;
identifying metadata applicable to two or more samples of the streamed data;
cause at least one timed metadata track to be created based on the identified metadata, the at least one timed metadata track describing a referred media track and the reception hint track, wherein the hint track refers to samples comprising instructions for constructing packets for transmission over an indicated communication protocol, wherein the media track refers to samples formatted according to a media compression format; and
form at least one group from the two or more samples of the streamed data, each sample in a group having identical metadata content for a metadata type;
select each sample to group box associated with the reception hint track and the media track; and
find a sample group description index of a particular reception hint sample or media sample.
2. The method of
3. The method of
4. The method of
5. The method of
6. The method of
7. The method of
8. The method of
9. The method of
11. The method of
identifying the group including samples from two or more tracks with metadata associated with all related tracks.
12. The method of
identifying the group including samples from two or more tracks with metadata associated with a master track.
13. The method of
14. The method of
15. The method of
16. The method of
18. The computer program product of
19. The computer program product of
20. The computer program product of
21. The computer program product of
22. The computer program product of
23. The computer program product of
24. The computer program product of
25. The computer program product of
26. The computer program product of
identify the group including samples from two or more tracks with metadata associated with all related tracks.
27. The computer program product of
identify the group including samples from two or more tracks with metadata associated with a master track.
28. The computer program product of
29. The computer program product of
30. The computer program product of
31. The computer program product of
33. The apparatus of
34. The apparatus of
35. The apparatus of
36. The apparatus of
37. The apparatus of
38. The apparatus of
39. The apparatus of
41. The apparatus of
42. The apparatus of
43. The apparatus of
44. The apparatus of
45. The apparatus of
46. The apparatus of
48. The apparatus of
49. The apparatus of
50. The apparatus of
51. The apparatus of
|
The present application claims priority to U.S. Provisional Application No. 60/992,057, filed Dec. 3, 2007, U.S. Provisional Application No. 60/971,193, filed Sep. 10, 2008, and U.S. Provisional Application No. 60/956,912, filed Aug. 20, 2007 all of which are incorporated herein by reference.
The present invention relates generally to the field of streamed multimedia data and, more specifically, to the organization of such multimedia data with metadata and/or indexes.
This section is intended to provide a background or context to the invention that is recited in the claims. The description herein may include concepts that could be pursued, but are not necessarily ones that have been previously conceived or pursued. Therefore, unless otherwise indicated herein, what is described in this section is not prior art to the description and claims in this application and is not admitted to be prior art by inclusion in this section.
The multimedia container file format is an important element in the chain of multimedia content production, manipulation, transmission and consumption. There are substantial differences between the coding format (a.k.a. elementary stream format) and the container file format. The coding format relates to the action of a specific coding algorithm that codes the content information into a bitstream. The container file format comprises means of organizing the generated bitstream in such way that it can be accessed for local decoding and playback, transferred as a file, or streamed, all utilizing a variety of storage and transport architectures. Furthermore, the file format can facilitate interchange and editing of the media as well as recording of received real-time streams to a file.
Available media file format standards include ISO base media file format (ISO/IEC 14496-12), MPEG-4 file format (ISO/IEC 14496-14, also known as the MP4 format), AVC file format (ISO/IEC 14496-15) and 3GPP file format (3GPP TS 26.244, also known as the 3GP format). Other formats are also currently in development.
It is desirable that such media file formats enable the use of metadata. In many applications, such as reception and storage of media streams, such as audio and video streams, it is desirable that the file format enable the use of metadata that is specific to segments of the content of the streamed data to facilitate organization and access to desired sections or segments of the streamed data.
One aspect of the invention relates to a method of organizing streamed data. The method includes storing streamed data in a file, identifying metadata applicable to a subset of the streamed data, and forming at least one group of one or more samples of the streamed data, each sample in a group having identical metadata content for a metadata type.
In one embodiment, the at least one group is identified in a file.
In one embodiment, the file is in accordance with ISO base media file format. The storing streamed data in a file may include storing in a reception hint track. The at least one group may be indicated in a sample group description box.
In one embodiment, the metadata type includes titles of segments of the streamed data.
In one embodiment, a group includes samples from two or more tracks. The method may further comprise identifying the group including samples from two or more tracks with metadata associated with all related tracks. Alternatively, the method may comprise identifying the group including samples from two or more tracks with metadata associated with a master track.
In one embodiment, the method further includes creating a timed metadata track for the streamed data. In one embodiment, the at least one group of one or more samples of the streamed data is indicated in a sample group description box for the timed metadata track through associating the at least one group of one or more samples of the streamed data with respective timed metadata samples in the timed metadata track.
In one embodiment, the timed metadata track associates timing metadata with one or more timelines. The timed metadata track may associate timing metadata with two or more timelines.
In another aspect, the invention relates to a computer program embodied on a computer-readable medium. The computer program is configured to provide a method comprising storing streamed data in a file, identifying metadata applicable to a subset of the streamed data, and forming at least one group of one or more samples of the streamed data, each sample in a group having identical metadata content for a metadata type.
In another aspect of the invention, a receiver adapted to receive streamed data includes a storage unit for storing streamed data in a file, and a processor. The processor is adapted to identify metadata applicable to a subset of the streamed data and form at least one group of one or more samples of the streamed data, each sample in a group having identical metadata content for a metadata type.
In various embodiments, the metadata type may be indicated by a grouping type and grouping instance data, the grouping type specifying semantics of the grouping instance data and the metadata content. The metadata content may comprise a metadata payload and zero or more metadata payload extensions, where the metadata payload is included in a first structure and the zero or more metadata payload extensions are included in a second structure.
These and other advantages and features of the invention, together with the organization and manner of operation thereof, will become apparent from the following detailed description when taken in conjunction with the accompanying drawings, wherein like elements have like numerals throughout the several drawings described below.
The coded media bitstream is transferred to a storage 120. The storage 120 may comprise any type of mass memory to store the coded media bitstream. The format of the coded media bitstream in the storage 120 may be an elementary self-contained bitstream format, or one or more coded media bitstreams may be encapsulated into a container file. Some systems operate “live”, i.e. omit storage and transfer coded media bitstream from the encoder 110 directly to the sender 130. The coded media bitstream is then transferred to the sender 130, also referred to as the server, on a need basis. The format used in the transmission may be an elementary self-contained bitstream format, a packet stream format, or one or more coded media bitstreams may be encapsulated into a container file. The encoder 110, the storage 120, and the server 130 may reside in the same physical device or they may be included in separate devices. The encoder 110 and server 130 may operate with live real-time content, in which case the coded media bitstream is typically not stored permanently, but rather buffered for small periods of time in the content encoder 110 and/or in the server 130 to smooth out variations in processing delay, transfer delay, and coded media bitrate.
The server 130 sends the coded media bitstream using a communication protocol stack. The stack may include but is not limited to Real-Time Transport Protocol (RTP), User Datagram Protocol (UDP), and Internet Protocol (IP). When the communication protocol stack is packet-oriented, the server 130 encapsulates the coded media bitstream into packets. For example, when RTP is used, the server 130 encapsulates the coded media bitstream into RTP packets according to an RTP payload format. Typically, each media type has a dedicated RTP payload format. It should be again noted that a system may contain more than one server 130, but for the sake of simplicity, the following description only considers one server 130.
The server 130 may or may not be connected to a gateway 140 through a communication network. The gateway 140 may perform different types of functions, such as translation of a packet stream according to one communication protocol stack to another communication protocol stack, merging and forking of data streams, and manipulation of data stream according to the downlink and/or receiver capabilities, such as controlling the bit rate of the forwarded stream according to prevailing downlink network conditions. Examples of gateways 140 include multipoint conference control units (MCUs), gateways between circuit-switched and packet-switched video telephony, Push-to-talk over Cellular (PoC) servers, IP encapsulators in digital video broadcasting-handheld (DVB-H) systems, or set-top boxes that forward broadcast transmissions locally to home wireless networks. When RTP is used, the gateway 140 is called an RTP mixer or an RTP translator and typically acts as an endpoint of an RTP connection.
The system includes one or more receivers 150, typically capable of receiving, de-modulating, and de-capsulating the transmitted signal into a coded media bitstream. The coded media bitstream is transferred to a recording storage 155. The recording storage 155 may comprise any type of mass memory to store the coded media bitstream. The recording storage 155 may alternatively or additively comprise computation memory, such as random access memory. The format of the coded media bitstream in the recording storage 155 may be an elementary self-contained bitstream format, or one or more coded media bitstreams may be encapsulated into a container file. If there are many coded media bitstreams, such as an audio stream and a video stream, associated with each other, a container file is typically used and the receiver 150 comprises or is attached to a container file generator producing a container file from input streams. Some systems operate “live,” i.e. omit the recording storage 155 and transfer coded media bitstream from the receiver 150 directly to the decoder 160. In some systems, only the most recent part of the recorded stream, e.g., the most recent 10-minute excerption of the recorded stream, is maintained in the recording storage 155, while any earlier recorded data is discarded from the recording storage 155.
The coded media bitstream is transferred from the recording storage 155 to the decoder 160. If there are many coded media bitstreams, such as an audio stream and a video stream, associated with each other and encapsulated into a container file, a file parser (not shown in
The codec media bitstream is typically processed further by a decoder 160, whose output is one or more uncompressed media streams. Finally, a renderer 170 may reproduce the uncompressed media streams with a loudspeaker or a display, for example. The receiver 150, recording storage 155, decoder 160, and renderer 170 may reside in the same physical device or they may be included in separate devices.
The Digital Video Broadcasting (DVB) organization is currently in the process of specifying the DVB File Format. The primary purpose of defining the DVB File Format is to ease content interoperability between implementations of DVB technologies, such as set-top boxes according to current (DVT-T, DVB-C, DVB-S) and future DVB standards, IP (Internet Protocol) television receivers, and mobile television receivers according to DVB-H and its future evolutions. The DVB File Format will allow exchange of recorded (read-only) media between devices from different manufacturers, exchange of content using USB mass memories or similar read/write devices, and shared access to common disk storage on a home network, as well as much other functionality.
The ISO file format is the basis for most current multimedia container file formats, generally referred to as the ISO family of file formats. The ISO base media file format may be the basis for the development of the DVB File Format as well.
Referring now to
According to ISO family of file formats, a file consists of media data and metadata that are enclosed in separate boxes, the media data (mdat) box 220 and the movie (moov) box 210, respectively. The movie box may contain one or more tracks, and each track resides in one track box 212, 214. A track can be one of the following types: media, hint or timed metadata. A media track refers to samples formatted according to a media compression format (and its encapsulation to the ISO base media file format). A hint track refers to hint samples, containing cookbook instructions for constructing packets for transmission over an indicated communication protocol. The cookbook instructions may contain guidance for packet header construction and include packet payload construction. In the packet payload construction, data residing in other tracks or items may be referenced (e.g., a reference may indicate which piece of data in a particular track or item is instructed to be copied into a packet during the packet construction process). A timed metadata track refers to samples describing referred media and/or hint samples. For the presentation one media type, typically one media track is selected.
ISO base media file format does not limit a presentation to be contained in one file, and it may be contained in several files. One file contains the metadata for the whole presentation. This file may also contain all the media data, whereupon the presentation is self-contained. The other files, if used, are not required to be formatted to ISO base media file format, are used to contain media data, and may also contain unused media data, or other information. The ISO base media file format concerns the structure of the presentation file only. The format of the media-data files is constrained the ISO base media file format or its derivative formats only in that the media-data in the media files must be formatted as specified in the ISO base media file format or its derivative formats.
Movie fragments can be used when recording content to ISO files in order to avoid losing data if a recording application crashes, runs out of disk, or some other incident happens. Without movie fragments, data loss may occur because the file format insists that all metadata (the Movie Box) be written in one contiguous area of the file. Furthermore, when recording a file, there may not be sufficient amount of RAM to buffer a Movie Box for the size of the storage available, and re-computing the contents of a Movie Box when the movie is closed is too slow. Moreover, movie fragments can enable simultaneous recording and playback of a file using a regular ISO file parser. Finally, smaller duration of initial buffering is required for progressive downloading (e.g., simultaneous reception and playback of a file, when movie fragments are used and the initial Movie Box is smaller compared to a file with the same media content but structured without movie fragments).
The movie fragment feature enables splitting of the metadata that conventionally would reside in the moov box 210 to multiple pieces, each corresponding to a certain period of time for a track. Thus, the movie fragment feature enables interleaving of file metadata and media data. Consequently, the size of the moov box 210 can be limited and the use cases mentioned above be realized.
The media samples for the movie fragments reside in an mdat box 220, as usual, if they are in the same file as the moov box. For the meta data of the movie fragments, however, a moof box is provided. It comprises the information for a certain duration of playback time that would previously have been in the moov box 210. The moov box 210 still represents a valid movie on its own, but in addition, it comprises an mvex box indicating that movie fragments will follow in the same file. The movie fragments extend the presentation that is associated to the moov box in time.
The metadata that can be included in the moof box is limited to a subset of the metadata that can be included in a moov box 210 and is coded differently in some cases. Details of the boxes that can be included in a moof box can be found from the ISO base media file format specifications ISO/IEC International Standard 14496-12, Second Edition, 2005-04-01, including Amendments 1 and 2, incorporated herein by reference in their entirety.
In addition to timed tracks, ISO files can contain any non-timed binary objects in a meta box, or “static” metadata. The meta box can reside at the top level of the file, within a movie box, and within a track box. At most one meta box may occur at each of the file level, movie level, or track level. The meta box is required to contain a ‘hdlr’ box indicating the structure or format of the “meta” box contents. The meta box may contain any number of binary items that can be referred and each one of them can be associated with a file name.
In order to support more than one meta box at any level of the hierarchy (file, movie, or track), a meta box container box (‘meco’) has been introduced in the ISO base media file format. The meta box container box can carry any number of additional meta boxes at any level of the hierarchy (file, move, or track). This allows, for example, the same meta-data to be presented in two different, alternative, meta-data systems. The meta box relation box (“mere”) enables describing how different meta boxes relate to each other (e.g., whether they contain exactly the same metadata, but described with different schemes, or if one represents a superset of another).
Referring now to
The SampleToGroup box is allowed to reside in a movie fragment. Hence, sample grouping can be done fragment by fragment.
A key feature of the DVB file format is known as reception hint tracks, which may be used when one or more packet streams of data are recorded according to the DVB file format. Reception hint tracks indicate the order, reception timing, and contents of the received packets among other things. Players for the DVB file format may re-create the packet stream that was received based on the reception hint tracks and process the re-created packet stream as if it was newly received. Reception hint tracks have an identical structure compared to hint tracks for servers, as specified in the ISO base media file format. For example, reception hint tracks may be linked to the elementary stream tracks (i.e., media tracks) they carry, by track references of type ‘hint’. Each protocol for conveying media streams has its own reception hint sample format.
Servers using reception hint tracks as hints for sending of the received streams should handle the potential degradations of the received streams, such as transmission delay jitter and packet losses, gracefully and ensure that the constraints of the protocols and contained data formats are obeyed regardless of the potential degradations of the received streams.
The sample formats of reception hint tracks may enable constructing of packets by pulling data out of other tracks by reference. These other tracks may be hint tracks or media tracks. The exact form of these pointers is defined by the sample format for the protocol, but in general they consist of four pieces of information: a track reference index, a sample number, an offset, and a length. Some of these may be implicit for a particular protocol. These ‘pointers’ always point to the actual source of the data. If a hint track is built ‘on top’ of another hint track, then the second hint track must have direct references to the media track(s) used by the first where data from those media tracks is placed in the stream.
Conversion of received streams to media tracks allows existing players compliant with the ISO base media file format process DVB files as long as the media formats are also supported. However, most media coding standards only specify the decoding of error-free streams, and consequently it should be ensured that the content in media tracks can be correctly decoded. Players for the DVB file format may utilize reception hint tracks for handling of degradations caused by the transmission, i.e., content that may not be correctly decoded is located only within reception hint tracks. The need for having a duplicate of the correct media samples in both a media track and a reception hint track can be avoided by including data from the media track by reference into the reception hint track.
Currently, three types of reception hint tracks are being specified: MPEG-2 transport stream (MPEG2-TS), Real-Time Transport Protocol (RTP), and Real-Time Transport Control Protocol (RTCP) reception hint tracks. Samples of an MPEG2-TS reception hint track contain MPEG2-TS packets or instructions to compose MPEG2-TS packets from references to media tracks. An MPEG-2 transport stream is a multiplex of audio and video program elementary streams and some metadata information. It may also contain several audiovisual programs. An RTP reception hint track represents one RTP stream, typically a single media type. An RTCP reception hint track may be associated with an RTP reception hint track and represents the RTCP packets received for the associated RTP stream.
RTP is used for transmitting continuous media data, such as coded audio and video streams in networks based on the Internet Protocol (IP). The Real-time Transport Control Protocol (RTCP) is a companion of RTP, i.e. RTCP should be used to complement RTP always when the network and application infrastructure allow. RTP and RTCP are usually conveyed over the User Datagram Protocol (UDP), which, in turn, is conveyed over the Internet Protocol (IP). There are two versions of IP, IPv4 and IPv6, differing by the number of addressable endpoints among other things. RTCP is used to monitor the quality of service provided by the network and to convey information about the participants in an on-going session. RTP and RTCP are designed for sessions that range from one-to-one communication to large multicast groups of thousands of endpoints. In order to control the total bitrate caused by RTCP packets in a multiparty session, the transmission interval of RTCP packets transmitted by a single endpoint is proportional to the number of participants in the session. Each media coding format has a specific RTP payload format, which specifies how media data is structured in the payload of an RTP packet.
The metadata requirements for the DVB file format can be classified to four groups based on the type of the metadata: 1) sample-specific timing metadata, such as presentation timestamps; 2) indexes; 3) segmented metadata; and 4) user bookmarks (e.g., of favorite locations in the content).
For sample-specific timing metadata, there can be different timelines to indicate sample-specific timing metadata. Timelines may be required to cover the entire length of the recorded streams. Further, timelines may be paused. For example, timeline A may be created in the final editing phase of a movie. A service provider or another entity may insert commercials and provided timeline B for the commercials. Timeline A may be paused while commercials are ongoing. Timelines may also be transmitted after the content itself. In one embodiment, a timeline samples may be carried within MPEG-2 program elementary streams (PES). A PES conveys an elementary audio or video bitstream, and hence, timelines are accurately synchronized with audio and video frames. In this regard, reference is made to ETSI TS 102 823, “Specification for the carriage of synchronized auxiliary data”, which is hereby incorporated by reference.
Indexes may include, for example, video access points and trick mode support (e.g., fast forward/backward, slow-motion). Such operations may require, for example, indication of self-decodable pictures, decoding start points, and indications of reference and non-reference pictures.
In the case of segmented metadata, the DVB services may be described with a service guide according to a specific metadata schema, such as Broadcast Content Guide (BCG), TV-Anytime, or Electronic Service Guide (ESG) for IP datacasting (IPDC). The description may apply to a part of the stream only. Hence, the file may have several descriptive segments (e.g., a description about that specific segment of the program, such as “Holiday in Corsica near Cargese”) information.
In addition, the metadata and indexing structures of the DVB file format are required to be extensible and user-defined indexes are required to be supported. Additionally, there should be a method for fast access to elements in the file (e.g., via index tables). Furthermore, it should be possible to signal the accuracy of any index. For example, the index might be perfectly accurate, accurate to ±a specified amount or a heuristic guess only.
Various techniques for performing indexing and implementing segmented metadata have been proposed. Such techniques include, for example, sample events and sample properties. The sample event mechanism enables the listing of samples associated with a particular event type and event description index in one data structure, referred to as the SampleToEvent box. Each entry in the list may be additionally accompanied by a value or description, having a determined size in terms of bytes. Samples are listed with reference to their sample number within the associated track. It is not required to have all samples listed for a particular event type. The allowed values of the event type are predetermined and specified, e.g., in the DVB File Format specification. For each used event type in any SampleToEvent box, there is a respective SampleEventDescription box associated with the same event type. A SampleEventDescription box contains a list of sample event description entries, each associated with a unique event description index (staring from 1 and incremented by 1 per each entry in the list). The syntax and semantics of sample event description entries are determined by the event type and are specified e.g. in the DVB File Format specification. The semantics specified for a particular sample description index and event type apply to those samples that are associated with that sample description index and event type in SampleToEvent boxes. The SampleToEvent box is allowed to reside in a movie fragment. Hence, a listing of sample events can be performed fragment by fragment. There may be multiple instances of the SampleToEvent and SampleEventDescription boxes for different event types, but there can be at most one SampleToEvent box of a particular event type per a movie fragment and at most one SampleEventDescription box of a particular event type within a file.
Syntax of the SampleToEvent box and the SampleEventDescription box is specified as follows:
aligned(8) class SampleToEventBox extends FullBox(,,steb“) {
unsigned int(32) event_type;
unsigned int(32) entry_count;
for(i=1; i<=entry_count; i++) {
unsigned int(32) event_desc_index;
unsigned int(32) sample_count;
for(j=1; j<=sample_count; j++) {
unsigned int(32) sample_number;
unsigned int(8) value[event_data_length];
}
}
}
aligned(8) class SampleEventDescriptionBox extends FullBox(,,sedb“) {
unsigned int(32) event_type;
unsigned int(32) entry_count;
for(i=1; i<=entry_count; i++) {
SampleEventDescriptionEntry( );
}
}
Semantics of the syntax elements used in the SampleToEvent box and the SampleEventDescription box are specified as follows:
event_type is a 32-bit unsigned integer which shall contain the four-character code of an event type, built of ASCII characters.
entry_count is a 32-bit unsigned integer with the number of entries in the following loop.
event_desc_index is a 32-bit unsigned integer with the index of the SampleEventDescriptionEntry.
sample_count is a 32-bit unsigned integer with the number of entries in the following loop which contains the sample numbers of the samples the event applies to.
value is an array of event-specific data. The length of this field is specified by event_data_length in the corresponding SampleEventDescriptionEntry. This field is constant size for all indexes in this loop and padding shall be applied for variable-length indexes.
In unicast, multicast, and broadcast streaming applications the available streams are announced and their coding formats are characterized to enable each receiver to conclude if it can decode and render the content successfully. Sometimes a number of different format options for the same content are provided, from which each receiver can choose the most suitable one for its capabilities and/or end-user wishes. The available media streams are often described with the corresponding media type and its parameters that are included in a session description formatted according to the Session Description Protocol (SDP). In unicast streaming applications the session description is usually carried by the Real-Time Streaming Protocol (RTSP), which is used to set up and control the streaming session. In broadcast and multicast streaming applications, the session description may be carried as part of the electronic service guide (ESG) for the service. The SDP information for a broadcast stream may be updated during the broadcast. Hence, SDP for a recorded stream should be stored in such a manner that it can be updated later.
Referring now to
Referring now to
One grouping type is defined per metadata or index type in the DVB file format specification. Examples of potential grouping types include the following:
Referring now to
In another embodiment, the DVB file generator 610 of the device 600 creates one or more reception hint tracks and file metadata for the moov box or the moof box. The device 600 also derives indexes, such as a randomly accessible intra picture indication, from received packets. This may be performed by the DVB file generator 610 or by another component of the device 600, such as the processor 630 or another component not shown in
In this embodiment, one metadata sample can be created for every reception hint sample, regardless of whether or not an index or a piece of metadata is associated with that reception hint sample. The payload of the corresponding metadata sample indicates if no index or piece of metadata is associated with a reception hint sample. Further, potential track fragments of a timed metadata track can be aligned to the corresponding reception hint track in terms of the first and last included sample in the fragments. It can be indicated within a file in a track-header-level box or structure whether or not there is one timed metadata sample per each reception hint sample, and whether or not track fragments of a timed metadata track are identical to those of the associated reception hint track. When one metadata sample is created per each reception hint sample, corresponding track fragments of a timed metadata track and a reception hint track may contain samples having identical sample numbers.
It is noted that the creation of sample grouping may be implemented in a number ways. For example, the creation of sample grouping may not happen simultaneously with the reception and storage of the samples themselves. Instead, sample grouping may be performed off-line, based on an intermediate format for storing metadata or based on the timed metadata track. Referring to
The file re-writer and the second recording storage referred to above may reside in the same device as the receiver 150, the recording storage 155, the decoder 160, or a different device. Moreover, the file re-writer and the second recording storage may reside in the same device or different devices from each other.
Received streams can be of different nature—they may contain several media types and programs in a multiplexed format (MPEG-2 transport streams) or they may contain a single media type of a single program (RTP streams). Consequently, the receiver operation and file format structures are adapted accordingly as follows.
It should be noted that the various embodiments of the present invention are not limited to any particular type of sample grouping, such as the sample grouping specified in ISO/IEC 14496-12. The following embodiment describes a sample grouping mechanism that can be applied with various embodiments of the invention.
In addition to the grouping type, a sample grouping may be characterized by zero or more global parameters. Examples of global parameters include a program identifier (PID) within an MPEG-2 transport stream and a timeline identifier. When parsing sample groups, only a particular value of the global parameter is of interest at a time, e.g., a desired timestamp within a particular timeline. In addition to enumerated sample group description entries, it can advantageous to allow local parameters that may change for each entry in the SampleToGroup box. For example, a timecode can be associated with an entry in the SampleToGroup box. When parsing a sample grouping, different values of local parameters are interleaved into the same structure, i.e. in the same SampleToGroup box.
Syntax of a sample grouping mechanism of this embodiment is presented below.
aligned(8) class DVBSampleToGroupBox
extends FullBox(‘dstg’, version = 0, flags)
{
unsigned int(32) grouping_type;
if (flags & 1) {
unsigned int(8) info_length;
unsigned int(8) instance_info[info_length];
}
if (flags & 16) {
unsigned int(8) static_length;
unsigned int(8) static_info[static_length];
}
if (flags & 2)
unsigned int(8) payload_length;
unsigned int(32) entry_count;
for (i=1; i <= entry_count; i++) {
if (flags & 4)
unsigned int(32) sample_number;
if (!(flags & 8))
unsigned int(32) sample_count;
if (flags & 2)
unsigned int(8) payload[payload_length];
else
unsigned int(32) group_description_index;
}
}
The presented DVBSampleToGroup box is structurally compatible with the SampleToGroup box, i.e., if flags (of the box header) are 0, then the syntax is identical to the SampleToGroup box. Global parameters as specified above are represented by grouping instance information (instance_info). static_info represents static values remaining unchanged for all grouped samples. The semantics of static_info depend on grouping_type. For example, static_info may be formatted according to the payload of the DVBIndexBox (presented subsequently), giving timing and sample number inaccuracy ranges. Local parameters are represented by payload or group_description_index. The DVBSampleToGroup box can be parsed without semantic knowledge of grouping_type thanks to info_length, static_length, and payload length. A file editor software program or device that is unaware of a particular grouping_type can edit the box (e.g. as response of removal or insertion of samples), even if it did not have knowledge about the semantics of a particular grouping_type. sample_number indicates the first sample for which the payload or group_description_index apply, and the persistence of the payload or group_description_index in terms of number of consecutive samples is indicated by sample_count. If sample_number is not present, then the first sample for which the payload or group_description_index apply is subsequent to the last sample for which the previous entry in the loop applied. Values of sample_count in a DVBSampleToGroup box is increasing in appearance order. A sample can be included only once in a DVBSampleToGroup box, i.e., the sum of previous values of sample_number and sample_count is less than the present value of sample_number. It is not required that each sample in a track is mapped in a DVBSampleToGroup box. The DVBSampleGroupDescription box and group_description_index can be used for enumerated indexes. The value of payload or group_description_index for those samples that are not mapped in a DVBSampleToGroup box is unspecified. The semantics of instance_info and payload are specified for each grouping_type.
An alternative definition for the syntax of the DVBSampleToGroup box is presented below. A primary difference compared to the previous case is that varying-length payloads are enabled.
aligned(8) class DVBSampleToGroupBox
extends FullBox(‘dstg’, version = 0, flags)
{
unsigned int(32) grouping_type;
if (flags & 1) {
unsigned int(8) info_length;
unsigned int(8) instance_info[info_length];
}
if (flags & 16) {
unsigned int(8) static_length;
unsigned int(8) static_info[static_length];
}
if (flags & 2)
unsigned int(8) payload_length;
unsigned int(32) entry_count;
for (i=1; i <= entry_count; i++) {
if (flags & 4)
unsigned int(32) sample_number;
if (!(flags & 8))
unsigned int(32) sample_count;
if (flags & 2)
unsigned int(8) payload[payload_length];
else if (flags & 32) {
unsigned int(8) varying_length;
unsigned int(8) payload[varying_length];
}
else
unsigned int(32) group_description_index;
}
}
The DVBSampleToGroup box 700 is allowed to reside in a movie fragment. Hence, sample grouping can be done fragment by fragment.
Syntax of DVBSampleGroupDescription box is presented below.
aligned(8) class DVBSampleGroupDescriptionBox
(unsigned int(32) handler_type)
extends FullBox(‘dsgd’, version, flags)
{
unsigned int(32) grouping_type;
if (flags & 1) {
unsigned int(8) info_length;
unsigned int(8) instance_info[info_length];
}
if (version>=1)
unsigned int(32) default_length;
unsigned int(32) entry_count;
int i;
for (i = 1; i <= entry_count ; i++){
if (version>=1 && default_length==0)
unsigned int(32) description_length;
switch (handler_type){
case ‘vide’: // for video tracks
VisualSampleGroupEntry ( );
break;
case ‘soun’: // for audio tracks
AudioSampleGroupEntry( );
break;
case ‘hint’: // for hint tracks
HintSampleGroupEntry( );
break;
}
}
}
The structure of the DVBSampleGroupDescription box is compatible with the SampleGroupDescription box, i.e., if flags (in the box header) are set to 0, the syntax is identical to the SampleGroupDescription box. Global parameters as defined above are represented by grouping instance information (instance_info). The inclusion of instance_info in the DVBSampleGroupDescription box enables different enumerations (i.e., sample group entries) for different values of the global parameters.
The syntax presented above for the DVBSampleToGroup box only allows a fixed-length payload or group_description_index being specified for each entry in the loop. It may be desirable for some grouping types to have varying-length payloads for any entry specified in a particular DVBSampleToGroup box. The following extension mechanism, referred to as DVBSampleToGroupExtension box is depicted below for this purpose.
aligned(8) class DVBSampleToGroupExtensionBox
extends FullBox(‘dsge’, version = 0, flags)
{
unsigned int(32) grouping_type;
if (flags & 1) {
unsigned int(8) info_length;
unsigned int(8) instance_info[info_length];
}
if (flags & 2)
unsigned int(32) extension_type;
if (!(flags & 4))
unsigned int(16) ext_length;
unsigned int(32) entry_count;
for (i=1; i<=entry_count; i++) {
if (flags & 4)
unsigned int(16) ext_length;
unsigned int(8) extension[ext_length];
}
}
Each entry in the loop corresponds to the respective entry (having the same loop counter value) in the DVBSampleToGroup box having identical grouping_type and global parameter values (i.e., identical values of info_length and instance_info). In other words, extension applies to the same samples as payload when the loop counter value is the same in respective DVBSampleToGroupExtension and DVBSampleToGroup boxes. If extension_type is not present, then the default semantics for an extension of a grouping_type are applied. When extension_type is present, then it determines the semantics of extension. Similarly to the DVBSampleToGroup box, a DVBSampleToGroupExtension box of a particular global parameter and extension_type can be included for each movie fragment.
In addition to varying-length local parameters for sample groups, the DVBSampleGroupExtension box can be used for providing new pieces of information for a sample grouping while maintaining compatibility with existing sample group definitions and parsers. In other words, a parser not capable of interpreting the DVBSampleToGroupExtension box or a particular extension_type can still parse the corresponding DVBSampleToGroup box successfully.
In one embodiment, the instance_info used above in the syntax for DVB SampleToGroup, DVB SampleGroupDescription, and DVBSampleToGroupExtension boxes is an index to the loop included in the DVBGroupingInstance box having an identical value of grouping_type as the aforementioned boxes. The syntax of the DVBGroupingInstance box can be specified as follows. The semantics of instance_descr are determined by grouping_type.
aligned(8) class DVBGroupingInstanceBox
extends FullBox(‘dgri’, version = 0, flags)
{
unsigned int(32) grouping_type;
unsigned int(32) default_length;
unsigned int(32) entry_count;
for (i=1; i<=entry_count; i++) {
if (default_length == 0) {
unsigned int(32) descr_length;
unsigned int(8) instance_descr[descr_length];
}
else
unsigned int(8) instance_descr[default_length];
}
}
In one embodiment, extension_type can be parameterized similarly to global and local parameters discussed previously. In other words, rather than identifying an extension based on its extension_type only, it can be associated with a parameter further specifying its intent and scope. The associated parameter can also be an index to a loop in another box, which contains the actual parameter values.
When sample grouping is applied to MPEG2-TS reception hint track (or any other multiplexed stream), the following holds:
When sample grouping is applied to a media track or RTP reception hint track, indices and metadata for a single media type can be represented by specific grouping types. However, as sample grouping mechanism operates only within one track, metadata that applies to more than one media track or reception hint track (later referred to as “multi-track” metadata) should be represented. This issue can materialize in two cases. First, both media tracks and reception hint track(s) may be created from the received stream. Second, there may be multiple RTP reception hint tracks. It is suggested that the sample grouping mechanism of the invention is associated with reception hint tracks whenever they are present in the file. Thus, the first case above can be avoided. For the second case, several options are contemplated.
First, “multi-track” metadata may be associated with any related track. When the metadata is associated to a reception hint track or a media track, the metadata implicitly applies to all reception hint tracks or media tracks, respectively, within the same period of decoding time. This alternative requires parsers to check all tracks for potential “multi-track” metadata.
Alternatively, “multi-track” metadata may be associated with a “master” track. The master track may be explicitly indicated with a new box within the moov box, for example. The new box may also list all the inter-linked tracks for the “multi-track” metadata.
In another embodiment, timed metadata track may be associated with many media tracks or RTP reception hint tracks. Sample grouping may be applied to the timed metadata track. The referred media and reception hint tracks may be listed as track references of type ‘cdsc’. This serves as an explicit indication which tracks the metadata pertains to. If used, sample duration can be set to match the persistence of each metadata item.
In one embodiment of the present invention, in the embodiment of the “master” track described above, the new box is referred to as a track relation box and indicates relationship between tracks. The track relation box is defined as follows:
Box Type: ‘trel’
Container: Movie Box (‘moov’)
Mandatory: No
Quantity: Zero or One
The syntax for the track relation box is as follows:
aligned(8) class TrackRelationBox
extends FullBox(‘trel’, version = 0, flags) {
int i,j;
if(flags & 0x000001 == 1) {
unsigned int(16) num_reception_groups;
for(i=1; i<=num_reception_groups; i++) {
unsigned int(16) num_tracks_in_reception_group;
for(j=1; j<=num_tracks_in_reception_group; j++)
unsigned int(32) track_id;
}
}
}
}
In the above syntax, “version” is an integer that specifies the version of the track relation box (0 as described above), and “flags” is a 24-bit integer with flags. The following bits are defined, where bit 0 is the least significant bit, bit 1 is the second least significant bit, etc. When bit 0 is equal to 1, this indicates that information of reception groups is present in this box. When bit 0 is equal to 0, this indicates that information of reception groups is not present in this box. It is obvious for a person skilled in the art that other types of track relations could also be specified and indicated with a different bit number than bit 0.
“num_reception_groups” indicates the number of reception groups that are signaled. A reception group comprises reception hint tracks that are received simultaneously or are otherwise part of the same recording. In addition, a reception group may comprise media tracks that are derived based on reception hint tracks. “track_id” indicates the track identifier of the j-th track in the i-th reception group that is signaled. The first track_id of a reception group is a master track that may contain segmented metadata that is indicated with sample groups and is associated jointly with all the tracks in the reception group. No other track than a master track may be allowed to contain segmented metadata associated with all the tracks in the reception group.
When timed metadata tracks for indexes or segmented metadata are created, in one embodiment, the following practices may be followed.
One timed metadata track is created for program-specific indexes and metadata of a single-program MPEG-2 transport stream. Program-specific indexes and metadata apply equally to audio and video streams of a program and to any other potential components of the program, such as subtitle streams.
One timed metadata track per program is created for program-specific indexes and metadata of a multi-program MPEG-2 transport stream. In this regard, a timed metadata track may contain metadata of only one program. The program can be identified by its program_number value, which is a 16-bit unique identifier for programs within an MPEG-2 transport stream, used in program association and program map tables of an MPEG-2 transport stream, for example. In one embodiment, the parameter program_number can be included in the sample entry structure for timed metadata tracks associated with MPEG2-TS reception hint tracks.
One timed metadata track is created for media-specific indexes of each elementary stream of an MPEG2-TS program. Media-specific indexes may apply only to a particular media type. For example, they can be indications of reference and non-reference frames of video or indications of the temporal scalability level of video.
One timed metadata track may be created for media-specific indexes for an RTP stream. Further, one timed metadata track may be created for program-specific indexes of multiple RTP streams. The timed metadata track is associated with the RTP reception hint tracks using track references. In another embodiment, the timed metadata track is associated with the “master” reception hint track with a track reference and the other associated reception hint tracks are indicated through the TrackRelationBox as presented above.
Even though one program-specific timed metadata track and one media-specific timed metadata track per elementary media stream may often be preferable, more than one timed metadata tracks can be created. For example, if an alternative timeline for the program is provided subsequently to the program itself, it is more practical from the file arrangement point of view to create a new timed metadata track for the provided timeline.
For embodiments using timed metadata tracks, a receiver in accordance with an embodiment of the present invention can operate as follows as a response to each received packet:
A receiver with a greater amount of buffer memory may arrange several metadata samples and reception hint samples in continuous chunks of memory and, therefore, save in the storage space required for the sample to chunk box and the chunk offset box.
Indexes and segmented metadata have the following characteristics when it comes to reception hint samples that are associated with them. An index may indicate a characteristic to be valid from the associated reception hint sample onwards, usually until the next index of the same type. For example, an index may indicate a polarity change of scrambling in MPEG-2 transport stream. An index may indicate a characteristic of a single reception hint sample or an event that is synchronized with a reception hint sample. A bookmark is an example of such an index.
An index may indicate a characteristic of the stream in between the associated reception hint sample and the previous reception hint sample. An indication of missing packets is such an index.
An index may indicate a characteristic of a coded media sample. It is noted that timed metadata tracks proposed in this invention are associated to reception hint samples, reception hint samples do not usually contain exactly one media sample, and data for one media sample may or may not reside in contiguous reception hint samples (e.g., because elementary audio and video streams are multiplexed in an MPEG-2 transport stream). Consequently, there are at least two options how to index media samples: (1) an index is associated only with the first reception hint sample containing data for a media sample; and (2) an index is associated with all reception hint samples containing data for a media sample.
Sample groups associated with timed metadata tracks that contain indexes and segmented metadata can be created in any of the following ways:
If an index is associated with a coded media sample, a sample group may indicate only the first reception hint sample containing data for the media sample or all reception hint samples containing data for the media sample.
Exemplary sample entry formats may be provided as follows:
abstract class IndexSampleEntry( ) extends MetadataSampleEntry
(‘ixse’) {
unsigned int(16) entry_count;
for (i = 1; i <= entry_count; i++)
unsigned int(32) index_type_4cc;
}
class MPEG2TSIndexSampleEntry( ) extends IndexSampleEntry
(‘m2ix‘) {
unsigned int(16) program_number;
}
class MPEG2TSMediaIndexSampleEntry( ) extends
IndexSampleEntry (‘m2mi‘) {
unsigned int(3) reserved;
unsigned int(13) pid;
}
class RTPIndexSampleEntry( ) extends IndexSampleEntry (‘rtix‘) {
}
class RTPReceptionGroupIndexSampleEntry( ) extends
IndexSampleEntry (‘rgix‘) {
}
“IndexSampleEntry” indicates the types of indexes that may be present in samples associated with this sample entry. If “entry_count” is equal to 0, any indexes may be included in samples associated with this sample entry. If “entry_count” is greater than 0, a loop of “index_type_4cc” values is given and each value of “index_type_4cc” indicates a four-character code for a box that may be present in samples associated with this sample entry. The sample entries for a timed metadata track therefore indicate the types of metadata that may occur in the timed metadata track. It is noted that other mechanisms to indicate the types of metadata present or allowed in a timed metadata track can be envisioned, e.g., as extensions of track header boxes.
“MPEG2TSIndexSampleEntry” is used for timed metadata tracks containing program-specific indexes and metadata and associated with an MPEG2-TS reception hint track. “program_number” identifies a program within the MPEG-2 transport stream.
“MPEG2TSMediaIndexSampleEntry” is used for timed metadata tracks containing media-specific indexes and metadata and associated with an elementary stream contained in an MPEG2-TS reception hint track. pid is an identifier for an elementary stream within an MPEG-2 transport stream.
“RTPIndexSampleEntry” is used for timed metadata tracks containing media-specific indexes and metadata and associated with an RTP reception hint track.
“RTPReceptionGroupIndexSampleEntry” is used for program-specific indexes and metadata and associated to more than one RTP reception hint track.
An exemplary sample format for a timed metadata track containing indexes and segmented metadata is provided below:
aligned(8) class IndexSample {
box index_box[ ];
}
“IndexSample” contains zero or more index boxes. The four-character code for the included index boxes must be among those indicated by the associated sample entry.
Exemplary index boxes which can be used with the invention and which were originally proposed in DVB document TM-FF0058r0 are provided as follows:
abstract aligned(8) class DVBIndexBox (type) extends Box(type) {
unsigned int(4)
time_accuracy;
unsigned int(4)
sample_accuracy;
if(time_accuracy >= 8)
unsigned int(32)
max_timing_inaccuracy;
if(sample_accuracy >= 8)
unsigned int(32)
max_sample_accuracy;
}
The following values are specified for “time_accuracy” and “sample_accuracy”: 0x0: accurate, 0x1: unspecified, 0x2: heuristic, 0x3: reserved (no maximum provided), 0x4-0x7: application-specific (no maximum provided), 0x8: maximum inaccuracy specified, 0x9: reserved (maximum inaccuracy provided), 0xA-0xF: application-specific (maximum inaccuracy provided).
aligned(8) class DVBVideoIndex extends DVBIndexBox(‘idvi’) {
unsigned int(8)
video_event_mask;
unsigned int(24)
video_event_length;
};
“video_event_mask” is a bit mask indicating the video event(s) that start in the indicated sample, as per Table 1 below.
TABLE 1
Mask values used for “video_event_mask”
Mask
Meaning
0x01
video decode start point (e.g. a Random Access Point)
0x02
Self decodable picture (e.g. I frame)
0x04
Reference Picture
0x08
P Picture
0x10
B Picture
“video_event length” is the number of samples (transport packets) that make up this video picture, including the current packet. The value ‘0’ shall be used to mean “unknown”. The Sync Sample Box can also carry the indexes to the events of type 0x01.
aligned(8) class DVBPCRIndex extends DVBIndexBox(‘idpi’) {
unsigned int(1)
PCR_discontinuity_flag;
unsigned int(5)
reserved_0;
unsigned int(42)
PCR_Value;
}
“PCR_discontinuity_flag” is a field that is set to ‘1’ if there is a program clock reference (PCR) discontinuity in the associated PCR event; otherwise it is set to 0′.
“PCR_value” is the 27-MHz value extracted from the PCR that is indexed, for example, as per equation (2-1) in ISO/IEC International Standard 13818-1.
aligned(8) class DVBPolarityChange extends DVBIndexBox(‘idpc’) {
unsigned int(8) polarity;
}
“polarity” is the polarity of the associated event, as per Table 2 below:
TABLE 2
Interpretation of Polarity values
Value
Meaning
0
Clear
1
Odd polarity
2
Even polarity
This value indicates the new polarity value which applies, and the timed metadata sample corresponds to the first reception hint sample with this new polarity. A polarity change index shall only be deemed to occur when the polarity of a stream of packets on a given PID changes, and not when it changes between packets of different PIDs.
aligned(8) class DVBCAIndex extends DVBIndexBox(‘idci’) {
unsigned int(8)
polarity;
unsigned int(8)
ca_event_data[ ];
}
“ca_event_data” is the bytes that comprise the packet carrying the conditional access (CA) event; typically, this will be an ECM. The “ca_event_data” continues until the end of the box. The length of the “ca_event_data” can be determined from the length of the box.
aligned(8) class DVBTimecodeIndex extends DVBIndexBox(‘idtc’) {
unsigned int(8)
timeline_id;
unsigned int(2)
reserved_0;
unsigned int(6)
tick_format; // as per table 6 in
TR 102 823
unsigned int(32)
absolute_ticks;
}
“timeline_id” is the identifier of the timeline, as per ETSI TS 102 823, clause 5.2.2.4, and “tick_format” is a field which specifies the format that the “absolute_ticks” field takes, as per the Synchronized Auxiliary Data (SAD) specification, ETSI TS 102 823. “absolute_ticks” is the timecode, coded as indicated by the field “tick_format”.
aligned(8) class DVBSectionUpdateIndex extends DVBIndexBox(‘idsu’) {
unsigned int(8)
table_id;
unsigned int(16)
table_id_extension;
unsigned int(8)
section_no;
unsigned int(n*8)
section_data; // optional
}
“DVBIDTable” box can be included, for example, in a sample table box for a timed metadata track. “table_id” is the table id of the section version update that is being indexed. “table_id_extension” (or “program_number” for PMT, or “transport_stream_id” for PAT) from the section version update is being indexed. “section no” is the section number to which this update applies. “section_data” is the field may not be present. If it is present it contains the section data of the new version. The section data shall continue until the end of the box. The length of the “section_data” can be determined from the length of the box.
aligned(8) class DVBIDIndex extends DVBIndexBox(‘didi‘) {
unsigned int(5)
reserved;
unsigned int(3)
running_status; // As per table
105 in 102 323
unsigned int(24)
ID_Table_index;
}
“running_status” is the field which indicates the status of the ID that is referenced by the “ID_Table_index” field (e.g., if the ID is running or paused). The value of this field is define in table 105 of TS 102 323. “ID_Table_index” is an index into the “DVBIDTableBox”, indicating the ID that applies at this location with the indicated running_status.
aligned(8) class DVBIDTable extends FullBox(‘didt‘, version = 0, 0) {
unsigned int(32) ID_count;
for(i=0;i<ID_count;i++) {
string ID; //in URI Format
}
}
“ID_count” is the number of IDs that follow in the “DVBIDTable”. “ID” is the URI formatted ID.
Further examples of index boxes are provided below:
aligned(8) class SDPUpdate extends DVBIndexBox(‘idsd’) {
string sdp_text;
}
“sdp_text” is a null-terminated string containing an SDP description that is valid starting from the indicated sample.
aligned(8) class KeyUpdate extends DVBIndexBox(‘idkm’) {
string key_message;
}
“key_message” contains the cryptographic key to be used for deciphering the packet payloads starting from the related reception hint sample.
aligned(8) class ErrorIndex extends DVBIndexBox(‘idei’) {
unsigned int(2) packet_header_error;
unsigned int(2) packet_payload_error;
unsigned int(2) packet_sequence_gap;
unsigned int(2) reserved;
}
For “packet_header_error”, a value 0x0 indicates that the packet header contains no errors. Value 0x1 indicates that the packet header may or may not contain errors. Value 0x2 indicates that the packet header contains errors. Value 0x3 is reserved.
For “packet_payload_error” a value 0x0 indicates that the packet payload contains no errors. Value 0x1 indicates that the packet payload may or may not contain errors. Value 0x2 indicates that the packet payload contains errors. Value 0x3 is reserved.
For “packet sequence_gap”, a value 0x0 indicates that the packet immediately follows the previous packet in the reception hint track in transmission order. Value 0x1 indicates that the packet may or may not immediately follow the previous packet in the reception hint track in transmission order. Value 0x2 indicates that the packet does not immediately follow the previous packet in the reception hint track in transmission order, for example, that a there is at least one missing packet preceding this packet. Value 0x3 is reserved.
In another embodiment, an Index Sample Entry format for a timed metadata track containing indexes is as follows:
class IndexSampleEntry( ) extends MetadataSampleEntry (‘ixse’) {
unsigned int(16) program_number;
unsigned int(16) entry_count;
int(32) sample_number_offset;
for (i = 1; i <= entry_count; i++) {
unsigned int(32) index_type_4cc;
unsigned int(8) instance_info_length;
unsigned int(8) payload_length;
unsigned int(8) extension_count;
}
}
An IndexSampleEntry indicates the types of indexes that may be present in samples associated with this sample entry. program_number identifies a program within the MPEG-2 transport stream. If entry_count is equal to 0, then any indexes may be included in samples associated with this sample entry, and default values (according to the first release of the DVB File Format specification) of instance_info_length, payload length, and extension_count are used. sample_number_offset specifies an offset to be added to the sample_number in the associated timed metadata samples to obtain the sample number in the referred track. index_type_4cc indicates an index type that may be present in samples associated with this sample entry. instance_info_length, payload length, and extension_count indicate the values of these fields used in index samples. A DVB File Format parser ignores those bytes index_instance_info, index_payload and index_extension that are beyond the byte range specified in this document for the respective index_type_4cc. This mechanism allows appending fields in a backward-compatible manner in later index format versions. If there are many timed metadata tracks for a reception hint track, then index_type_4cc values can be used to locate the track containing the desired indexes.
An index sample contains one or more index events, each having the following abstract syntax. The semantics of index_instance_info and index_payload are determined by the 32-bit index_type and the semantics of index_extension is determined by extension_type as specified subsequently. index_type corresponds to grouping_type in sample grouping.
abstract class IndexEvent(index_type, info_length, payload_length,
extension_count) {
unsigned int(8) index_instance_info[info_length];
unsigned int(8) index_payload[payload_length];
for (i=1; i<=extension_count; i++) {
unsigned int(32) extension_type;
unsigned int(16) extension_length;
unsigned int(8) index_extension[extension_length];
}
unsigned int(8) pad[ ];
}
abstract class IndexInstanceInfo(index_type, info_length)
{
}
abstract class IndexPayload(index_type, payload_length)
{
}
abstract class IndexExtension(extension_type, extension_length)
{
}
The semantics of index_instance_info, index_payload, extension_type, and index_extension in the IndexEvent structure are identical to the semantics of index_instance_info, index_payload, extension_type, and index_extension, respectively, used in DVBSampleToGroup, DVBSampleGroupDescription, and DVBSampleToGroupExtension boxes. This match enables the straightforward conversion between indexes included in timed metadata tracks and indexes indicated by DVBSampleToGroup boxes.
Syntax of some index events are provided below, where the semantics of the syntax elements is the same as described in the embodiment presented previously.
class VideoIndex extends IndexEvent(“idvi”, 0, 2, 0);
class VideoIndexPayload extends IndexPayload(“idvi”, 2) {
unsigned int(2)
reserved = 0;
unsigned int(2)
sample_depends_on;
unsigned int(2)
sample_is_depended_on;
unsigned int(2)
sequence_depends_on;
unsigned int(8)
dependency_level;
}
class PCRIndex extends IndexEvent(“idpc”, 0, 6, 0);
class PCRIndexPayload extends IndexPayload(“idpc”, 6) {
unsigned int(1) PCR_discontinuity_flag;
unsigned int(5) reserved_0;
unsigned int(42) PCR_Value;
}
class PolarityIndex extends IndexEvent(“idca”, 2, 1, 1);
class PolarityInstanceInfo extends IndexInstanceInfo(“idca”, 2) {
unsigned int(13) pid;
unsigned int(3) reserved;
}
class PolarityIndex extends IndexEvent(“idca”, 0, 1, 1);
class PolarityIndexPayload extends IndexPayload(“idca”, 1) {
unsigned int(6) reserved_0;
unsigned int(2) polarity;
}
class PolarityIndexExtension extends IndexExtension(“idca”,
extension_length) {
unsigned int(8) ca_event_data[extension_length];
}
class TimecodeIndex extends IndexEvent(“idtc”, 1, 5, 0);
class TimecodeIndexInstanceInfo extends IndexInstanceInfo(“idtc”, 1) {
unsigned int(8) timeline_id;
}
class TimecodeIndexPayload extends IndexPayload(“idtc”, 5) {
unsigned int(2) reserved_0;
unsigned int(6) tick_format; // as per table 6 in TR 102 823
unsigned int(32) absolute_ticks;
}
class TimeIndex extends IndexEvent(“idti”, 0, 0, 0);
class SectionUpdateIndex extends IndexEvent(“idsu”, 0, 4, 1);
class SectionUpdateIndexPayload extends IndexPayload(“idsu”, 4) {
unsigned int(8) table_id;
unsigned int(16) table_id_extension;
unsigned int(8) section_no;
}
class SectionUpdateIndexExtension extends IndexExtension(“idsu”,
extension_length) {
unsigned int(8) section_data[extension_length];
}
There are at least two types of usage for descriptive metadata. First, when a broadcast is recorded continuously (e.g., as a response of user initiating an instantaneous record operation rather than a scheduled recording of a particular program), content from more than one program may be recorded to a single file. Thus, Electronic Service Guide (ESG) information is also stored for more than one program within the file. An indexing mechanism can be used to map a group of samples comprising a program with the respective fragments of ESG information. Second, different “storylines”, such as summaries or special cuts, can be provided by broadcasters. For example, a broadcaster can provide information, in particular the time periods and descriptions, about the goals of a soccer game, general highlights of the game, or highlights of a particular player during the game. These “storylines” can be provided after the recording and may use a specific timeline (as discussed in the timecode index presented earlier). An indexing mechanism can be used for indicating “storylines” in a recorded file.
The DVBSampleToGroup box can be used for descriptive or segmented metadata as follows. The value of grouping_type can be differentiate ESG and “storyline” metadata, even though they may be structurally identical and the same grouping_type could also be used. The global parameter, instance_info, is used to differentiate between storylines. For ESG metadata, instance_info can contain the item_id of the metadata item for initiating the ESG metadata (such as giving information about the ESG provider) or it may be used to differentiate between different programs of ESG metadata. In practice, instance_info can contain the item_id of a metadata item containing the information about storylines or ESG. Alternatively, instance_info may point to the DVBGroupingInstance box, which can e.g. contain a textual name of each storyline or URIs pointing to descriptions. The entries in the loop included in the DVBSampleToGroup box can be used as follows. If the constant-size payload field is used, it may contain the item_id of the metadata item containing a description of the grouped samples, such as respective ESG fragment or descriptive data about an event in the storyline. If group_description_index is used within the loop, the DVBSampleGroupDescription box may contain the URIs of the respective ESG fragments or descriptive XML fragments about an event in the storyline. Alternatively, the DVBSampleGroupDescription box may contain textual description of the storyline events. When the syntax for DVBSampleToGroup box that allows varying-length payloads is used, the payload can be an URI used for purposes mentioned above or a textual description of an event.
When sample groups are included in a reception hint track or a media track, seeking to a position in a file based on a desired segmented metadata or index property can be done by a file parser with the following steps:
1. Select an appropriate sample group description box (based on its grouping type).
2. Select a desired group description index within the sample group description box.
3. Seek for the desired group description index from the SampleToGroup box(es) (of the selected grouping type). This reveals the sample number of the desired metadata/index property.
4. Use the sample table box to locate the sample in a file based on its sample number.
When sample groups according to the DVBSampleToGroup box are included in a reception hint track or a media track and payload rather than group_description_index is used, seeking a position in a file based on a desired segmented metadata or index property can be done by a file parser with the following processes:
1. Select an appropriate DVBSampleToGroup box (based on its grouping type and instance_info).
2. Seek for the desired value of payload in the DVBSampleToGroup box(es). This reveals the sample number having the desired metadata/index property.
3. Use the sample table box to locate the sample in a file based on its sample number.
When sample groups are included in a timed metadata track, seeking to a position in a file based on a desired segmented metadata or index property can be done by a file parser with the following processes:
1. Select an appropriate sample group description box (based on its grouping type).
2. Select a desired group description index within the sample group description box.
3. Seek for the desired group description index from the SampleToGroup box(es) (of the selected grouping type). This reveals the sample number of the desired metadata/index property within the timed metadata track.
4. Map the sample of the timed metadata track to a sample of the associated reception hint track. If one metadata sample is created for each reception hint sample, the desired sample number in the reception hint track is identical to that of the timed metadata track. Otherwise, derive the decoding timestamp of the timed metadata sample (using the decoding time to sample box) and seek for a reception hint sample having an identical decoding timestamp (or a reception hint sample that pertains at the time of the decoding timestamp)
5. Use the sample table box of the reception hint track to locate the sample in a file based on the sample number of the reception hint sample identified in the previous step.
When sample groups are included in a reception hint track or a media track, obtaining the indexes and segmented metadata valid for a particular reception hint sample or a media sample, respectively, can be done by a file parser with the following processes:
1. Select each sample to group box (one at a time) associated with the reception hint track or the media track and find the sample group description index of the of the particular reception hint sample or media sample, respectively.
2. Obtain the index or segmented metadata corresponding to the sample group description index from the respective sample group description box.
When sample groups are included in a timed metadata track, obtaining the indexes and segmented metadata valid for a particular reception hint sample can be done by a file parser with the following processes:
1. Map the reception hint sample to a metadata sample as follows. If one metadata sample is created for each reception hint sample, the desired sample number in the reception hint track is identical to that of the timed metadata track. Otherwise, derive the decoding timestamp of the reception hint sample (using the decoding time to sample box) and seek for a timed metadata sample having an identical decoding timestamp (or a timed metadata sample that pertains at the time of the decoding timestamp).
2. Select each sample to group box (one at a time) associated with the timed metadata track and find the sample group description index of the of the particular timed metadata sample, respectively.
3. Obtain the index or segmented metadata corresponding to the sample group description index from the respective sample group description box.
A player, typically comprising a file parser, a decoder 160, and a renderer 170, can provide a list of indices, bookmarks, or other types of metadata to the end-user by analyzing and re-producing some part of the content of the sample group description boxes.
A player may provide bookmarking operation to end-users. Bookmarks can be stored to a file using sample grouping similarly as described above for the receiver operation
Thus, in accordance with embodiments of the present invention, sample groups can be used with movie fragments seamlessly, because SampleToGroup box can reside in a track fragment box. Hence, simultaneous recording and playback of a single file is fully supported. Further, the sample group mechanism is extensible, as new grouping types can be specified. Persistence of metadata is clear: contiguous samples with the same value of group_description_index (of the same grouping type). Further, getting a table of contents of the file based on SampleGroupDescription boxes is straightforward. Thus, when storing metadata associated with single samples, e.g. bookmarks, the sample grouping mechanism can be better than timed metadata tracks.
While embodiments of the invention have mostly concerned a receiver 150, a recording storage 155, a file parser, and a decoder 160, the presented file structures can also be used by an encoder 110 when creating a file. In particular, when a file is provided for playback without converting its media contents to a packetized format, it is often useful to provide segmented metadata and indices structured as proposed in the invention. In accordance with the ISO base media file format, the encoder 110 derives file metadata that will be later stored within the moov box or the moof box, depending on whether or not movie fragments are in use. For any received piece of metadata, the encoder 110 identifies its type and creates a new sample group description box, if the type of metadata did not exist before. The encoder 110 also checks whether or not the content of the piece of metadata differs from all metadata stored earlier in the file. If the content is novel, then a new entry is created for the sample group description box and the metadata content is copied into the entry. The SampleToGroup box is maintained for each grouping type and samples of a media track or a hint track are mapped to one of the entries in the respective sample group description box.
It is noted that the above-described creation of sample grouping may be implemented in a number ways. For example, the creation of sample grouping may not happen simultaneously with the encoding and storage 120 of the samples themselves. Instead, sample grouping may be performed off-line, based on an intermediate format for storing metadata. Referring again to
The server file re-writer and the second storage referred to above may reside in the same device as the encoder 110, the storage 120, the server 130, or a different device. Moreover, the server file re-writer and the second storage may reside in the same device or different devices from each other.
In one embodiment, the server 130 typically includes a file parser (not shown in
While particular embodiments of the present invention have been disclosed, it is to be understood that various different modifications and combinations are possible and are contemplated within the true spirit and scope of the appended claims. There is no intention, therefore, of limitations to the exact abstract and disclosure herein presented.
Various embodiments of present invention described herein are described in the general context of method steps, which may be implemented in one embodiment by a program product including computer-executable instructions, such as program code, executed by computers in networked environments. A computer-readable medium may include removable and non-removable storage devices including, but not limited to, Read Only Memory (ROM), Random Access Memory (RAM), compact discs (CDs), digital versatile discs (DVD), etc. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. Computer-executable instructions, associated data structures, and program modules represent examples of program code for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represents examples of corresponding acts for implementing the functions described in such steps. Various embodiments of the present invention can be implemented directly in software using any common programming language, e.g. C/C++ or assembly language.
Software and web implementations of the present invention could be accomplished with standard programming techniques with rule based logic and other logic to accomplish the various database searching steps, correlation steps, comparison steps and decision steps. It should also be noted that the words “component” and “module,” as used herein and in the claims, is intended to encompass implementations using one or more lines of software code, and/or hardware implementations, and/or equipment for receiving manual inputs.
Individual and specific structures described in the foregoing examples should be understood as constituting representative structure of means for performing specific functions described in the following the claims, although limitations in the claims should not be interpreted as constituting “means plus function” limitations in the event that the term “means” is not used therein. Additionally, the use of the term “step” in the foregoing description should not be used to construe any specific limitation in the claims as constituting a “step plus function” limitation. To the extent that individual references, including issued patents, patent applications, and non-patent publications, are described or otherwise mentioned herein, such references are not intended and should not be interpreted as limiting the scope of the following claims.
The foregoing description of embodiments of the present invention have been presented for purposes of illustration and description. It is not intended to be exhaustive or to limit the present invention to the precise form disclosed, and modifications and variations are possible in light of the above teachings or may be acquired from practice of the present invention. The embodiments were chosen and described in order to explain the principles of the present invention and its practical application to enable one skilled in the art to utilize the present invention in various embodiments and with various modifications as are suited to the particular use contemplated. The features of the embodiments described herein may be combined in all possible combinations of methods, apparatus, modules, systems, and computer program products.
Patent | Priority | Assignee | Title |
10778743, | Jul 01 2014 | Canon Kabushiki Kaisha | Method for identifying objects across time periods and corresponding device |
11070893, | Mar 27 2017 | Canon Kabushiki Kaisha | Method and apparatus for encoding media data comprising generated content |
11265622, | Mar 27 2017 | Canon Kabushiki Kaisha | Method and apparatus for generating media data |
Patent | Priority | Assignee | Title |
7110664, | Apr 20 2001 | Rovi Technologies Corporation | Methods and apparatus for indexing and archiving encoded audio-video data |
7155531, | Jun 12 2001 | Network Appliance | Storage methods and apparatus for streaming media data |
7483532, | Jul 03 2003 | Microsoft Technology Licensing, LLC | RTP payload format |
20020009172, | |||
20030061369, | |||
20030110503, | |||
20040167925, | |||
20050002525, | |||
20050102371, | |||
20070016594, | |||
20070186005, | |||
20080256326, | |||
20090282077, | |||
20100114848, | |||
20100198887, | |||
CN1578311, | |||
CN1650628, | |||
DE10392281, | |||
EP1500002, | |||
KR2004008541, | |||
WO3073768, | |||
WO3098475, | |||
WO2007080500, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Aug 19 2008 | Nokia Technologies Oy | (assignment on the face of the patent) | / | |||
Sep 05 2008 | HANNUKSELA, MISKA M | Nokia Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 021819 | /0108 | |
Jan 16 2015 | Nokia Corporation | Nokia Technologies Oy | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 040811 | /0991 |
Date | Maintenance Fee Events |
Jun 09 2021 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
Dec 26 2020 | 4 years fee payment window open |
Jun 26 2021 | 6 months grace period start (w surcharge) |
Dec 26 2021 | patent expiry (for year 4) |
Dec 26 2023 | 2 years to revive unintentionally abandoned end. (for year 4) |
Dec 26 2024 | 8 years fee payment window open |
Jun 26 2025 | 6 months grace period start (w surcharge) |
Dec 26 2025 | patent expiry (for year 8) |
Dec 26 2027 | 2 years to revive unintentionally abandoned end. (for year 8) |
Dec 26 2028 | 12 years fee payment window open |
Jun 26 2029 | 6 months grace period start (w surcharge) |
Dec 26 2029 | patent expiry (for year 12) |
Dec 26 2031 | 2 years to revive unintentionally abandoned end. (for year 12) |