Dicom correction Proposal




Yüklə 56.97 Kb.
tarix10.04.2016
ölçüsü56.97 Kb.
DICOM Correction Proposal

STATUS

Final Text

Date of Last Update

2014/01/07

Person Assigned

David Clunie mailto:dclunie@dclunie.com

Submitter Name

Florian Knicker florian.knicker@eu.panasonic.com

Submission Date

2013/03/18



Correction Number CP-1304

Log Summary: Various corrections related to MPEG-4 AVC/H.264 Transfer Syntax

Name of Standard

PS 3.3, PS 3.4, PS 3.5, PS 3.17 2011



Rationale for Correction:

Sup 149 added the MPEG-4 AVC/H.264 Transfer Syntaxes to PS 3.5. This CP adds clarifications on the usage and missing references for MPEG-4 AVC/H.264.



Correction Wording:

PS 3.3 Section 2: Add reference to MPEG-4 AVC/H.264

2 Normative references


ISO/IEC 13818-4:2004 Information technology -- Generic coding of moving pictures and associated audio information -- Part 4: Conformance testing



ISO/IEC 14496-10:2009 Information technology – Coding of audio-visual objects – Part 10: Advanced Video Coding

ISO 15076-1:2005 Image technology colour management — Architecture, profile format, and data structure

Note: Also available as ICC.1:2004-10 (Profile version 4.2.0.0), International Color Consortium.

PS 3.3 Annex A: Add MPEG-4 AVC/H.264 color sampling definition

A.8.5.4 Multi-frame True Color SC Image IOD Content Constraints


The VOI LUT module shall not be present.

In the Image Pixel Module, the following constraints apply:

- Samples per Pixel (0028,0002) shall be 3

- Photometric Interpretation (0028,0004) shall be RGB for uncompressed or lossless compressed transfer syntaxes that do not involve color space transformations, YBR_ICT for irreversible JPEG 2000 transfer syntaxes, YBR_RCT for reversible JPEG 2000 transfer syntaxes, YBR_PARTIAL_420 for MPEG2 and MPEG-4 AVC/H.264 transfer syntaxes and YBR_FULL_422 for other lossy compressed transfer syntaxes

Note: Future lossless and lossy transfer syntaxes may lead to the need for new definitions and choices for Photometric Interpretation.


A.36.4.3.1 Enhanced MR Color Image IOD Content Constraints

The General Image Module, Overlay Plane Module and VOI LUT Module shall not be used in a Standard Extended SOP Class of the Enhanced MR Color Image.

Note: In order to annotate images, whether during acquisition or subsequently, SOP Instances of the Color Softcopy Presentation State Storage or the Structured Report Storage SOP Classes that reference the image SOP Instance, may be used.

No standard mechanism is provided for inclusion of annotations within the image SOP Instance itself, and implementers are discouraged from using private extensions to circumvent this restriction.

Color Softcopy Presentation State Storage Instances that are generated during acquisition may be referenced from the Image SOP Instance by using the Referenced Presentation State Sequence in the MR Image and Spectroscopy Instance Macro invoked from the Enhanced MR Image Module. See C.8.13.2.

The Photometric Interpretation (0028,0004) shall be RGB for uncompressed or lossless compressed transfer syntaxes that do not involve color space transformations, YBR_ICT for irreversible JPEG2000 transfer syntaxes, YBR_RCT for reversible JPEG2000 transfer syntaxes, YBR_PARTIAL_420 for MPEG2 and MPEG-4 AVC/H.264 transfer syntaxes and YBR_FULL_422 for other lossy compressed transfer syntaxes.

PS 3.3 Annex C: Clarify MPEG-2 reference

C.7.6.5 Cine Module


Table C.7-13 specifies the Attributes of a Multi-frame Cine Image.

Table C.7-13
CINE MODULE ATTRIBUTES


Attribute Name

Tag

Type

Attribute Description

...

...

...

...

Actual Frame Duration

(0018,1242)

3

Elapsed time of data acquisition in msec per each frame.

Multiplexed Audio Channels Description Code Sequence

(003A,0300)

2C

Description of any multiplexed audio channels. See Section C.7.6.5.1.3.

Zero or more items may be included in this sequence.

Required if the Transfer Syntax used to encode the multi-frame image contains multiplexed (interleaved) audio channels, such as is possible with MPEG2 Systems (see ISO/IEC 13818-1:2000).


>Channel Identification Code

(003A,0301)

1

A reference to the audio channel as identified within Transfer Syntax encoded bit stream (1 for the main channel, 2 for the second channel and 3 to 9 to the complementary channels).

...

...

...

...

PS 3.4 Annex Y: Add key frames note for MPEG-4 AVC/H.264
Y.3.2.1.3 Time Range

The range may be expanded as a consequence of the format in which the information is stored. Where such expansion occurs, any embedded audio data shall be similarly selected. Under all circumstances, the returned Composite SOP Instance shall retain the relationship between image and audio data.

Note: For MPEG-2 and MPEG-4 AVC/H.264 this would be to the nearest surrounding Key Frames

For JPEG 2000 Part 2, this would be to nearest surrounding precinct or tile boundary



PS 3.5 Section 2: Add references to Audio Formats and MP4 container formats

ISO/IS 14495-1 Lossless and near-lossless coding of continuous tone still images (JPEG-LS)



ISO/IEC 14496-3:2009 Information technology -- Coding of audio-visual objects -- Part 3: Audio

ISO/IEC 14496-10:2009 Information technology – Coding of audio-visual objects – Part 10: Advanced Video Coding



ISO/IEC 14496-12:2003 Information technology – Coding of audio-visual objects – Part 12: ISO base media file format

ISO/IEC 14496-14:2003 Information technology – Coding of audio-visual objects – Part 14: MP4 file format

ISO/IEC 15444-1 JPEG 2000 Image Coding System

RFC 1554 ISO-2022-JP-2: Multilingual Extension of ISO-2022-JP



Note: RFC 1554 is available from “http://www.ietf.org/rfc/rfc1554.txt”.

ISO/IEC 11172-3:1993 Information technology -- Coding of moving pictures and associated audio for digital storage media at up to about 1,5 Mbit/s -- Part 3: Audio

ISO/IEC 13818-1:2000 Information technology -- Generic coding of moving pictures and associated audio information: Systems

ISO/IEC 13818-4:1998 Information technology -- Generic coding of moving pictures and associated audio information -- Part 4: Conformance testing



ISO/IEC 13818-7:1997 Information technology -- Generic coding of moving pictures and associated audio information -- Part 7: Advanced Audio Coding (AAC)

ISO/IEC 15444-9 Information technology -- JPEG 2000 image coding system: Interactivity tools, APIs and protocols



PS 3.5 Section 8: Correct typo in reference to MPEG-2 and MPEG-4 AVC/H.264

8.2.5 MPEG2 MP@ML IMAGE COMPRESSION


The Basic Offset Table shall be empty (present but zero length).

Note: The Basic Offset Table is not used because MPEG2 contains its own mechanism for describing navigation of frames. To enable decoding of only a part of the sequence, MPEG2 manages a header in any group of pictures (GOP) containing a time_code – a 25-bit integer containing the following: drop_frame_flag, time_code_hours, time_code_minutes, marker_bit, time_code_seconds and time_code_pictures.

The container format for the video bitstream shall be MPEG-2 Transport Stream, a.k.a. MPEG-TS (see ISO/IEC 13818-1) or MPEG-4, a.k.a. MP4 container (see ISO/ENC 14496-12 and ISO/ENC 14496-14).

Any audio components present within the MPEG bit stream shall comply with the following restrictions:

- CBR MPEG-1 LAYER III (MP3) Audio Standard

- up to 24 bits

- 32 kHz, 44.1 kHz or 48 kHz for the main channel (the complementary channels can be sampled at the half rate, as defined in the Standard)

- one main mono or stereo channel, and optionally one or more complementary channel(s)

Notes : 1. MPEG-1 Layer III is standardized in Part 3 of the MPEG-1 standard (see ISO/IEC 11172-3)
2.
Although MPEG describes each channel as including up to 5 signals (e.g. for surround effects), it is recommended to limit each of the two channels to 2 signals each one (stereo).

8.2.6 MPEG2 MP@HL IMAGE COMPRESSION


One fragment shall contain the whole MPEG2 bit stream.

Notes: 1. If a video stream exceeds the maximum length of one fragment (approximately 4 GB), it may be sent as multiple SOP Instances.

2. This constraint limits the length of the compressed bit stream to no longer than 232-2 bytes.

The Basic Offset Table in the Pixel Data (7FE0,0010) shall be empty (present but zero length).

Note: The Basic Offset Table is not used because MPEG2 contains its own mechanism for describing navigation of frames. To enable decoding of only a part of the sequence, MPEG2 manages a header in any group of pictures (GOP) containing a time_code – a 25-bit integer containing the following: drop_frame_flag, time_code_hours, time_code_minutes, marker_bit, time_code_seconds and time_code_pictures.



The container format for the video bitstream shall be MPEG-2 Transport Stream, a.k.a. MPEG-TS (see ISO/IEC 13818-1) or MPEG-4, a.k.a. MP4 container (see ISO/ENC 14496-12 and ISO/ENC 14496-14).

Any audio components present within the MPEG2 MP@HL bit stream shall comply with the restrictions as for MPEG2 MP@ML as stated in Section 8.2.5.


8.2.7 MPEG-4 AVC/H.264 High Profile / Level 4.1 Video Compression


The requirements are:

The possible values for Rows (0028,0010), Columns (0028,0011), Cine Rate (0018,0040) , and Frame Time (0018,1063) or Frame Time Vector (0018,1065) depend on the used transfer syntax.



    • For MPEG-4 AVC/H.264 High Profile / Level 4.1 transfer syntax, the values for these data elements shall be compliant with the High Profile / Level 4.1 of the MPEG-4 AVC/H.264 standard (ISO/IEC 14496-10:2009) and restricted to a square pixel aspect ratio.



Table 8-5
MPEG-4 AVC/H.264 High Profile / Level 4.1 IMAGE TRANSFER SYNTAX FRAME RATE ATTRIBUTES


Video Type

Spatial resolution layer

Frame Rate (see Note 23)

Frame Time (see Note 34)

30 Hz HD

Single level, Enhancement

30

33.33 ms

25 Hz HD

Single level, Enhancement

25

40.0 ms

60 Hz HD

Single level, Enhancement

60

16.17 ms

50 Hz HD

Single level, Enhancement

50

20.00 ms

One fragment shall contain the whole MPEG-4 AVC/H.264 bit stream.

Note: If a video stream exceeds the maximum length of one fragment (approximately 4 GB), it may be sent as multiple SOP Instances.



The container format for the video bitstream shall be MPEG-2 Transport Stream (see ISO/IEC 13818-1) or MPEG-4, a.k.a. MP4 container (see ISO/ENC 14496-12 and ISO/ENC 14496-14). The PTS/DTS of the transport stream shall be used in the MPEG coding. Any audio components present within the bit stream shall be interleaved in either LPCM, or AC-3, AAC, MP3 or MPEG-1 Layer II audio format and shall comply with the following restrictions:

Tabelle: Table 8-x
Allowed audio formats


Audio Format

MPEG-2 TS container

MP4 container

LPCM

Allowed

-

AC-3

Allowed

-

AAC

Allowed

Allowed

MP3

Allowed

Allowed

MPEG-1 Audio Layer II

Allowed

Allowed




  • LPCM

    • Maximum bitrate: 4.608 Mbps

    • Sampling frequency: 48, 96 kHz

    • Bits per sample: 16, 20 or 24 bits

    • Number of channels: 2 channels

Note: If LPCM is used for Audio components, the container format shall be MPEG-2 TS.

  • AC-3

    • Maximum bitrate: 640kbps

    • Sampling frequency: 48kHz

    • Bits per sample: 16 bits

    • Number of channels: 2 or 5.1 channels

Notes: 1. AC-3 is standardized in ETSI TS 102 366

2. If AC-3 is used for Audio components, the container format shall be MPEG-2 TS.


  • AAC

    • Maximum bitrate: 640kbps

    • Sampling frequency: 48kHz

    • Bits per sample: 16, 20 or 24 bits

    • Number of channels: 2 or 5.1 channels

Note: AAC is standardized in Part 7 of the MPEG-2 standard (see ISO/IEC 13818-7, and Subpart 4 in Part 3 of the MPEG-4 standard (see ISO/IEC 14496-3).

  • CBR MPEG-1 LAYER III (MP3) Audio Standard

    • Maximum bitrate: 320 bits

    • Sampling frequency: 32 kHz, 44.1 kHz or 48 kHz for the main channel (the complementary channels can be sampled at the half rate, as defined in the Standard)

    • Bits per sample: up to 24 bits

    • Number of channels: one main mono or stereo channel, and optionally one or more complementary channel(s)

Notes: 1. MPEG-1 Layer III is standardized in Part 3 of the MPEG-1 standard (see ISO/IEC 11172-3)
2. Although MPEG describes each channel as including up to 5 signals (e.g. for surround effects), it is recommended to limit each of the two channels to 2 signals each one (stereo).


  • MPEG-1 LAYER II (MP2)

    • Maximum bitrate: 384kbps

    • Sampling frequency: 32 kHz, 44.1 kHz or 48 kHz

    • Bits per sample: up to 24 bits

    • Number of channels: 2

Note: MPEG-1 Layer II is standardized in Part 3 of the MPEG-1 standard (see ISO/IEC 11172-3)

PS 3.17 Annex KK: Add MPEG-4 AVC/H.264 to Retrieval based on report references

KK.1.1 Retrieval based on report references

Features of interest in many long “video” examinations (e.g. endoscopy) are commonly referenced as times from the start of the examination. The same benefits of reduced WAN bandwidth use could be obtained by shortening the MPEG-2, MPEG-4 AVC/H.264, or JPEG 2000 Part 2 Multi-component based stream prior to transmission.



PS 3.17 Annex LL: Add MPEG-4 AVC/H.264 to RETRIEVAL OF SELECTED FRAME COMPOSITE INSTANCES FROM MPEG-2 VIDEO

LL.3 Retrieval of SELECTED FRAME Composite instances from MPEG-2 OR MPEG-4 AVC/H.264 Video


Features of interest in many long “video” examinations (e.g. endoscopy) are commonly referenced as

If the image has been stored in MPEG-2 or MPEG-4 AVC/H.264 format, and if the SCU has knowledge independent of DICOM as to which section of a “video” is required for viewing (e.g. perhaps notes from an endoscopy) then the SCU can perform the following steps:


  1. Use known configuration information to identify the available transfer syntaxes

  2. If MPEG-2, MPEG-4 AVC/H.264, or JPEG 2000 Part 2 Multi-component transfer syntaxes are available, then issue a request to retrieve the required section.

The data received may be slightly longer than that requested, depending on the position of key frames in the data.

PS 3.17 Annex MM: Add MPEG-4 AVC/H.264

MM.2.5 MPEG-4 AVC/H.264 and JPEG 2000 Part 2 Multi-Component Transform


Similar issues exist as for MPEG-2 data and similar solutions apply.


MM.2.8 A “Single-Frame” Multi-Frame Image


Similar issues exist as for MPEG-2 data and similar solutions apply.

However, a single frame may well cause problems for a transfer syntax based on “video” such as those using MPEG-2 or MPEG-4 H.264/AVC, and therefore the SCU when negotiating a C-GET should consider this problem, and include one or more transfer syntaxes suitable for holding single or non-contiguous frames where such a retrieval request is being made.




MM.6 Audio


Where the original object is MPEG-2 or MPEG-4 AVC/H.264 video with interleaved audio data in the MPEG-2 dataSystem stream, and where the retrieved object is also MPEG-2 or MPEG-4 AVC/H.264 encoded, then audio could normally be preserved and maintain synchronization, but in other cases, the audio may be lost.

Page




Verilənlər bazası müəlliflik hüququ ilə müdafiə olunur ©azrefs.org 2016
rəhbərliyinə müraciət

    Ana səhifə