Professional Documents
Culture Documents
姓名或名称 联合专利有限责任公
用户代码 电话
司
③无效宣告请求人
居民身份证件号码或统一社会信用代码/组织机构代码
电子邮箱
国籍或注册国家(地区) 美国 邮政编码
经常居所地或营业所所在地 美国 华盛顿
省、自治区、直辖市 市县
姓名 郭婧婧 电话 021-51096606
④
电子邮箱 jzmc@iprtop.com 邮政编码 200041
收
件
省、自治区、直辖市 上海市 市县 静安区
人
城区(乡)、街道、门牌号 南京西路 580 号仲益大厦 32 楼 C 座
⑤
名称 上海光华专利事务所(普通合伙) 代码 31219
专
利 姓 名 李艳春 姓 名 郭婧婧
代 代 代
理 理 执业证号 3121933782.5 理 执业证号 3121913718.1
机 人 人
(1) 电 话 021-51096606-850 (2) 电 话 021-51096606-852
构
⑦无效宣告请求的理由、范围及所依据的证据
理 由 范 围 依据的证据
专利法第 三十三 条,第 0 款 权利要求 1-4 审查过程文件
实施细则第 0 条,第 0 款
1
101001
2019.4
专 利 权 无 效 宣 告 请 求 书
由于意见陈述内容过多,请见附件,谢谢。
⑨附件清单:
【附件名称】发明专利无效宣告请求书正文 【附件属性】 电子件
【附件名称】审查过程文件 【附件属性】 电子件
【附件名称】涉案专利-CN101035291 【附件属性】 电子件
【附件名称】证据 1-CN1327345A 【附件属性】 电子件
【附件名称】证据 2-US2001040700A1 【附件属性】 电子件
【附件名称】证据 3-T-REC-H.263-1998 【附件属性】 电子件
【附件名称】证据 2 和 3 译文 【附件属性】 电子件
【附件名称】公司主体证明 【附件属性】 电子件
【附件名称】无效宣告程序授权委托书 【附件属性】 电子件
⑩无效宣告请求人或专利代理机构签字或者盖章 11 国家知识产权局处理意见
○
上海光华专利事务所(普通合伙)
2021 年 05 月 13 日 年 月 日
2
101001
2019.4
专 利 权 无 效 宣 告 程 序 授 权 委 托 书
专 利 号 200710097060X 案件编号
发明创造名称 动态图像编码方法及动态图像解码方法
无效宣告请求人 联合专利有限责任公司
专 利 权 人 知识产权之桥一号有限责任公司
委托人:
姓名或名称 联合专利有限责任公司 电话 +1
650-999-0889
通 信 地 址 美 国 华 盛 顿 特 区 西 北 康 涅 狄 格 大 道 1875 号 10 层
邮编 20009
被委托人:
专利代理机构名称 上海光华专利事务所(普通合伙) 代
码 31219
通 信 地 址 上 海 市 南 京 西 路 580 号 仲 益 大 厦 32 楼 C 座
邮编 200041
现委托上列被委托人指定的代理人在上述专利的专利权无效宣告程序中为我方代理人,其委托权限仅
限于办理无效宣告程序有关事务。
其中:
委托人(签章) 被委托人(签章)
1
101003
2019.4
专 利 权 无 效 宣 告 程 序 授 权 委 托 书
联合专利有限责任公司 上海光华专
利事务所(普通合伙)
2021 年 05 月 13 日 2021 年 05 月 13 日
重要提示:在填写本授权委托书前,请务必仔细阅读后面的注意事项。
【此处插入专利权无效宣告程序授权委托书扫描文件】
【图片描述】 无效宣告程序授权委托书
2
101003
2019.4
宣告 200710097060.X 号发明专利无效的事实和理由
(发明专利无效宣告请求书正文)
请求人呈交本请求书。
根据《中华人民共和国专利法》及《中华人民共和国专利法实施细则》的有
关规定,联合专利有限责任公司(以下简称“请求人”
)请求宣告中华人民共和
国知识产权局专利局(以下简称“专利局”) 2011 年 05 月 04 日公告授权的、
名称为“动态图像编码方法及动态图像解码方法”的第 200710097060.X 号发明
专利(以下简称“涉案专利”)无效。
请求人委托上海光华专利事务所(普通合伙)呈交本请求书,并同时请求口
审。
一、 第 200710097060.X 号专利的基本情况
附件 1 是涉案专利授权公告的发明专利说明书复印件。涉案专利的专利权人
为“知识产权之桥一号有限责任公司”,涉案专利名称为“动态图像编码方法及
动态图像解码方法”,申请日为 2003 年 4 月 16 日,优先权日为 2002 年 4 月 19
日,授权公告日为 2011 年 5 月 04 日。涉案专利为申请号为 03800473.9 的中国
专利的分案申请。涉案专利的权利要求共 4 项,其中第 1、3 项为独立权利要求。
二、 法律依据
专利法第二十二条第 3 款规定:
“创造性,是指与现有技术相比,该发明具有突出的实质性特点和显著的进
步,该实用新型具有实质性特点和进步。”
1
第三十三条 申请人可以对其专利申请文件进行修改,但是,对发明和实用
新型专利申请文件的修改不得超出原说明书和权利要求书记载的范围,对外观设
计专利申请文件的修改不得超出原图片或者照片表示的范围。
第二十六条第四款规定:
“权利要求书应当以说明书为依据,清楚、简要地限定要求专利保护的范
围。”
第四十五条 自国务院专利行政部门公告授予专利权之日起,任何单位或者
个人认为该专利权的授予不符合本法有关规定的,可以请求专利复审委员会宣告
该专利权无效。
三、 证据
证据 1 CN1327345A 公开日:2001 年 12 月 19 日
证据 2 US2001/0040700A1 公开日:2001 年 11 月 15 日
证据 3 ITU-T H.263, 低比特率通信的视频 公开日:1998 年 2 月 28 日
编 码 ( Video coding for low bit rate (出版月份的最后一日)
communication)
四、 事实与理由
(一)专利权人在实质审查过程中对涉案专利权利要求的修改超出了原说
明书和权利要求书记载的范围,不符合《专利法》第三十三条的规定。
1. 涉案专利对权利要求 1 的修改超出了原说明书和权利要求书记载的范
围,不符合《专利法》第三十三条的规定。
涉案专利的公开文本中的权利要求 1 如下:
一种图像解码方法,其基于像块从多个参照图像中选择一个参照图像,并在
待解码的当前图像中的像块上执行预测解码,所述方法包括:
当对由多个像块构成的多像块图像单元解码时,判断在所述多像块图像
单元的共同信息区域中是否描述了用于识别用于共同参照的共同参照图像的
2
信息;
在判断所述共同信息区域中描述了用于识别所述共同参照图像的所述信
息的情况下,使用所述共同参照图像,生成所述多像块图像单元中包含的当
前像块的预测图像;
在判断所述共同信息区域中没有描述用于识别所述共同参照图像的所述
信息的情况下,使用基于像块指定的所述参照图像,生成所述多像块图像单
元中包含的当前像块的预测图像;和
使用所述预测图像对所述当前像块解码。
涉案专利的授权公告文本中的权利要求 1 如下:
一种图像解码方法,其基于像块从多个参照图像中选择一个参照图像,并在
待解码的当前图像中的、在编码的动态图像数据中包含的像块上执行预测解码,
所述方法包括:
当对由所述编码的动态图像数据中包含的多个像块构成的多像块图像单元
解码时,在(a)所述编码的动态图像数据中包含的,以及(b)位于所述多像块图像
单元前的共同信息区域中描述了用于识别待被共同参照的共同参照图像的信息
的情况下,使用所述共同参照图像,生成所述多像块图像单元中包含的当前像块
的预测图像;
在基于像块识别了针对包含在所述多像块图像单元中的所述当前像块识别
的参照图像的情况下,使用所述参照图像,生成所述当前像块的预测图像;和
使用所述预测图像对所述当前像块解码。
相比公开文本,授权公告的权利要求 1 存在如下修改:
一种图像解码方法,其基于像块从多个参照图像中选择一个参照图像,并在
待解码的当前图像中的、在编码的动态图像数据中包含的像块上执行预测解码,
所述方法包括:
当对由所述编码的动态图像数据中包含的多个像块构成的多像块图像单
元解码时,在(a)所述编码的动态图像数据中包含的,以及(b)位于所述多像块
图像单元前的共同信息区域中描述了用于识别待被共同参照的共同参照图像
的信息的情况下,使用所述共同参照图像,生成所述多像块图像单元中包含
的当前像块的预测图像;
3
在基于像块识别了针对包含在所述多像块图像单元中的所述当前像块识别
的参照图像的情况下,使用所述参照图像,生成所述当前像块的预测图像;和
使用所述预测图像对所述当前像块解码。
申请人在答复第二次审查意见通知书时对涉案专利的权利要求 1 进行了修
改,修改后的权利要求 1 包括特征“(b)位于所述多像块图像单元前的共同信息
区域中描述了用于识别待被共同参照的共同参照图像的信息的情况下”。
根据原说明书和权利要求书记载的内容(说明书第[0039]段),共同信息区
域位于动态图像编码数据中的多个像块内。原说明书和权利要求书均未公开共同
信息区域位于所述多像块图像单元前,而且申请人给出的原说明书第 14 页倒数
第 2 段至第 15 页第 4 段也未公开该内容,因此专利权人在实质审查过程中对涉
案专利权利要求的修改超出了原说明书和权利要求书记载的范围,不符合《专利
法》第三十三条的规定。因此,请求人恳请宣告涉案专利的独立权利要求 1,及
其相应的从属权利要求 2 无效。
2. 涉案专利对权利要求 3 的修改超出了原说明书和权利要求书记载的范
围,不符合《专利法》第三十三条的规定。
涉案专利的公开文本中的权利要求 3 如下:
一种图像解码装置,其基于像块从多个参照图像中选择一个参照图像,并在
待解码的当前图像中的像块上执行预测解码,所述装置包括:
进行判断操作的单元,当对由多个像块构成的多像块图像单元解码时,判断
在所述多像块图像单元的共同信息区域中是否描述了用于识别用于共同参照的
共同参照图像的信息;
进行生成操作的单元,在判断所述共同信息区域中描述了用于识别所述共同
参照图像的所述信息的情况下,使用所述共同参照图像,生成所述多像块图像单
元中包含的当前像块的预测图像;
进行生成操作的单元,在判断所述共同信息区域中没有描述用于识别所述共
同参照图像的所述信息的情况下,使用基于像块指定的所述参照图像,生成所述
多像块图像单元中包含的当前像块的预测图像;以及
4
进行解码操作的单元,使用所述预测图像对所述当前像块解码。
涉案专利的授权公告文本中的权利要求 3 如下:
一种图像解码装置,其基于像块从多个参照图像中选择一个参照图像,并在
待解码的当前图像中的、在编码的动态图像数据中包含的像块上执行预测解码,
所述装置包括:
进行生成操作的单元,i)当对由所述编码的动态图像数据中包含的多个像块
构成的多像块图像单元解码时,在(a)所述编码的动态图像数据中包含的,以及(b)
位于所述多像块图像单元前的共同信息区域中描述了用于识别待被共同参照的
共同参照图像的信息的情况下,使用所述共同参照图像,生成所述多像块图像单
元中包含的当前像块的预测图像;以及 ii)在基于像块识别了针对包含在所述多
像块图像单元中的所述当前像块识别的参照图像的情况下,使用所述参照图像,
生成所述当前像块的预测图像;以及
进行解码操作的单元,使用所述预测图像对所述当前像块解码。
相比公开文本,授权公告的权利要求 3 存在如下修改:
一种图像解码装置,其基于像块从多个参照图像中选择一个参照图像,并在
待解码的当前图像中的、在编码的动态图像数据中包含的像块上执行预测解码,
所述装置包括:
进行生成操作的单元,i)当对由所述编码的动态图像数据中包含的多个像块
构成的多像块图像单元解码时,在(a)所述编码的动态图像数据中包含的,以及(b)
位于所述多像块图像单元前的共同信息区域中描述了用于识别待被共同参照的
共同参照图像的信息的情况下,使用所述共同参照图像,生成所述多像块图像单
元中包含的当前像块的预测图像;以及 ii)在基于像块识别了针对包含在所述多
像块图像单元中的所述当前像块识别的参照图像的情况下,使用所述参照图像,
生成所述当前像块的预测图像;以及
进行解码操作的单元,使用所述预测图像对所述当前像块解码。
申请人在答复第二次审查意见通知书时对涉案专利的权利要求 3 进行了修
改,修改后的权利要求 3 包括特征“(b)位于所述多像块图像单元前的共同信息
区域中描述了用于识别待被共同参照的共同参照图像的信息的情况下”。
根据原说明书和权利要求书记载的内容(说明书第[0039]段),共同信息区
5
域位于动态图像编码数据中的多个像块内。原说明书和权利要求书均未公开共同
信息区域位于所述多像块图像单元前,而且申请人给出的原说明书第 14 页倒数
第 2 段至第 15 页第 4 段也未公开该内容,因此专利权人在实质审查过程中对涉
案专利权利要求的修改超出了原说明书和权利要求书记载的范围,不符合《专利
法》第三十三条的规定。因此,请求人恳请宣告涉案专利的独立权利要求 3,及
其相应的从属权利要求 4 无效。
即便不考虑修改超范围的问题,将特征“(b)位于所述多像块图像单元前的
共同信息区域中描述了用于识别待被共同参照的共同参照图像的信息的情况下”
根据说明书中的内容解读为“共同信息区域位于所述多像块图像单元前的某个位
置(比如信头),在共同信息区域中描述了用于识别待被共同参照的共同参照图
像的信息的情况下”,基于下面具体陈述的理由(三)和(四),涉案专利的权利
要求 1 和权利要求 3 也不具备创造性。
6
涉案专利权利要求 3 中的技术特征“(b)位于所述多像块图像单元前的共同
信息区域中描述了用于识别待被共同参照的共同参照图像的信息的情况下”的表
述是不清楚的,从而导致涉案专利权利要求 3 的保护范围不清楚。
对于“位于所述多像块图像单元前的共同信息区域”,本领域技术人员不清
楚共同信息区域如何位于多像块图像单元前,即本领域技术人员不能确定共同信
息区域是位于多像块图像单元前面的某个位置,还是在时间上先于多像块图像单
元。因此该表述也是不清楚的,从而导致涉案专利权利要求 3 的保护范围不清楚,
不符合《专利法》第二十六条第四款的规定。因此,请求人恳请宣告涉案专利的
独立权利要求 3,及其相应的从属权利要求 4 无效。
7
的图像块译码方法,包括一个译码器,它提供运动矢量及要重新组成的图像与一
个估计图像之间的差分,一个运动补偿预测器,它接收运动矢量及一个基准图像
并提供一个估计图像,一个基准图像存储器,存储至少两个基准图像并向所述预
测器提供所述基准图像的一个,及一个块效应过滤器,用于在基准图像存储在基
准图像存储中之前过滤基准图像中的块效应。
(相当于当对由编码的动态图像数
据中包含的多个像块构成的多像块图像单元解码时)
证据 1 公开了(说明书第 2 页第 3 段)在编码一个图像时使用多个基准图像,
它们经过块效应过滤,从所述基准图像中选择一个基准图像,选择用于预测编码
的存储图像被发送到译码器,它使用适当的存储图像作为从编码器接收的指令函
数。证据 1 还公开了(说明书第 1 页倒数第 2 段)译码处理使用与重新组成每个
图像相反的步骤。证据 1 还公开了(说明书第 5 页倒数第 2 段)译码器选择用于
使用开关解码接收到的图像的参考图像,该开关可以被作为从图 1 的编码系统通
过输入信道接收的信息的函数来控制。因此由于选择的参考图像被发送至译码
器,本领域技术人员将知晓证据 1 的发送至译码器应当包括识别参考图像的信息
(相当于在描述了用于识别待被共同参照的共同参照图像的信息的情况下)。
如图 2 所示,证据 1 公开了(说明书第 3 页第 1 段)使用预测器为使用选定
的共同参考图像译码的宏块生成预测图像。而且证据 1 公开了(说明书第 3 页第
1 段)包括运动补偿预测器的译码器。运动估计器提供相对于在前图像的一个图
像的诸块或诸宏块的运动估计。为了估计目的,预测器还在其输入端从一个基准
图像存储器接收一个基准图像。也即是说,预测器还在其输入端接收来自图像的
宏块和选定的共同参考图像的运动矢量。证据 1 公开了(说明书第 2 页倒数第 4
段)运动补偿预测器接收运动矢量及一个基准图像,并且提供一个估计图像。证
据 1 还公开了(说明书第 4 页第 3 段)运动补偿预测器提供基于基准图像和运动
矢量的估计图像。也即是说,使用这些输入,预测器为每个宏块生成预测图像。
(相当于使用所述共同参照图像,生成所述多像块图像单元中包含的当前像块的
预测图像)
证据 1 公开了(说明书第 2 页第 3 段)在编码一个图像时使用多个基准图像,
它们经过块效应过滤,从所述基准图像中选择一个基准图像,选择用于预测编码
的存储图像被发送到译码器,它使用适当的存储图像作为从编码器接收的指令函
8
数。证据 1 还公开了(说明书第 1 页倒数第 2 段)译码处理使用与重新组成每个
图像相反的步骤。证据 1 还公开了(说明书第 5 页倒数第 2 段)译码器选择用于
使用开关解码接收到的图像的参考图像,该开关可以被作为从图 1 的编码系统通
过输入信道接收的信息的函数来控制。因此由于选择的参考图像被发送至译码
器,本领域技术人员将知晓证据 1 的发送至译码器应当包括识别参考图像的信息
(相当于在基于像块识别了针对包含在所述多像块图像单元中的所述当前像块
识别的参照图像的情况下)。
如图 2 所示,证据 1 公开了(说明书第 3 页第 1 段)使用预测器为使用选定
的共同参考图像译码的宏块生成预测图像。而且证据 1 公开了(说明书第 3 页第
1 段)包括运动补偿预测器的译码器。运动估计器提供相对于在前图像的一个图
像的诸块或诸宏块的运动估计。为了估计目的,预测器还在其输入端从一个基准
图像存储器接收一个基准图像。也即是说,预测器还在其输入端接收来自图像的
宏块和选定的共同参考图像的运动矢量。证据 1 公开了(说明书第 2 页倒数第 4
段)运动补偿预测器接收运动矢量及一个基准图像,并且提供一个估计图像。证
据 1 还公开了(说明书第 4 页第 3 段)运动补偿预测器提供基于基准图像和运动
矢量的估计图像。也即是说,使用这些输入,预测器为每个宏块生成预测图像。
(相当于使用所述参照图像,生成所述当前像块的预测图像)
证据 1 公开了(说明书第 5 页第 3 段)该译码系统接收在输入缓冲器 30 中
以低的比特率通过信道发送的信息。所接收的信息被传送给可变长度译码器 32。
可变长度译码器 32 提供运动矢量,它们被馈入运动补偿预测器 34,并且将输入
图像与估计图像之间的差分馈入量化反相器 36,并且馈入反相器 38 中。
证据 1 还公开了(说明书第 5 页倒数第 1 段)加法器 42 接收由运动补偿预
测器提供的估计图像及由反相器 38 提供的估计图像与参考图像的差分。它在其
输出端产生重新组成的图像。也即是说,可变长度译码器使用来自预测器的估计
图像输出进行译码。(相当于使用所述预测图像对所述当前像块解码)
证据 1 公开了特征“以及(b)位于所述多像块图像单元前的共同信息区域
中描述了用于识别待被共同参照的共同参照图像的信息”。
证据 1 公开了(说明书第 2 页第 3 段)在编码一个图像时使用多个基准图像,
9
它们经过块效应过滤,从所述基准图像中选择一个基准图像,选择用于预测编码
的存储图像被发送到译码器,它使用适当的存储图像作为从编码器接收的指令函
数。证据 1 还公开了(说明书第 5 页倒数第 2 段)译码器选择用于使用开关解码
接收到的图像的参考图像,该开关可以被作为从图 1 的编码系统通过输入信道接
收的信息的函数来控制。因此由于选择的参考图像被发送至译码器,本领域技术
人员知晓证据 1 的发送至译码器应当包括识别参考图像的信息。
证据 1 公开了(说明书第 2 页第 6 段)图像块编码系统具有运动估计器,它
接收要被编码的图像及提供运动矢量;运动补偿预测器,它接收运动矢量及一个
基准图像,并且提供一个估计图像。在提供估计图像时,基准图像存储器存储至
少两个基准图像,并且将所述基准图像的一个提供给所述预测器。因此,对于待
编码的图像的每个块或宏块,因为运动矢量是基于选择的基准图像生成的(一个
图像的诸块或诸宏块的估计运动),候选基准图像中的一个(待被共同参照的共
同参照图像)被提供给预测器,本领域技术人员清楚证据 1 公开了从多个参照图
像中选择一个用于“一个图像的诸块或诸宏块”共同参照的共同参照图像。
并且,当此共同参照图像被传输至译码器时,共同参照图像的信息也被传输。
即,本领域技术人员可以理解,证据 1 至少隐含公开了在“多像块图像单元前的
共同信息区域中描述了用于识别待被共同参照的共同参照图像的信息”。
证据 1 公开了“在(a)所述编码的动态图像数据中包含的…共同信息区域
中描述了用于识别待被共同参照的共同参照图像的信息”。
根据前述内容,证据 1 公开了在多像块图像单元的共同信息区域中,描述用
于识别所选择的共同参照图像的共同信息。证据 1 公开了多像块图像单元包含在
编码的动态图像数据中。本领域技术人员清楚证据 1 中描述用于识别待被共同参
照的共同参照图像的信息包含在所述编码的动态图像数据中的共同信息区域中。
针对技术特征“以及(b)位于所述多像块图像单元前的共同信息区域中描
述了用于识别待被共同参照的共同参照图像的信息”,除了证据 1 公开的前述内
容之外,证据 2 也公开了该特征。
证据 2 公开了(说明书第[0019]段)“编码器可用这种指示符来指示解码器
哪些图像极为类似于当前参考图像,使得如果实际参考图像在传输过程中丢失,
10
则这些图像之一可用作备用参考图像”。证据 2 公开了(说明书第[0020]段)备
用参考图像编号通常包含在图像信头中,以识别解码器的备用参考图片。
证据 2 公开了(说明书第[0012]段)图像层的数据包含影响整个图像区以及
图像数据解码的参数。这种数据的绝大部分安排在所谓的图像信头中。证据 2
公开了“包含在图像信头中的备用参考图像编号是影响整个图像区以及图像数据
解码的参数”。
因此证据 2 公开了备用参考图像被编码器通过图像信头共同识别,如果实际
参考图像在传输过程中丢失,译码器使用识别的备用参考图像。本领域技术人员
知晓证据 2 公开了在图片的共同信息区域(备用参考图像编号,对应在被编码的
图像的信头中的位置)中,识别选择的共同参照图像(备用参考图像)。
本领域技术人员有动机使用证据 2 中的教导(备用参考图像编号包含在图像
信头中)来改进证据 1 中的共同参照图像。在 H.263 标准下,本领域技术人员有
动机去寻找关于参考图像选择的现有技术,以识别选择的共同参考图像。
证据 2 公开了(说明书第[0020]段)“在本发明的一种最佳实现中,视频信
号是根据 H.263 标准编码的,而指示符包含在附加增强信息中”。
证据 1 公开了(说明书第 3 页倒数第 2 段)图 1 提出使用块边缘过滤及在
H.263+标准中提出的预测块编码系统中的多个存储帧。
因此根据证据 2 公开的内容,本领域技术人员可以清楚地知道共同参照图像
可以在图像信头中被识别。证据 2 中的图像信头提供共同信息区域,用于存储证
据 1 中共同参照图像的标识符。即证据 1 和 2 的结合公开了特征“以及(b) 位于
多像块图像单元前的共同信息区域中描述了用于识别待被共同参照的共同参照
图像的信息”。
涉案专利权利要求 1 中的技术特征已被证据 1 和 2 所公开,并且在证据 1
和 2 中所起的作用与在涉案专利中起到的作用相同,都是确定共同信息区域的位
置。在此基础上,本领域技术人员容易想到将证据 2 与证据 1 相结合,从而得到
涉案专利权利要求 1 的技术方案。
综上所述,请求人认为,涉案专利的权利要求 1 与证据 2 和 1 的结合相比,
不具备突出的实质性特点和显著的进步,不符合专利法第 22 条第 3 款规定的创
造性。因此,请求人恳请宣告涉案专利的权利要求 1 无效。
11
2、涉案专利的权利要求 2 不具备创造性,不符合《专利法》第二十二条第
三款的规定。
涉案专利的权利要求 2 为:
如权利要求 1 所述的图像解码方法,其中,所述多像块图像单元是多个图像
单元,一个图像单元,片段单元和宏像块单元中的一个。
证据 1 公开了(说明书第 4 页第 1 段)对图像单元执行译码。运动估计器提
供对相对于在前图像的一个图像的诸块或诸宏块的运动估计。本领域技术人员知
晓证据 1 中的一个图像对应于权利要求 2 中的一个图像单元,证据 1 中的宏块对
应于权利要求 2 中的宏像块单元,因此证据 1 公开了多像块图像单元是一个图像
单元或宏像块单元,即公开了多像块图像单元是多个图像单元,一个图像单元,
片段单元和宏像块单元中的一个。
综上所述,请求人认为,涉案专利的权利要求 2 与证据 2 和 1 的结合相比,
不具备突出的实质性特点和显著的进步,不符合专利法第 22 条第 3 款规定的创
造性。因此,请求人恳请宣告涉案专利的权利要求 2 无效。
3、涉案专利的权利要求 3 不具备创造性,不符合《专利法》第二十二条第
三款的规定。
证据 1 公开了(说明书第 5 页第 3 段)一种具有运动预测的图像块译码系统
(相当于一种图像解码装置)。该系统从已经经过块效应过滤的多个已接收的基准
图像中所选择的基准图像重新组成一个发送图像。证据 1 还公开了(说明书第 3
页第 5 段)在译码时,从已经经过块效应过滤的所接收的多个基准图像中选择的
一个基准图像重新组成所发送的图像(相当于其基于像块从多个参照图像中选择
一个参照图像)。
证据 1 中公开了(说明书第 3 页第 5 段)在译码时,从已经经过块效应过滤
的所接收的多个基准图像中选择的一个基准图像重新组成所发送的图像,本领域
技术人员知晓从基准图像重新生成所发送的图像即是执行预测解码。而且证据 1
公开了(发明名称)涉及用运动预测和块效应过滤进行视频源编码,其中具有运
动预测的图像块译码方法包括译码器提供运动矢量,运动补偿预测器接收运动矢
12
量及基准图像并提供一个估计图像,即本领域技术人员知晓像块包含在编码的动
态图像数据中(相当于并在待解码的当前图像中的、在编码的动态图像数据中包
含的像块上执行预测解码,所述装置包括:)
证据 1 公开了(说明书第 3 页第 1 段)本发明进一步提供一种具有运动预测
的图像块译码方法,包括一个译码器,它提供运动矢量及要重新组成的图像与一
个估计图像之间的差分,一个运动补偿预测器,它接收运动矢量及一个基准图像
并提供一个估计图像,一个基准图像存储器,存储至少两个基准图像并向所述预
测器提供所述基准图像的一个,及一个块效应过滤器,用于在基准图像存储在基
准图像存储中之前过滤基准图像中的块效应。(相当于进行生成操作的单元,i)
当对由所述编码的动态图像数据中包含的多个像块构成的多像块图像单元解码
时)
证据 1 公开了(说明书第 2 页第 3 段)在编码一个图像时使用多个基准图像,
它们经过块效应过滤,从所述基准图像中选择一个基准图像,选择用于预测编码
的存储图像被发送到译码器,它使用适当的存储图像作为从编码器接收的指令函
数。证据 1 还公开了(说明书第 1 页倒数第 2 段)译码处理使用与重新组成每个
图像相反的步骤。证据 1 还公开了(说明书第 5 页倒数第 2 段)译码器选择用于
使用开关解码接收到的图像的参考图像,该开关可以被作为从图 1 的编码系统通
过输入信道接收的信息的函数来控制。因此由于选择的参考图像被发送至译码
器,本领域技术人员将知晓证据 1 的发送至译码器应当包括识别参考图像的信息
(相当于在…描述了用于识别待被共同参照的共同参照图像的信息的情况下)。
如图 2 所示,证据 1 公开了(说明书第 3 页第 1 段)使用预测器为使用选定
的共同参考图像译码的宏块生成预测图像。而且证据 1 公开了(说明书第 3 页第
1 段)包括运动补偿预测器的译码器。运动估计器提供相对于在前图像的一个图
像的诸块或诸宏块的运动估计。为了估计目的,预测器还在其输入端从一个基准
图像存储器接收一个基准图像。也即是说,预测器还在其输入端接收来自图像的
宏块和选定的共同参考图像的运动矢量。证据 1 公开了(说明书第 2 页倒数第 4
段)运动补偿预测器接收运动矢量及一个基准图像,并且提供一个估计图像。证
据 1 还公开了(说明书第 4 页第 3 段)运动补偿预测器提供基于基准图像和运动
矢量的估计图像。也即是说,使用这些输入,预测器为每个宏块生成预测图像。
13
(相当于使用所述共同参照图像,生成所述多像块图像单元中包含的当前像块的
预测图像)
证据 1 公开了(说明书第 2 页第 3 段)在编码一个图像时使用多个基准图像,
它们经过块效应过滤,从所述基准图像中选择一个基准图像,选择用于预测编码
的存储图像被发送到译码器,它使用适当的存储图像作为从编码器接收的指令函
数。证据 1 还公开了(说明书第 1 页倒数第 2 段)译码处理使用与重新组成每个
图像相反的步骤。证据 1 还公开了(说明书第 5 页倒数第 2 段)译码器选择用于
使用开关解码接收到的图像的参考图像,该开关可以被作为从图 1 的编码系统通
过输入信道接收的信息的函数来控制。因此由于选择的参考图像被发送至译码
器,本领域技术人员将知晓证据 1 的发送至译码器应当包括识别参考图像的信息
(相当于 ii)在基于像块识别了针对包含在所述多像块图像单元中的所述当前像
块识别的参照图像的情况下)。
如图 2 所示,证据 1 公开了(说明书第 3 页第 1 段)使用预测器为使用选定
的共同参考图像译码的宏块生成预测图像。而且证据 1 公开了(说明书第 3 页第
1 段)包括运动补偿预测器的译码器。运动估计器提供相对于在前图像的一个图
像的诸块或诸宏块的运动估计。为了估计目的,预测器还在其输入端从一个基准
图像存储器接收一个基准图像。也即是说,预测器还在其输入端接收来自图像的
宏块和选定的共同参考图像的运动矢量。证据 1 公开了(说明书第 2 页倒数第 4
段)运动补偿预测器接收运动矢量及一个基准图像,并且提供一个估计图像。证
据 1 还公开了(说明书第 4 页第 3 段)运动补偿预测器提供基于基准图像和运动
矢量的估计图像。也即是说,使用这些输入,预测器为每个宏块生成预测图像。
(相当于使用所述共同参照图像,生成所述多像块图像单元中包含的当前像块的
预测图像)
证据 1 公开了(说明书第 5 页第 3 段)该译码系统接收在输入缓冲器 30 中
以低的比特率通过信道发送的信息。所接收的信息被传送给可变长度译码器 32。
可变长度译码器 32 提供运动矢量,它们被馈入运动补偿预测器 34,并且将输入
图像与估计图像之间的差分馈入量化反相器 36,并且馈入反相器 38 中。
证据 1 还公开了(说明书第 5 页倒数第 1 段)加法器 42 接收由运动补偿预
测器提供的估计图像及由反相器 38 提供的估计图像与参考图像的差分。它在其
14
输出端产生重新组成的图像。也即是说,可变长度译码器使用来自预测器的估计
图像输出进行译码。
(相当于进行解码操作的单元,使用所述预测图像对所述当
前像块解码)
证据 1 公开了“以及(b)位于所述多像块图像单元前的共同信息区域中描
述了用于识别待被共同参照的共同参照图像的信息”。
证据 1 公开了(说明书第 2 页第 3 段)在编码一个图像时使用多个基准图像,
它们经过块效应过滤,从所述基准图像中选择一个基准图像,选择用于预测编码
的存储图像被发送到译码器,它使用适当的存储图像作为从编码器接收的指令函
数。证据 1 还公开了(说明书第 5 页倒数第 2 段)译码器选择用于使用开关解码
接收到的图像的参考图像,该开关可以被作为从图 1 的编码系统通过输入信道接
收的信息的函数来控制。因此由于选择的参考图像被发送至译码器,本领域技术
人员知晓证据 1 的发送至译码器应当包括识别参考图像的信息。
证据 1 公开了(说明书第 2 页第 6 段)图像块编码系统具有运动估计器,它
接收要被编码的图像及提供运动矢量;运动补偿预测器,它接收运动矢量及一个
基准图像,并且提供一个估计图像。在提供估计图像时,基准图像存储器存储至
少两个基准图像,并且将所述基准图像的一个提供给所述预测器。因此,对于待
编码的图像的每个块或宏块,因为运动矢量是基于选择的基准图像生成的(一个
图像的诸块或诸宏块的估计运动),候选基准图像中的一个(待被共同参照的共
同参照图像)被提供给预测器,本领域技术人员清楚证据 1 公开了从多个参照图
像中选择一个用于“一个图像的诸块或诸宏块”共同参照的共同参照图像。
并且,当此共同参照图像被传输至译码器时,共同参照图像的信息也被传输。
即,本领域技术人员可以理解,证据 1 至少隐含公开了“(b)位于所述多像块图
像单元前的共同信息区域中描述了用于识别待被共同参照的共同参照图像的信
息”。
证据 1 公开了“在(a)所述编码的动态图像数据中包含的…共同信息区域
中描述了用于识别待被共同参照的共同参照图像的信息”。
根据前述内容,证据 1 公开了在多像块图像单元的共同信息区域中,描述用
于识别所选择的共同参照图像的共同信息。证据 1 公开了多像块图像单元包含在
15
编码的动态图像数据中。本领域技术人员清楚证据 1 中描述用于识别待被共同参
照的共同参照图像的信息包含在所述编码的动态图像数据中的共同信息区域中。
针对技术特征“以及(b)位于所述多像块图像单元前的共同信息区域中描
述了用于识别待被共同参照的共同参照图像的信息”,除了证据 1 公开的前述内
容之外,证据 2 也公开了该特征。
证据 2 公开了(说明书第[0019]段)“编码器可用这种指示符来指示解码器
哪些图像极为类似于当前参考图像,使得如果实际参考图像在传输过程中丢失,
则这些图像之一可用作备用参考图像”。证据 2 公开了(说明书第[0020]段)备
用参考图像编号通常包含在图像信头中,以识别解码器的备用参考图片。本领域
技术人员清楚信头位于多像块图像单元的前面。
证据 2 公开了(说明书第[0012]段)图像层的数据包含影响整个图像区以及
图像数据解码的参数。这种数据的绝大部分安排在所谓的图像信头中。证据 2
公开了“包含在图像信头中的备用参考图像编号是影响整个图像区以及图像数据
解码的参数”。
因此证据 2 公开了备用参考图像被编码器通过图像信头共同识别,如果实际
参考图像在传输过程中丢失,译码器使用识别的备用参考图像。本领域技术人员
知晓证据 2 公开了在图片的共同信息区域(备用参考图像编号,对应在被编码的
图像的信头中的位置)中,识别选择的共同参照图像(备用参考图像)。
本领域技术人员有动机使用证据 2 中的教导(备用参考图像编号包含在图像
信头中)来改进证据 1 中的共同参照图像。在 H.263 标准下,本领域技术人员有
动机去寻找关于参考图像选择的现有技术,以识别选择的共同参考图像。
证据 2 公开了(说明书第[0020]段)“在本发明的一种最佳实现中,视频信
号是根据 H.263 标准编码的,而指示符包含在附加增强信息中”。
证据 1 公开了(说明书第 3 页倒数第 2 段)图 1 提出使用块边缘过滤及在
H.263+标准中提出的预测块编码系统中的多个存储帧。
因此根据证据 2 公开的内容,本领域技术人员可以清楚地知道共同参照图像
可以在图像信头中被识别。证据 2 中的图像信头提供共同信息区域,用于存储证
据 1 中共同参照图像的标识符。即证据 1 和 2 的结合公开了特征“以及(b)位
16
于所述多像块图像单元前的共同信息区域中描述了描述用于识别待被共同参照
的共同参照图像的信息”。
证据 1 公开了(说明书第 4 页第 1 段)对图像单元执行译码。运动估计器提
供对相对于在前图像的一个图像的诸块或诸宏块的运动估计。本领域技术人员知
晓证据 1 中的一个图像对应于权利要求 4 中的一个图像单元,证据 1 中的宏块对
应于权利要求 4 中的宏像块单元,因此证据 1 公开了多像块图像单元是一个图像
单元或宏块,即公开了多像块图像单元是多个图像单元,一个图像单元,片段单
元和宏像块单元中的一个。
综上所述,请求人认为,涉案专利的权利要求 4 与证据 2 和 1 的结合相比,
不具备突出的实质性特点和显著的进步,不符合专利法第 22 条第 3 款规定的创
造性。因此,请求人恳请宣告涉案专利的权利要求 4 无效。
17
三款的规定。
证据3公开了(第13页,“5.句法和语义”)…视频句法排列成具有4个基本
层的分级结构。从上到下,依次为:图像;块组;宏块;块。
证据 3 还公开了(第 22 页,
“5.1 图像层”第一段)每个图像的数据包括一
个图像信头以及跟随其后的像块组的数据或者片段的数据。
证据 3 公开了低比特率通信的视频译码,而本领域技术人员知晓视频序列由
一组图像帧组成(相当于一种图像解码方法)。
证据 3 公开了(第 9 页最后一段)每个图像被分割成块组(GOB,group of
block)或切片。GOB 和切片均包括多个宏块。证据 3 还公开了(第 10 页第 1
段)每个 GOB 的数据由 GOB 信头(可以是空的)和其后的宏块数据组成。对
于 H.263 标准,宏块是基于参考图像的预测和运动矢量生成的处理单元(证据 3
的第 12 页第 3 段公开了解码器将接收每个宏块的一个矢量…)。因此本领域技术
人员知晓证据 3 中的宏块相当于涉案专利权利要求 1 中的像块。
证据 3 公开了(第 2 页 3.2)视频译码器实施与视频编码器相反的处理。
证据 3 公开了(第 96-97 页,以及图 N.1)参照图像选择模式,其中视频编
码器从多个候选参照图像中选择一个参照图像用于帧间预测。证据 3 公开了(第
12 页第 3 段)解码器将接收每个宏块的一个矢量,(第 11 页)当进行帧间预测
时,INTRA 编码模式可以在宏块级别上标示,从而解决在宏块基础上进行帧间
预测的问题,(第 12 页最后一段)解释量化-在为每个宏块生成运动矢量之后的
(第 33-40 页)提供宏块层的概
编码过程中的一个步骤-也发生在宏块的基础上,
述,并解释如何在内部/之间基础上对单个宏块进行编码。可以看出,帧间预测
是基于宏块进行的,对于每个宏块,当对所编码的图像的宏块与所选参照图像的
相应宏块进行比较时,生成单个运动矢量(相当于其基于像块从多个参照图像中
选择一个参照图像)。
证据 3 公开了(第 96 页)在参照图像选择模式中,用于通知选择用于预测
的图像的信息包含在编码的比特流中,也即是说,一旦选定参照图像,它在编码
的比特流中被识别。证据 3 还公开了选择的参照图像可以在 GOB 层的句法中被
识别。具体地,每个 GOB 包括识别选择的参照图像的信头信息,视频编码器使
用信头信息对 GOB 各自的宏块进行帧间预测。
18
证据 3 公开了(第 9 页)视频源编码算法,其中主要内容是预测、块变换和
量化。证据 3 公开了(第 11 页)以宏块为最低级编码单元的预测编码的性能,
在进行帧间预测时,可以在宏块级别上发出内部编码模式的信号,从而解决在宏
块基础上进行帧间预测的问题,(第 12 页)解码器将接收每个宏块的一个矢量。
(相当于并在待解码的当前图像中的、在编码的动态图像数据中包含的像块上执
行预测解码,所述方法包括)
证据 3 公开了(第 97 页,图 N.1)参照图像选择模式中使用的视频编码器。
这个图显示了使用一些图像存储器的结构。视频编码器依赖于一些附加的图像存
储器,包括存储器 AP1、AP2 和 AP3,来存储所选择的用于编码的参照图像。证
据 3 公开了源编码器可以选择附加图像存储器之一来抑制帧间编码引起的时间
误差传播。所选择的参照图像存储在一个图像存储器中。
如前面所述,证据 3 公开了视频编码器基于宏块执行预测编码。本领域技术
人员知晓证据 3 中的作为最低级别编码单元的宏块对应于涉案专利权利要求 1
中的像块。证据 3 还公开了(第 10-11,98-99 页)GOB 包括多个宏块(相当于
当对由所述编码的动态图像数据中包含的多个像块构成的多像块图像单元解码
时)。
如前面所述,证据 3 公开了从多个参照图像中选择参照对象,以对像块执行
预定的编码。证据 3 公开了(第 97 页)源编码器可以选择多个附加的图像存储
器中的一个来抑制由于帧间编码而产生的时间误差传播。用于描述所选择的用于
预测的图像的信息包含在编码的比特流中。即,源编码器从存储在附加图像存储
器(AP1、AP2 和 AP3)中的图像中选择一幅图像作为组成该 GOB 的宏块共同参照
的参照图像。证据 3 公开了(第 98 页倒数第 3 段)在编码比特流中识别所选择
的参照图像时,编码比特流仅在 GOB 或切片层中更改。图 N.2 示出应用参照图
像选择模式时 GOB 层的句法。图 N.2 示出每个 GOB 包括一个“TRP(Temporal
Reference for Prediction,用于预测的时域参照)”字段。GOB 对所选择的参照图
像在 TRP 字段的识别应用于组成 GOB 的宏块。因此,本领域技术人员知晓在
GOB 的 TRP 字段被识别所选择的图像是构成 GOB 的宏块的共同参照图像(相
当于在…共同信息区域中描述了用于识别待被共同参照的共同参照图像的信息
19
的情况下)。
证据 3 公开了(第 9 页)视频编码器包括预测、块变换和量化。具体地,证
据 3 公开了(第 97 页)源编码器可以选择多个附加的图像存储器中的一个来抑
制由于帧间编码而产生的时间误差传播。用于描述所选择的用于预测的图像的信
息包含在编码的比特流中。即,源编码器从存储在附加图像存储器(AP1、AP2 和
AP3)中的图像中选择一幅图像作为组成该 GOB 的宏块共同参照的参照图像。即
源编码器使用所选择的共同参照图像为待编码的当前块生成预测图像(相当于使
用所述共同参照图像,生成所述多像块图像单元中包含的当前像块的预测图像)。
如前面所述,证据 3 公开了从多个参照图像中选择参照对象,以对像块执行
预定的编码。证据 3 公开了(第 97 页)源编码器可以选择多个附加的图像存储
器中的一个来抑制由于帧间编码而产生的时间误差传播。用于描述所选择的用于
预测的图像的信息包含在编码的比特流中。即,源编码器从存储在附加图像存储
器(AP1、AP2 和 AP3)中的图像中选择一幅图像作为组成该 GOB 的宏块共同参照
的参照图像。证据 3 公开了(第 98 页倒数第 3 段)在编码比特流中识别所选择
的参照图像时,编码比特流仅在 GOB 或切片层中更改。图 N.2 示出应用参照图
像选择模式时 GOB 层的句法。图 N.2 示出每个 GOB 包括一个“TRP(Temporal
Reference for Prediction,用于预测的时域参照)”字段。GOB 对所选择的参照图
像在 TRP 字段的识别应用于组成 GOB 的宏块。因此,本领域技术人员知晓在
GOB 的 TRP 字段被识别所选择的图像是构成 GOB 的宏块的共同参照图像(相
当于在基于像块识别了针对包含在所述多像块图像单元中的所述当前像块识别
的参照图像的情况下)。
证据 3 公开了(第 9 页)视频编码器包括预测、块变换和量化。具体地,证
据 3 公开了(第 97 页)源编码器可以选择多个附加的图像存储器中的一个来抑
制由于帧间编码而产生的时间误差传播。用于描述所选择的用于预测的图像的信
息包含在编码的比特流中。即,源编码器从存储在附加图像存储器(AP1、AP2 和
AP3)中的图像中选择一幅图像作为组成该 GOB 的宏块共同参照的参照图像。即
源编码器使用所选择的共同参照图像为待编码的当前块生成预测图像(相当于使
用所述参照图像,生成所述当前像块的预测图像)。
证据 3 公开了(第 9 页)视频编码器使用预测的图像执行块变换和量化。本
20
领域技术人员知晓使用预测的图像执行块变换和量化即是进行编码和解码(相当
于使用所述预测图像对所述当前像块解码)。
证据 3 公开了“以及(b)位于所述多像块图像单元前的共同信息区域中描
述了用于识别待被共同参照的共同参照图像的信息”。
根据前述内容,每个 GOB 包括一个 TRP 字段。GOB 对所选择的参照图像
在 TRP 字段的识别应用于组成 GOB 的宏块。本领域技术人员知晓存储在 TRP
字段中的 TRP 数据是 GOB 的用于识别所选择的共同参照图像的共同信息。此外,
本领域技术人员还知晓当应用参照图像选择模式时,TRP 字段通常用于 GOB 层
中所有 GOB 的语法中,它是一个共同信息描述单元,因为它保存了 GOB 的宏
块所共有的信息,且图 N.3 (证据 3 第 99 页)显示 TRP 字段是在宏块数据之前。
GOB 信头包括 GOB 的共同信息,GOB 信头对应于多像块图像单元的共同信息
区域。对于正在编码的 GOB,TRP 字段是共同信息区域,因为 TRP 数据由组成
GOB 的宏块共同使用(证据 3 第 98 页)。即证据 3 公开了在“以及(b)位于所
述多像块图像单元前的共同信息区域中描述了用于识别待被共同参照的共同参
照图像的信息”。
证据 3 公开了“在(a)所述编码的动态图像数据中包含的…共同信息区域
中描述了用于识别待被共同参照的共同参照图像的信息”。
根据前述内容,证据 3 公开了在多像块图像单元的共同信息区域中,描述用
于识别所选择的共同参照图像的共同信息。证据 3 公开了多像块图像单元包含在
编码的动态图像数据中。本领域技术人员清楚证据 3 中描述用于识别待被共同参
照的共同参照图像的信息包含在所述编码的动态图像数据中的共同信息区域中。
针对技术特征“以及(b)位于所述多像块图像单元前的共同信息区域中描
述了用于识别待被共同参照的共同参照图像的信息”,除了证据 3 公开的前述内
容之外,证据 2 也公开了该特征。
证据 2 公开了(说明书第[0019]段)“编码器可用这种指示符来指示解码器
哪些图像极为类似于当前参考图像,使得如果实际参考图像在传输过程中丢失,
则这些图像之一可用作备用参考图像”。证据 2 公开了(说明书第[0020]段)备
用参考图像编号通常包含在图像信头中,以识别解码器的备用参考图片。本领域
技术人员清楚信头位于多像块图像单元的前面。
21
证据 2 公开了(说明书第[0012]段)图像层的数据包含影响整个图像区以及
图像数据解码的参数。这种数据的绝大部分安排在所谓的图像信头中。证据 2
公开了“包含在图像信头中的备用参考图像编号是影响整个图像区以及图像数据
解码的参数”。
因此证据 2 公开了备用参考图像被编码器通过图像信头共同识别,如果实际
参考图像在传输过程中丢失,译码器使用识别的备用参考图像。本领域技术人员
知晓证据 2 公开了在图片的共同信息区域(备用参考图像编号,对应在被编码的
图像的信头中的位置)中,识别选择的共同参照图像(备用参考图像)。
证据 2 公开了(说明书第[0020]段)“在本发明的一种最佳实现中,视频信
号是根据 H.263 标准编码的,而指示符包含在附加增强信息中”。
本领域技术人员有动机使用证据 2 中的教导(备用参考图像编号包含在图像
信头中)来改进证据 3 中的共同参照图像。
因此根据证据 2 公开的内容,本领域技术人员可以清楚地知道共同参照图像
可以在图像信头中被识别。证据 2 中的图像信头提供共同信息区域,用于存储证
据 1 中共同参照图像的标识符。即证据 3 和 2 的结合公开了特征“以及(b)位
于所述多像块图像单元前的共同信息区域中描述了用于识别待被共同参照的共
同参照图像的信息”。
涉案专利权利要求 1 中的技术特征已被证据 3 和 2 所公开,并且在证据 3
和 2 中所起的作用与在涉案专利中起到的作用相同,都是生成当前像块的预测图
像。在此基础上,本领域技术人员容易想到将证据 2 与证据 3 相结合,从而得到
涉案专利权利要求 1 的技术方案。
综上所述,请求人认为,涉案专利的权利要求 1 与证据 2 和 3 的结合相比,
不具备突出的实质性特点和显著的进步,不符合专利法第 22 条第 3 款规定的创
造性。因此,请求人恳请宣告涉案专利的权利要求 1 无效。
2、涉案专利的权利要求 2 不具备创造性,不符合《专利法》第二十二条第
三款的规定。
证据 3 公开了(第 9 页最后一段)每个图像被分割成像块(GOB)或切片。
GOB 和切片均包括多个宏块。证据 3 还公开了(第 10 页第 1 段)每个 GOB 的
数据由 GOB 信头(可以是空的)和其后的宏块数据组成。对于 H.263 标准,宏
22
块是基于参考图像的预测和运动矢量生成的处理单元(证据 3 的第 12 页第 3 段
。因此本领域技术人员知晓证据 3
公开了解码器将接收每个宏块的一个矢量…)
中的宏块相当于涉案专利权利要求 2 中的像块。另外证据 3 中的切片(slice)与
涉案专利权利要求 2 中的片段(slice)单元,仅是表述上的不同,其本质是相同
内容的不同表述。因此本领域技术人员知晓证据 3 中公开了多像块图像单元是多
个图像单元,一个图像单元,片段单元和宏像块单元中的一个。
综上所述,请求人认为,涉案专利的权利要求 2 与证据 2 和 3 的结合相比,
不具备突出的实质性特点和显著的进步,不符合专利法第 22 条第 3 款规定的创
造性。因此,请求人恳请宣告涉案专利的权利要求 2 无效。
3、涉案专利的权利要求 3 不具备创造性,不符合《专利法》第二十二条第
三款的规定。
证据3公开了(第13页,“5.句法和语义”)…视频句法排列成具有4个基本
层的分级结构。从上到下,依次为:图像;块组;宏块;块。
证据 3 还公开了(第 22 页,
“5.1 图像层”第一段)每个图像的数据包括一
个图像信头以及跟随其后的像块组的数据或者片段的数据。
证据 3 公开了低比特率通信的视频译码,而本领域技术人员知晓视频序列由
一组图像帧组成(相当于一种图像解码装置)。
证据 3 公开了(第 9 页最后一段)每个图像被分割成块组(GOB,group of
block)或切片。GOB 和切片均包括多个宏块。证据 3 还公开了(第 10 页第 1
段)每个 GOB 的数据由 GOB 信头(可以是空的)和其后的宏块数据组成。对
于 H.263 标准,宏块是基于参考图像的预测和运动矢量生成的处理单元(证据 3
的第 12 页第 3 段公开了解码器将接收每个宏块的一个矢量…)。因此本领域技术
人员知晓证据 3 中的宏块相当于涉案专利权利要求 1 中的像块。
证据 3 公开了(第 2 页 3.2)视频译码器实施与视频编码器相反的处理。
证据 3 公开了(第 96-97 页,以及图 N.1)参照图像选择模式,其中视频编
码器从多个候选参照图像中选择一个参照图像用于帧间预测。证据 3 公开了(第
12 页第 3 段)解码器将接收每个宏块的一个矢量,(第 11 页)当进行帧间预测
时,INTRA 编码模式可以在宏块级别上标示,从而解决在宏块基础上进行帧间
23
预测的问题,(第 12 页最后一段)解释量化-在为每个宏块生成运动矢量之后的
(第 33-40 页)提供宏块层的概
编码过程中的一个步骤-也发生在宏块的基础上,
述,并解释如何在内部/之间基础上对单个宏块进行编码。可以看出,帧间预测
是基于宏块进行的,对于每个宏块,当对所编码的图像的宏块与所选参照图像的
相应宏块进行比较时,生成单个运动矢量(相当于其基于像块从多个参照图像中
选择一个参照图像)。
证据 3 公开了(第 96 页)在参照图像选择模式中,用于通知选择用于预测
的图像的信息包含在编码的比特流中,也即是说,一旦选定参照图像,它在编码
的比特流中被识别。证据 3 还公开了选择的参照图像可以在 GOB 层的句法中被
识别。具体地,每个 GOB 包括识别选择的参照图像的信头信息,视频编码器使
用信头信息对 GOB 各自的宏块进行帧间预测。
证据 3 公开了(第 9 页)视频源编码算法,其中主要内容是预测、块变换和
量化。证据 3 公开了(第 11 页)以宏块为最低级编码单元的预测编码的性能,
在进行帧间预测时,可以在宏块级别上发出内部编码模式的信号,从而解决在宏
块基础上进行帧间预测的问题,(第 12 页)解码器将接收每个宏块的一个矢量。
(相当于并在待解码的当前图像中的、在编码的动态图像数据中包含的像块上执
行预测解码,所述装置包括)
证据 3 公开了(第 97 页,图 N.1)参照图像选择模式中使用的视频编码器。
这个图显示了使用一些图像存储器的结构。视频编码器依赖于一些附加的图像存
储器,包括存储器 AP1、AP2 和 AP3,来存储所选择的用于编码的参照图像。证
据 3 公开了源编码器可以选择附加图像存储器之一来抑制帧间编码引起的时间
误差传播。所选择的参照图像存储在一个图像存储器中。
如前面所述,证据 3 公开了视频编码器基于宏块执行预测编码。本领域技术
人员知晓证据 3 中的作为最低级别编码单元的宏块对应于涉案专利权利要求 1
中的像块。证据 3 还公开了(第 10-11,98-99 页)GOB 包括多个宏块(相当于
进行生成操作的单元,i)当对由所述编码的动态图像数据中包含的多个像块构
成的多像块图像单元解码时)。
如前面所述,证据 3 公开了从多个参照图像中选择参照对象,以对像块执行
24
预定的编码。证据 3 公开了(第 97 页)源编码器可以选择多个附加的图像存储
器中的一个来抑制由于帧间编码而产生的时间误差传播。用于描述所选择的用于
预测的图像的信息包含在编码的比特流中。即,源编码器从存储在附加图像存储
器(AP1、AP2 和 AP3)中的图像中选择一幅图像作为组成该 GOB 的宏块共同参照
的参照图像。证据 3 公开了(第 98 页倒数第 3 段)在编码比特流中识别所选择
的参照图像时,编码比特流仅在 GOB 或切片层中更改。图 N.2 示出应用参照图
像选择模式时 GOB 层的句法。图 N.2 示出每个 GOB 包括一个“TRP(Temporal
Reference for Prediction,用于预测的时域参照)”字段。GOB 对所选择的参照图
像在 TRP 字段的识别应用于组成 GOB 的宏块。因此,本领域技术人员知晓在
GOB 的 TRP 字段被识别所选择的图像是构成 GOB 的宏块的共同参照图像(相
当于在…共同信息区域中描述了用于识别待被共同参照的共同参照图像的信息
的情况下)。
证据 3 公开了(第 9 页)视频编码器包括预测、块变换和量化。具体地,证
据 3 公开了(第 97 页)源编码器可以选择多个附加的图像存储器中的一个来抑
制由于帧间编码而产生的时间误差传播。用于描述所选择的用于预测的图像的信
息包含在编码的比特流中。即,源编码器从存储在附加图像存储器(AP1、AP2 和
AP3)中的图像中选择一幅图像作为组成该 GOB 的宏块共同参照的参照图像。即
源编码器使用所选择的共同参照图像为待编码的当前块生成预测图像(相当于使
用所述共同参照图像,生成所述多像块图像单元中包含的当前像块的预测图像)。
如前面所述,证据 3 公开了从多个参照图像中选择参照对象,以对像块执行
预定的编码。证据 3 公开了(第 97 页)源编码器可以选择多个附加的图像存储
器中的一个来抑制由于帧间编码而产生的时间误差传播。用于描述所选择的用于
预测的图像的信息包含在编码的比特流中。即,源编码器从存储在附加图像存储
器(AP1、AP2 和 AP3)中的图像中选择一幅图像作为组成该 GOB 的宏块共同参照
的参照图像。证据 3 公开了(第 98 页倒数第 3 段)在编码比特流中识别所选择
的参照图像时,编码比特流仅在 GOB 或切片层中更改。图 N.2 示出应用参照图
像选择模式时 GOB 层的句法。图 N.2 示出每个 GOB 包括一个“TRP(Temporal
Reference for Prediction,用于预测的时域参照)”字段。GOB 对所选择的参照图
像在 TRP 字段的识别应用于组成 GOB 的宏块。因此,本领域技术人员知晓在
25
GOB 的 TRP 字段被识别所选择的图像是构成 GOB 的宏块的共同参照图像(相
当于以及 ii)在基于像块识别了针对包含在所述多像块图像单元中的所述当前像
块识别的参照图像的情况下)。
证据 3 公开了(第 9 页)视频编码器包括预测、块变换和量化。具体地,证
据 3 公开了(第 97 页)源编码器可以选择多个附加的图像存储器中的一个来抑
制由于帧间编码而产生的时间误差传播。用于描述所选择的用于预测的图像的信
息包含在编码的比特流中。即,源编码器从存储在附加图像存储器(AP1、AP2 和
AP3)中的图像中选择一幅图像作为组成该 GOB 的宏块共同参照的参照图像。即
源编码器使用所选择的共同参照图像为待编码的当前块生成预测图像(相当于使
用所述参照图像,生成所述当前像块的预测图像)。
证据 3 公开了(第 9 页)视频编码器使用预测的图像执行块变换和量化。本
领域技术人员知晓使用预测的图像执行块变换和量化即是进行编码和解码(相当
于进行解码操作的单元,使用所述预测图像对所述当前像块解码)。
证据 3 公开了“以及(b)位于所述多像块图像单元前的共同信息区域中描
述了用于识别待被共同参照的共同参照图像的信息”。
根据前述内容,每个 GOB 包括一个 TRP 字段。GOB 对所选择的参照图像
在 TRP 字段的识别应用于组成 GOB 的宏块。本领域技术人员知晓存储在 TRP
字段中的 TRP 数据是 GOB 的用于识别所选择的共同参照图像的共同信息。此外,
本领域技术人员还知晓当应用参照图像选择模式时,TRP 字段通常用于 GOB 层
中所有 GOB 的语法中,它是一个共同信息描述单元,因为它保存了 GOB 的宏
块所共有的信息,且图 N.3 (证据 3 第 99 页)显示 TRP 字段是在宏块数据之前。
GOB 信头包括 GOB 的共同信息,GOB 信头对应于多像块图像单元的共同信息
区域。对于正在编码的 GOB,TRP 字段是共同信息区域,因为 TRP 数据由组成
GOB 的宏块共同使用(证据 3 第 98 页)。即证据 3 公开了“以及(b)位于所述
多像块图像单元前的共同信息区域中描述了用于识别待被共同参照的共同参照
图像的信息”。
证据 3 公开了“在(a)所述编码的动态图像数据中包含的…共同信息区域
中描述了用于识别待被共同参照的共同参照图像的信息”。
根据前述内容,证据 3 公开了在多像块图像单元的共同信息区域中,描述用
26
于识别所选择的共同参照图像的共同信息。证据 3 公开了多像块图像单元包含在
编码的动态图像数据中。本领域技术人员清楚证据 3 中描述用于识别待被共同参
照的共同参照图像的信息包含在所述编码的动态图像数据中的共同信息区域中。
针对技术特征“以及(b)位于所述多像块图像单元前的共同信息区域中描
述了用于识别待被共同参照的共同参照图像的信息”,除了证据 3 公开的前述内
容之外,证据 2 也公开了该特征。
证据 2 公开了(说明书第[0019]段)“编码器可用这种指示符来指示解码器
哪些图像极为类似于当前参考图像,使得如果实际参考图像在传输过程中丢失,
则这些图像之一可用作备用参考图像”。证据 2 公开了(说明书第[0020]段)备
用参考图像编号通常包含在图像信头中,以识别解码器的备用参考图片。本领域
技术人员清楚信头位于多像块图像单元的前面。
证据 2 公开了(说明书第[0012]段)图像层的数据包含影响整个图像区以及
图像数据解码的参数。这种数据的绝大部分安排在所谓的图像信头中。证据 2
公开了“包含在图像信头中的备用参考图像编号是影响整个图像区以及图像数据
解码的参数”。
因此证据 2 公开了备用参考图像被编码器通过图像信头共同识别,如果实际
参考图像在传输过程中丢失,译码器使用识别的备用参考图像。本领域技术人员
知晓证据 2 公开了在图片的共同信息区域(备用参考图像编号,对应在被编码的
图像的信头中的位置)中,识别选择的共同参照图像(备用参考图像)。
证据 2 还公开了(说明书第[0020]段)“在本发明的一种最佳实现中,视频
信号是根据 H.263 标准编码的,而指示符包含在附加增强信息中”。
本领域技术人员有动机使用证据 2 中的教导(备用参考图像编号包含在图像
信头中)来改进证据 3 中的共同参照图像。
因此根据证据 2 公开的内容,本领域技术人员可以清楚地知道共同参照图像
可以在图像信头中被识别。证据 2 中的图像信头提供共同信息区域,用于存储证
据 1 中共同参照图像的标识符。即证据 3 和 2 的结合公开了特征“以及(b)位
于所述多像块图像单元前的共同信息区域中描述了用于识别待被共同参照的共
同参照图像的信息”。
涉案专利权利要求 3 中的技术特征已被证据 3 和 2 所公开,并且在证据 3
27
和 2 中所起的作用与在涉案专利中起到的作用相同,都是生成当前像块的预测图
像。在此基础上,本领域技术人员容易想到将证据 2 与证据 3 相结合,从而得到
涉案专利权利要求 3 的技术方案。
综上所述,请求人认为,涉案专利的权利要求 3 与证据 2 和 3 的结合相比,
不具备突出的实质性特点和显著的进步,不符合专利法第 22 条第 3 款规定的创
造性。因此,请求人恳请宣告涉案专利的权利要求 3 无效。
4、涉案专利的权利要求 4 不具备创造性,不符合《专利法》第二十二条第
三款的规定。
证据 3 公开了(第 9 页最后一段)每个图像被分割成像块(GOB)或切片。
GOB 和切片均包括多个宏块。证据 3 还公开了(第 10 页第 1 段)每个 GOB 的
数据由 GOB 信头(可以是空的)和其后的宏块数据组成。对于 H.263 标准,宏
块是基于参考图像的预测和运动矢量生成的处理单元(证据 3 的第 12 页第 3 段
。因此本领域技术人员知晓证据 3
公开了解码器将接收每个宏块的一个矢量…)
中的宏块相当于涉案专利权利要求 4 中的像块。另外证据 3 中的切片(slice)与
涉案专利权利要求 4 中的片段(slice)单元,仅是表述上的不同,其本质是相同
内容的不同表述。因此本领域技术人员知晓证据 3 中公开了多像块图像单元是多
个图像单元,一个图像单元,片段单元和宏像块单元中的一个。
综上所述,请求人认为,涉案专利的权利要求 4 与证据 2 和 3 的结合相比,
不具备突出的实质性特点和显著的进步,不符合专利法第 22 条第 3 款规定的创
造性。因此,请求人恳请宣告涉案专利的权利要求 4 无效。
28
ЁढҎ⇥݅ᆊⶹ䆚ѻᴗሔ
*CN101035291B*
থᯢϧ߽
ᥜᴗ݀ਞো&1%
ᥜᴗ݀ਞ᮹
⬇䇋ো; ᇍ↨᭛ӊ
&1$ ܼ᭛
⬇䇋᮹
(3$ ܼ᭛
Ӭܜᴗ᭄ -3⡍ᓔᑇ খ㾕䇈
-3 ᯢк ↉ˈ ↉ˈ ↉ˈ䰘
ߚḜॳ⬇䇋᭄
-3⡍ᓔ $ ܼ᭛
ϧ߽ᴗҎ ᵒϟ⬉఼ѻϮ᷾ᓣӮ⼒ &1$ ܼ᭛
ഄഔ ᮹ᴀ䯾ᑰ ᅵᶹਬᓴђḶ
থᯢҎ 㖑但䆮㾦䞢ⳳг䖥㮸ᬣᖫ
ᅝ⏙ס
ϧ߽ҷ⧚ᴎᵘ ∌ᮄϧ߽ଚᷛҷ⧚᳝䰤݀ৌ
ҷ⧚Ҏ ⥟㣅
,QW&O
+1
+1
+1
+1
+1
ᵳ࡙㾱≲Җ亥䈤᰾Җ亥䱴മ亥
থᯢৡ⿄
ࡼᗕڣ㓪ⷕᮍ⊩ঞࡼᗕڣ㾷ⷕᮍ⊩
ᨬ㽕
ࡼᗕڣ㓪ⷕ㺙㕂ˈࣙᣀ ˖ ᐙখ✻ڣ
ᦦؐ乘⌟ᯊᇚ ᐙখ✻ڣЁⱘ ᐙᅮЎ䕧
ܹⱘ咬䅸খ✻ڣ㓪ো 'HI5HI1R ᠔ᣛ⼎ⱘখ✻
ڣǃ䖯㸠䖤ࡼᅮⱘ䖤ࡼᅮ䚼 ˗ᇍ↣Ͼ
ڣഫᇍ⅟Ꮒ㓪ⷕ᭄ (5HVǃ乘⌟⾡㉏ 3UHG7\SHǃ
খ ✻ ڣ㓪 ো 5HI1R ঞ 䖤 ࡼ ⶶ 䞣 09ǃ09 䖯
㸠ৃব䭓㓪ⷕˈᇍ↣ᐙڣᇍ咬䅸খ✻ڣ㓪ো
'HI5HI1Rˈ䕧ߎࡼᗕڣ㓪ⷕ᭄ 6WU 䖯㸠ৃব
&1%
䭓㓪ⷕⱘৃব䭓㓪ⷕ䚼 DŽ
&1% ᴗ߽㽕∖к 义
ϔ⾡ڣ㾷ⷕᮍ⊩ˈ݊ѢڣഫҢϾখ✻ڣЁ䗝ᢽϔϾখ✻ˈڣᑊᕙ㾷ⷕ
ⱘᔧࠡڣЁⱘǃ㓪ⷕⱘࡼᗕڣ᭄ЁࣙⱘڣഫϞᠻ㸠乘⌟㾷ⷕˈ᠔䗄ᮍ⊩ࣙᣀ ˖
ᔧᇍ⬅᠔䗄㓪ⷕⱘࡼᗕڣ᭄ЁࣙⱘϾڣഫᵘ៤ⱘڣഫڣऩܗ㾷ⷕᯊˈ
D ᠔䗄㓪ⷕⱘࡼᗕڣ᭄Ёࣙⱘˈҹঞ E ԡѢ᠔䗄ڣഫڣऩ݅ⱘࠡܗৠֵᙃ
ऎඳЁᦣ䗄њ⫼Ѣ䆚߿ᕙ㹿݅ৠখ✻ⱘ݅ৠখ✻ֵⱘڣᙃⱘᚙމϟˈՓ⫼᠔䗄݅ৠখ✻
⫳ˈڣ៤᠔䗄ڣഫڣऩܗЁࣙⱘᔧࠡڣഫⱘ乘⌟˗ ڣ
Ѣڣഫ䆚߿њ䩜ᇍࣙ᠔䗄ڣഫڣऩܗЁⱘ᠔䗄ᔧࠡڣഫ䆚߿ⱘখ✻
ⱘڣᚙމϟˈՓ⫼᠔䗄খ✻⫳ˈڣ៤᠔䗄ᔧࠡڣഫⱘ乘⌟˗ ڣ
Փ⫼᠔䗄乘⌟ڣᇍ᠔䗄ᔧࠡڣഫ㾷ⷕDŽ
བᴗ߽㽕∖ ᠔䗄ⱘڣ㾷ⷕᮍ⊩ˈ
݊Ёˈ᠔䗄ڣഫڣऩܗᰃϾڣऩˈܗϔϾڣऩˈܗ⠛↉ऩܗᅣڣഫऩܗЁ
ⱘϔϾDŽ
ϔ⾡ڣ㾷ⷕ㺙㕂ˈ݊ѢڣഫҢϾখ✻ڣЁ䗝ᢽϔϾখ✻ˈڣᑊᕙ㾷ⷕ
ⱘᔧࠡڣЁⱘǃ㓪ⷕⱘࡼᗕڣ᭄ЁࣙⱘڣഫϞᠻ㸠乘⌟㾷ⷕˈ ᠔䗄㺙㕂ࣙᣀ ˖
䖯㸠⫳៤᪡ⱘऩˈܗL ᔧᇍ⬅᠔䗄㓪ⷕⱘࡼᗕڣ᭄ЁࣙⱘϾڣഫᵘ៤ⱘ
ڣഫڣऩܗ㾷ⷕᯊˈ D ᠔䗄㓪ⷕⱘࡼᗕڣ᭄Ёࣙⱘˈҹঞ E ԡѢ᠔䗄ڣഫ
ڣऩ݅ⱘࠡܗৠֵᙃऎඳЁᦣ䗄њ⫼Ѣ䆚߿ᕙ㹿݅ৠখ✻ⱘ݅ৠখ✻ֵⱘڣᙃⱘᚙ
މϟˈՓ⫼᠔䗄݅ৠখ✻⫳ˈڣ៤᠔䗄ڣഫڣऩܗЁࣙⱘᔧࠡڣഫⱘ乘⌟˗ ڣ ҹ
ঞ LL Ѣڣഫ䆚߿њ䩜ᇍࣙ᠔䗄ڣഫڣऩܗЁⱘ᠔䗄ᔧࠡڣഫ䆚߿ⱘখ✻
ⱘڣᚙމϟˈՓ⫼᠔䗄খ✻⫳ˈڣ៤᠔䗄ᔧࠡڣഫⱘ乘⌟˗ ڣ ҹঞ
䖯㸠㾷ⷕ᪡ⱘऩˈܗՓ⫼᠔䗄乘⌟ڣᇍ᠔䗄ᔧࠡڣഫ㾷ⷕDŽ
བᴗ߽㽕∖ ᠔䗄ⱘڣ㾷ⷕ㺙㕂ˈ
݊Ёˈ᠔䗄ڣഫڣऩܗᰃϾڣऩˈܗϔϾڣऩˈܗ⠛↉ऩܗᅣڣഫऩܗЁ
ⱘϔϾDŽ
&1% 䇈ᯢк 义
ࡼᗕڣ㓪ⷕᮍ⊩ঞࡼᗕڣ㾷ⷕᮍ⊩
ᡔᴃ乚ඳ
>@ ᴀথᯢ⍝ঞϔ⾡ᇍࡼᗕڣ᭄䖯㸠㓪ⷕঞ㾷ⷕⱘᮍ⊩ˈҹঞ䆄ᔩњ⫼Ѣ䗮䖛䕃
ӊᅲ⦄䆹ᮍ⊩ⱘᑣⱘ䆄ᔩၦԧDŽ
㚠᱃ᡔᴃ
>@䖥ᑈᴹˈ䱣ⴔၦԧᑨ⫼ⱘথሩˈڣǃໄ䷇ǃ᭛ᴀㄝ᠔᳝ၦԧֵᙃϔ㠀㛑㒳ϔ
ഄ໘⧚DŽ䖭ᯊˈ䗮䖛ᇚ᠔᳝ⱘၦԧ᭄ᄫ࣪ˈ㛑㒳ϔഄ໘⧚ၦԧDŽԚᰃˈ⬅Ѣ᭄ᄫ࣪њⱘ
ڣ᳝ᑲⱘ᭄䞣ˈℸЎњᄬټǃӴ䕧ˈֵⱘڣᙃय़㓽ᡔᴃϡৃ㔎DŽ㗠ЎњⳌѦ
ᑨ⫼य़㓽ৢⱘڣ᭄ˈय़㓽ᡔᴃⱘᷛ࣪ޚгᕜ䞡㽕DŽЎڣय़㓽ᡔᴃⱘᷛޚ㾘Ḑˈ᳝
,78 䰙⬉⇨䗮ֵ㘨ড়⬉⇨䗮ֵᷛ࣪ޚ䚼 ⱘ +ǃ+ˈ,62 䰙ᷛ࣪ޚᴎᵘ ⱘ
03(* 䖤ࡼڣϧᆊ㒘 ǃ03(*ǃ03(* ㄝDŽ
>@ Ў䖭ѯࡼᗕڣ㓪ⷕᮍᓣ݅ৠⱘᡔᴃˈ᳝Ԉ䱣䖤ࡼ㸹ٓⱘڣ䯈乘⌟DŽ䖭ѯ
ࡼᗕڣ㓪ⷕᮍᓣⱘ䖤ࡼ㸹ٓЁˈᇚ䕧ܹⱘࡆߚڣ៤乘ᅮᇣⱘڣഫ EORFNˈ ᇍ↣Ͼڣ
ഫˈḍ㸼⼎ڣ䯈ⱘ䖤ࡼⱘ䖤ࡼⶶ䞣⫳៤乘⌟ڣDŽ03(* ⱘڣ䯈乘⌟ˈ Փ⫼ḍ ᐙ
ᰒ⼎ᯊࠏϞ䍙ࠡѢ㓪ⷕᇍ䈵ⱘڣڣ䖯㸠ⱘࠡᮍ乘⌟ǃḍ ᐙᰒ⼎ᯊࠏϞ⒲ৢѢ㓪
ⷕᇍ䈵ⱘڣڣ䖯㸠ⱘৢᮍ乘⌟ǃḍᰒ⼎ᯊࠏϞ䍙ࠡѢ㓪ⷕᇍ䈵ڣঞᰒ⼎ᯊ
ࠏϞ⒲ৢѢ㓪ⷕᇍ䈵݅ⱘڣ䅵 ᐙڣ䖯㸠ڣ㋴ᦦؐ乘⌟ⱘঠ乘⌟ খ✻՟བ ,62
,(& ˖
( ֵᙃᡔᴃ 㾚ᇍ䈵㓪ⷕ 3DUW ˖
ڣ3 ᯊ
䯈乘⌟㒧ᵘ DŽ
>@ 03(* ᇍѢ⬏䴶䯈乘⌟ⱘ⾡㉏ଃϔഄއᅮՓ⫼ⱘখ✻⬏䴶ˈϡ㛑䗝ᢽӏᛣഄখ✻⬏
䴶DŽ㗠 ,78 ⱘℷ໘Ѣᷛ࣪ޚПЁⱘ + ℷ䅼䆎ᠽሩⱘ ᮍ乘⌟ˈҹ֓㛑Ϣ㓪ⷕᇍ
䈵ⱘڣᰒ⼎ᯊࠏ᮴݇ഄҢᄬټڣᄬ఼ټЁⱘᐙᏆ㓪ⷕᅠ↩ⱘڣЁ䗝ᢽӏᛣⱘ
ᐙখ✻ڣDŽ
Ў㸼⼎ + Ёⱘࡼᗕڣ㓪ⷕ㺙㕂ⱘᵘ៤ⱘᮍḚDŽ ⱘҹࠡⱘࡼᗕ
>@
ڣ㓪ⷕ㺙㕂ڣ䯈乘⌟ᯊˈ䞛⫼ᠻ㸠㛑ҢᐙڣЁ䗝ᢽখ✻ࡼⱘڣᗕڣ㓪ⷕᮍ
ᓣⱘ㺙㕂DŽ
>@䆹ࡼᗕڣ㓪ⷕ㺙㕂བ ᠔⼎ࣙᣀ䖤ࡼᅮ䚼 ǃڣ㋴ᦦؐ䚼 ǃ఼⊩ޣ
ǃڣ㓪ⷕ䚼 ǃڣ㾷ⷕ䚼 ǃࡴ⊩఼ ǃৃব䭓㓪ⷕ䚼 ǃᏻ㓧 ఼ކঞᓔ
݇ DŽ
䆹ࡼᗕڣ㓪ⷕ㺙㕂ᇚ䕧ܹⱘڣ᭄ ,PJ ߚࡆ៤ڣഫˈᇍ䆹↣Ͼڣഫ䖯㸠໘
>@
⧚DŽ ఼⊩ޣҢ䕧ܹࠄࡼᗕڣ㓪ⷕ㺙㕂Ёⱘڣ᭄Ёޣএ乘⌟ڣ᭄ 3UHGˈ Ў
⅟Ꮒ᭄ 5HV 䕧ߎDŽڣ㓪ⷕ䚼 ᇍ䕧ܹⱘ⅟Ꮒ᭄ 5HV 䖯㸠ℷѸবᤶǃ䞣ᄤ࣪ㄝڣ㓪
ⷕ໘⧚ˈ䕧ߎࣙ䞣ᄤ࣪ℷѸবᤶ㋏᭄ㄝⱘ⅟Ꮒ㓪ⷕ᭄ (5HVDŽڣ㾷ⷕ䚼 ᇍ䕧ܹⱘ
&1% 䇈ᯢк 义
&1% 䇈ᯢк 义
⌟ᓣ 3UHG7\SHDŽ
>@ Ў㸼⼎ҹᕔⱘࡼᗕڣ㾷ⷕ㺙㕂ⱘᵘ៤ⱘᮍḚDŽ
>@ བ ᠔⼎ˈ䆹ࡼᗕڣ㾷ⷕ㺙㕂ࣙᣀৃব䭓ᑺ㾷ⷕ䚼 ǃ䖤ࡼ㸹ٓ䚼 ǃڣ
㾷ⷕ䚼 ǃࡴ⊩఼ ǃڣ㋴ᦦؐ䚼 ǃᏻ㓧 ఼ކᓔ݇ DŽ
>@ ৃব䭓ᑺ㾷ⷕ䚼 ᇍ䕧ܹⱘࡼᗕڣ㓪ⷕ᭄ 6WU 䖯㸠ৃব䭓ᑺ㾷ⷕˈ䕧ߎ⅟
Ꮒ㓪ⷕ᭄ (5HVǃ䖤ࡼⶶ䞣 09ǃ09ǃখ✻ⱘڣ㓪ো 5HI1Rǃ5HI1Rǃ乘⌟⾡㉏ 3UHG7\SHDŽ
ڣ㾷ⷕ䚼 ᇍ䕧ܹⱘ⅟Ꮒ㓪ⷕ᭄ (5HV 䖯㸠䗚䞣ᄤ࣪ǃ䗚ℷѸবᤶㄝڣ㾷ⷕ໘⧚ˈ
䕧ߎ⅟Ꮒ㾷ⷕ᭄ '5HVDŽࡴ⊩఼ ᇚ⅟Ꮒ㾷ⷕ᭄ '5HV Ϣ乘⌟ڣ᭄ 3UHG Ⳍࡴˈ
Ў㾷ⷕڣ᭄ 'OPJ 䕧ߎࠄࡼᗕڣ㾷ⷕ㺙㕂DŽᏻ㓧ֱ ఼ކᄬ䖯㸠ڣ䯈乘⌟
ⱘ㾷ⷕڣ᭄ 'OPJDŽ
>@ 䖤ࡼ㸹ٓ䚼 ḍ乘⌟⾡㉏ 3UHG7\SH 䕧ߎڣ䯈乘⌟᠔ᖙ䳔ⱘখ✻ڣഫⱘখ
✻ⱘڣ㓪ো 15HI1Rǃ15HI1R 䖤ࡼⶶ䞣 09ǃ09ˈ
ᣛ⼎ᏻ㓧 ఼ކ䕧ߎখ✻ڣഫDŽ
ᔧ乘⌟⾡㉏ 3UHG7\SH 㸼⼎ᐙখ✻ؐᦦڣ乘⌟ᯊˈᏻ㓧 ఼ކ䕧ߎϢখ✻ⱘڣ
㓪ো 15HI1R 䖤ࡼⶶ䞣 1PY Ⳍᇍᑨⱘখ✻ڣഫ 5HI%ON Ϣখ✻ⱘڣ㓪ো 15HI1R
䖤ࡼⶶ䞣 109 Ⳍᇍᑨⱘখ✻ڣഫ 5HI%ONDŽڣ㋴ᦦؐ䚼 ⫼ᑇഛؐㄝᦦؐ Ͼখ✻
ڣഫ 5HI%ONǃ5HI%ON ⱘᇍᑨⱘڣ㋴ؐˈ䕧ߎᦦؐڣഫ 5HI3RDŽ㗠乘⌟⾡㉏ 3UHG7\SH
㸼⼎ᐙখ✻ؐᦦڣ乘⌟ҹⱘڣ䯈乘⌟ᯊˈᏻ㓧 ఼ކ䕧ߎϢখ✻ⱘڣ㓪ো
15HI1R 䖤ࡼⶶ䞣 109 Ⳍᇍᑨⱘখ✻ڣഫ 5HI%ONDŽ
ᑊϨˈ乘⌟⾡㉏ 3UHG7\SH 㸼⼎ᐙখ✻ؐᦦڣ乘⌟ᯊˈᓔ݇ ߛᤶࠄĀā
>@
ϔջˈᇚᦦؐڣഫ 5HI3RO Ў乘⌟ڣ᭄ 3UHG Փ⫼DŽ㗠ᔧ乘⌟⾡㉏ 3UHG7\SH 㸼⼎ᐙ
খ✻ؐᦦڣ乘⌟ҹⱘڣ䯈乘⌟ᮍ⊩ᯊˈ ᓔ݇ ߛᤶࠄĀāϔջˈᇚখ✻ڣഫ 5HI%ON
Ў乘⌟ڣ᭄ 3UHG Փ⫼DŽ䗮䖛Ϟ䗄䇈ᯢ䖛ⱘ໘⧚ˈ ࡼᗕڣ㾷ⷕ㺙㕂ᇚࡼᗕڣ㓪ⷕ
᭄ 6WU 㾷ⷕˈ䕧ߎڣ㾷ⷕ᭄ 'OPJDŽ
Ԛᰃˈ 03(* ⱘࡼᗕڣ㓪ⷕᮍ⊩ЁˈᅮНњབϟⱘᐙখ✻ؐᦦڣ乘⌟ᮍ
>@
⊩ˈՓ⫼㹿⿄Ўঠ乘⌟ⱘڣᐙখ✻ؐᦦڣ乘⌟ൟⱘڣЁˈ䗮䖛ḍᏆ㓪ⷕᅠ
↩ⱘ䖤ࡼⶶ䞣ˈ䅵ㅫ㹿⿄ЎⳈᓣⱘ⬅ᦦؐࠊ乘⌟ڣ᠔Փ⫼ⱘ ᐙখ✻ⱘڣ䖤ࡼ
ⶶ䞣ˈⳕ⬹ڣഫⱘ㓪ⷕ᭄Ёⱘ䖤ࡼⶶ䞣ঞখ✻ڣ㓪োDŽ
>@ Ў 03(* ⱘⳈᓣⱘ䇈ᯢDŽ䖭䞠ˈ ڣSLF 㸼⼎㓪ⷕᇍ䈵ˈڣ ڣ5HI
㸼⼎ᰒ⼎ᯊࠏ䍙ࠡѢ㓪ⷕᇍ䈵 ڣSLF ⱘখ✻ˈڣ ڣ5HI 㸼⼎ᰒ⼎ᯊࠏ⒲ৢѢ㓪ⷕᇍ
䈵 ڣSLF ⱘখ✻ڣˈڣഫ %ON 㸼⼎㓪ⷕᇍ䈵ڣഫˈڣഫ %ON 㸼⼎খ✻ ڣ5HI Ё⬏䴶
ԡ㕂Ϣ㓪ⷕᇍ䈵 %ON ⳌৠⱘڣഫDŽᑊϨˈ䖤ࡼⶶ䞣 09 㸼⼎ҹ㓪ⷕڣഫ %ON ᯊՓ⫼ⱘ
ڣ5HI Ўখ✻ⱘڣǃᣛࠡᮍⱘ䖤ࡼⶶ䞣ˈ 䖤ࡼⶶ䞣 09 㸼⼎ᣛখ✻ ڣ5HI ⱘ㓪ⷕ
ᇍ䈵ڣഫⱘ䖤ࡼⶶ䞣ˈ䖤ࡼⶶ䞣 09 㸼⼎ᣛখ✻ ڣ5HI ⱘ㓪ⷕᇍ䈵ڣഫⱘ䖤ࡼⶶ䞣ˈ
ڣഫ 5HI%ON 㸼⼎㹿䖤ࡼⶶ䞣 09 খ✻ⱘখ✻ڣഫˈڣഫ 5HI%ON 㸼⼎㹿䖤ࡼⶶ䞣 09 খ
✻ⱘখ✻ڣഫDŽ
>@⫼Ѣ㓪ⷕᇍ䈵ڣഫ %ON খ✻ⱘ ᐙˈڣՓ⫼ᰒ⼎ᯊࠏ⒲ৢⱘǃ⾏ᕫ᳔䖥ⱘڣ
5HI Ўৢᮍখ✻ˈڣՓ⫼㓪ⷕڣഫ %ON ᯊখ✻䖛ⱘࠡᮍখ✻ ڣ5HI Ўࠡᮍখ✻
ڣDŽ
>@ 䖤ࡼⶶ䞣ⱘ䅵ㅫ؛ᅮڣП䯈䖤ࡼϔᅮ≵᳝䖤ࡼ䖯㸠DŽℸᯊˈབᵰ؛䆒㓪ⷕᇍ
&1% 䇈ᯢк 义
&1% 䇈ᯢк 义
఼Ё≵ֱ᳝ᄬᰒ⼎ᯊࠏ⒲ৢѢ㓪ⷕᇍ䈵ⱘڣˈڣℸϡ㛑Փ⫼Ϟ䗄ҹᕔⱘ⬅ᰒ⼎ᯊࠏ
⒲ৢѢ㓪ⷕᇍ䈵ⱘڣއڣᅮ䖤ࡼⶶ䞣ⱘⳈᓣDŽ
থᯢݙᆍ
>@ ℸˈᴀথᯢህᰃ䡈ѢϞ䗄䯂乬ˈⳂⱘᰃᦤկϔ⾡ᐙখ✻ؐᦦڣ乘⌟ᯊˈ㛑
᳝ᬜഄᅲ⦄㓪ⷕᑊ㛑ࠞޣ໘⧚䞣ⱘࡼᗕڣ㓪ⷕᮍ⊩ঞࡼᗕڣ㾷ⷕᮍ⊩DŽ
>@ Ўњ䖒ࠄϞ䗄Ⳃⱘˈᴀথᯢⱘࡼᗕڣ㓪ⷕᮍ⊩Ўҹڣഫऩԡᇍᵘ៤䕧ܹⱘڣ
ڣ䖯㸠㓪ⷕⱘࡼᗕڣ㓪ⷕᮍ⊩ˈ݊⡍ᕕѢˈࣙᣀ ˖
އᅮখ✻Ꮖ㓪ⷕᅠ↩ⱘڣ䖯㸠
㓪ⷕⱘϾڣഫ᠔݅ৠখ✻ⱘ݅ⱘڣৠখ✻އڣᅮℹ偸 ˗
⫼Ϟ䗄݅ৠখ✻ⱘ⫳ڣ៤
乘⌟ⱘڣ乘⌟⫳ڣ៤ℹ偸 ˗
⫼Ϟ䗄乘⌟ڣᇍ㓪ⷕᇍ䈵ڣഫ䖯㸠㓪ⷕⱘ㓪ⷕℹ偸DŽ
ℸˈ⫼খ✻⫳ڣ៤乘⌟ڣᯊˈ⬅Ѣϡ䳔䖯㸠ᇍ↣ϾڣഫҢᐙᏆ㓪ⷕᅠ
>@
↩ⱘڣЁ䗝ᢽЎখ✻ⱘڣⱘڣ໘⧚ˈℸৃҹࠞޣ໘⧚䞣DŽᑊϨˈ
⬅Ѣϡᖙᇍ↣Ͼ
ڣഫ㓪ⷕ䆹খ✻ˈڣℸৃҹࠞޣҷⷕ䞣DŽϔ㠀ഄˈ
ڣ᭄Ёⱘ䚼ߚⱘڣഫ䗝ᢽⳌৠ
ⱘڣЎ᳔ড়䗖ⱘখ✻ৃⱘڣ㛑ᗻᕜ催DŽℸˈ䗮䖛՟བҹڣഫऩԡՓখ✻ڣЎ݅
ৠⱘখ✻ˈڣ㛑㓈ᣕ催㓪ⷕᬜ⥛ⱘᚙމϟࠞޣ໘⧚䞣DŽ
㗙ˈᴀথᯢⱘࡼᗕڣ㓪ⷕᮍ⊩Ўҹڣഫऩԡᇍᵘ៤䕧ܹⱘڣڣ䖯㸠㓪
>@
ⷕⱘࡼᗕڣ㓪ⷕᮍ⊩ˈ݊⡍ᕕѢˈࣙᣀ ˖އᅮখ✻ ᐙᏆ㓪ⷕᅠ↩ⱘڣ䖯㸠㓪ⷕⱘ
Ͼڣഫ᠔݅ৠখ✻ⱘ ݅ⱘڣৠখ✻އڣᅮℹ偸 ˗ খ✻Ϟ䗄 ڣҢϾڣഫ
Ꮖ㓪ⷕᅠ↩ⱘڣЁ䗝ᢽⱘ ⫳ڣ៤乘⌟ⱘڣ乘⌟⫳ڣ៤ℹ偸 ˗ ⫼Ϟ䗄乘⌟ڣ
ᇍ㓪ⷕᇍ䈵ڣഫ䖯㸠㓪ⷕⱘ㓪ⷕℹ偸DŽ
ℸˈ⫼ ᐙڣЎখ✻⫳ڣ៤乘⌟ڣᯊˈ⬅ѢᇍѢ ᐙখ✻ڣϡ䳔㽕
>@
ᇍ↣ϾڣഫҢᐙᏆ㓪ⷕᅠ↩ⱘڣЁ䗝ᢽ ᐙⱘڣ໘⧚ˈ ℸ㛑ࠞޣ໘⧚䞣DŽᑊϨˈ
⬅Ѣϡᖙᇍ↣Ͼڣഫ㓪ⷕ䆹খ✻ˈڣℸৃҹࠞޣҷⷕ䞣DŽϔ㠀ഄˈ ڣ᭄Ёⱘ䚼ߚ
ⱘڣഫ䗝ᢽⳌৠⱘڣЎ᳔ড়䗖ⱘখ✻ৃⱘڣ㛑ᗻᕜ催DŽℸˈ䗮䖛՟བҹڣഫऩԡ
Փϔᮍⱘখ✻ڣЎ݅ৠⱘখ✻ˈڣ 㛑㓈ᣕ催㓪ⷕᬜ⥛ⱘᚙމϟࠞޣ໘⧚䞣DŽ
>@ 䖭䞠ˈϞ䗄ࡼᗕڣ㓪ⷕᮍ⊩䖬ৃҹࣙᣀᇚ⫼Ѣ⹂ᅮϞ䗄݅ৠখ✻ֵⱘڣᙃ䆄
䗄⫳៤ⱘࡼᗕڣ㓪ⷕ᭄ЁⱘϾڣഫⱘ݅ৠֵᙃऎඳֵⱘݙᙃ䆄䗄ℹ偸DŽℸˈৃ
ҹᇚ⹂ᅮ݅ৠⱘখ✻ֵⱘڣᙃ䆄䗄ࡼᗕڣ㓪ⷕ᭄Ёˈ䖯㸠䕧ߎˈ㾷ⷕࡼᗕڣ
㓪ⷕ᭄ᯊ㛑⹂ޚഄ⹂ᅮখ✻ڣDŽ
ᴀথᯢⱘࡼᗕڣ㾷ⷕᮍ⊩Ўᇍڣҹڣഫऩԡ㹿㓪ⷕৢⱘࡼᗕڣ㓪ⷕ᭄
>@
䖯㸠㾷ⷕⱘࡼᗕڣ㾷ⷕᮍ⊩ˈ݊⡍ᕕѢˈࣙᣀ ˖
އᅮখ✻Ꮖ㾷ⷕᅠ↩ⱘڣ䖯㸠㾷ⷕ
ⱘϾڣഫ᠔݅ৠখ✻ⱘ݅ⱘڣৠখ✻އڣᅮℹ偸 ˗⫼Ϟ䗄݅ৠখ✻ⱘ⫳ڣ៤乘⌟
ⱘڣ乘⌟⫳ڣ៤ℹ偸 ˗
⫼Ϟ䗄乘⌟ڣᇍ㾷ⷕᇍ䈵ڣഫ䖯㸠㾷ⷕⱘ㾷ⷕℹ偸DŽ
>@ ℸˈ㾷ⷕᯊ㛑ℷ⹂ഄ㾷ⷕ໘⧚⫼݅ৠⱘখ✻ڣ㓪ⷕৢ䕧ߎⱘࡼᗕڣ㓪
ⷕ᭄DŽ
㗙ˈᴀথᯢⱘࡼᗕڣ㾷ⷕᮍ⊩Ўᇍڣҹڣഫऩԡ㹿㓪ⷕৢⱘࡼᗕڣ㓪
>@
ⷕ᭄䖯㸠㾷ⷕⱘࡼᗕڣ㾷ⷕᮍ⊩ˈ݊⡍ᕕѢˈࣙᣀ ˖އᅮখ✻ ᐙᏆ㾷ⷕᅠ↩ⱘڣ
䖯㸠㾷ⷕⱘϾڣഫ᠔݅ৠখ✻ⱘ ݅ⱘڣৠখ✻އڣᅮℹ偸 ˗ খ✻Ϟ䗄 ڣ
ҢϾڣഫᏆ㾷ⷕᅠ↩ⱘڣЁ䗝ᢽⱘ ⫳ڣ៤乘⌟ⱘڣ乘⌟⫳ڣ៤ℹ偸 ˗ ⫼
&1% 䇈ᯢк 义
Ϟ䗄乘⌟ڣᇍ㾷ⷕᇍ䈵ڣഫ䖯㸠㾷ⷕⱘ㾷ⷕℹ偸DŽ
>@ ℸˈ㾷ⷕᯊ㛑ℷ⹂ഄ㾷ⷕ໘⧚⫼݅ৠⱘখ✻ڣ↣Ͼڣഫⱘখ✻ڣ㓪
ⷕৢ䕧ߎⱘࡼᗕڣ㓪ⷕ᭄DŽ
>@ 䖭䞠ˈϞ䗄ࡼᗕڣ㾷ⷕᮍ⊩䖬ৃҹࣙᣀҢϞ䗄ࡼᗕڣ㓪ⷕ᭄ЁⱘϾڣഫ
ⱘ݅ৠֵᙃऎඳݙᢑߎ⫼Ѣ⹂ᅮϞ䗄݅ৠⱘখ✻ֵⱘڣᙃⱘֵᙃᢑߎℹ偸DŽℸˈ㛑
Ңࡼᗕڣ㓪ⷕ᭄Ёᢑߎ⹂ᅮ݅ৠⱘখ✻ֵⱘڣᙃˈ㛑⹂ޚഄ⡍ᅮখ✻ڣDŽ
>@ ˈᴀথᯢϡҙৃҹᅲ⦄䖭ḋⱘࡼᗕڣ㓪ⷕᮍ⊩ࡼᗕڣ㾷ⷕᮍ⊩ˈ㗠Ϩ
ৃҹᅲ⦄᳝䖭ḋⱘࡼᗕڣ㓪ⷕᮍ⊩ࡼᗕڣ㾷ⷕᮍ⊩᠔⡍᳝ⱘℹ偸Ў㺙㕂ⱘࡼ
ᗕڣ㓪ⷕ㺙㕂ࡼᗕڣ㾷ⷕ㺙㕂DŽᑊϨˈ䖬ৃҹᅲ⦄䅵ㅫᴎЁᠻ㸠䖭ѯℹ偸ⱘᑣ
㗙ᅲ⦄⫼Ϟ䗄ࡼᗕڣ㓪ⷕᮍ⊩㓪ⷕ䖛ⱘࡼᗕڣ㓪ⷕ᭄DŽᑊϨϡ⫼䇈ˈ䖭ḋⱘᑣ
ࡼᗕڣ㓪ⷕ᭄ৃҹ䗮䖛 &'520 ㄝ䆄ᔩၦԧ⡍㔥ㄝӴ䕧ၦԧথ䗕DŽ
>@ 䰘ⱘㅔ㽕䇈ᯢ
>@ 㸼⼎ҹᕔⱘࡼᗕڣ㓪ⷕ㺙㕂ⱘᵘ៤ⱘᮍḚ ˗
>@ ⫼ᐙখ✻ڣ䖯㸠ᦦؐⱘὖᗉ ˗
>@ ҹᕔⱘࡼᗕڣ㓪ⷕ㺙㕂ⱘࡼᗕڣ㓪ⷕ᭄ⱘḐᓣⱘὖᗉ ˗
>@ 㸼⼎ҹᕔⱘࡼᗕڣ㾷ⷕ㺙㕂ⱘᵘ៤ⱘᮍḚ ˗
>@ ҹᕔⱘⳈᓣⱘ䇈ᯢ ˗
>@ 㸼⼎ᅲᮑᔶᗕ ⱘࡼᗕڣ㓪ⷕ㺙㕂ⱘᵘ៤ⱘᮍḚ ˗
>@ ᅲᮑᔶᗕ ⱘࡼᗕڣ㓪ⷕ᭄ⱘḐᓣⱘὖᗉ ˗
>@ 㸼⼎ᅲᮑᔶᗕ ⱘࡼᗕڣ㾷ⷕ㺙㕂ⱘᵘ៤ⱘᮍḚ ˗
>@ 㸼⼎ᅲᮑᔶᗕ ⱘࡼᗕڣ㓪ⷕ㺙㕂ⱘᵘ៤ⱘᮍḚ ˗
>@ ᅲᮑᔶᗕ ⱘࡼᗕڣ㓪ⷕ᭄ⱘḐᓣⱘὖᗉ ˗
>@ 㸼⼎ᅲᮑᔶᗕ ⱘࡼᗕڣ㓪ⷕ㺙㕂ⱘবᔶ՟ⱘᵘ៤ⱘᮍḚ ˗
>@ ᅲᮑᔶᗕ ⱘবᔶ՟ⱘࡼᗕڣ㓪ⷕ᭄ⱘḐᓣⱘὖᗉ ˗
>@ 㸼⼎ᅲᮑᔶᗕ ⱘࡼᗕڣ㓪ⷕ㺙㕂ⱘবᔶ՟ⱘᵘ៤ⱘᮍḚ ˗
>@ 㸼⼎ᅲᮑᔶᗕ ⱘࡼᗕڣ㾷ⷕ㺙㕂ⱘᵘ៤ⱘᮍḚ ˗
>@ 㸼⼎ᅲᮑᔶᗕ ⱘࡼᗕڣ㾷ⷕ㺙㕂ⱘবᔶ՟ⱘᵘ៤ⱘᮍḚ ˗
>@ ᅲᮑᔶᗕ ⱘᰒ⼎乎ᑣֵᙃ䍙ࠡѢ㓪ⷕᇍ䈵ⱘڣᐙখ✻Ⳉⱘڣ
ᓣⱘ䇈ᯢ ˗
>@ ᅲᮑᔶᗕ ⱘᰒ⼎乎ᑣֵᙃ⒲ৢѢ㓪ⷕᇍ䈵ⱘڣᐙখ✻Ⳉⱘڣ
ᓣⱘ䇈ᯢ ˗
>@ ᅲᮑᔶᗕ ⱘ䏇䎗ᓣᯊⱘڣ䯈乘⌟ⱘ䇈ᯢ ˗
>@ ᇍᄬ⫼ټ䅵ㅫᴎ㋏㒳ᴹᅲ⦄Ϟ䗄ᅲᮑᔶᗕⱘࡼᗕڣ㓪ⷕᮍ⊩ҹঞࡼᗕ
ڣ㾷ⷕᮍ⊩ⱘᑣⱘᄬټၦԧⱘ䇈ᯢˈD Ў㸼⼎њᄬټၦԧⱘᴀԧे䕃⺕Ⲭⱘ⠽⧚
Ḑᓣⱘ՟ᄤⱘ䇈ᯢˈE Ў㸼⼎њҢ䕃⺕Ⲭⱘℷ䴶᠔ⳟࠄⱘ㾖ǃ῾ᮁ䴶㒧ᵘҹঞ䕃⺕Ⲭ
ⱘ䇈ᯢˈF Ў㸼⼎њ⫼Ѣ䕃⺕Ⲭ )' Ϟ䖯㸠Ϟ䗄ᑣⱘ䆄ᔩⱘ⫳ݡᵘ៤ⱘ䇈ᯢ ˗
>@ 㸼⼎ᅲ⦄ݙᆍথ䗕᳡ࡵ఼ⱘݙᆍᦤկ㋏㒳ⱘܼ䚼ᵘ៤ⱘᮍḚ ˗
>@ 㸼⼎⿏ࡼ⬉䆱ᴎⱘϔ՟ⱘㅔ ˗
>@ 㸼⼎⿏ࡼ⬉䆱ᴎⱘݙ䚼ᵘ៤ⱘᮍḚ ˗
&1% 䇈ᯢк 义
>@ 㸼⼎᭄ᄫ᪁ᬒ⫼㋏㒳ⱘᭈԧᵘ៤ⱘᮍḚDŽ
>@ ᴀথᯢⱘ᳔Շᅲᮑᔶᗕ
>@ ϟ䴶খ✻䰘ህᴀথᯢⱘԧᅲᮑᔶᗕ䖯㸠䇈ᯢDŽ
>@ ᅲᮑᔶᗕ
>@ Ў㸼⼎ᴀথᯢⱘᅲᮑᔶᗕ ⱘࡼᗕڣ㓪ⷕ㺙㕂ⱘᵘ៤ⱘᮍḚDŽϢ Ё
㸼⼎ҹᕔⱘࡼᗕڣ㓪ⷕ㺙㕂ⱘᵘ៤ⱘᮍḚⳌخৠࡼⱘऩܗ㗙Ⳍৠⱘࡼ᭄⏏
ࡴⳌৠⱘ䰘ᷛ䆄ˈⳕ⬹݊䇈ᯢDŽˈϟ䴶䇈ᯢⱘᅲᮑᔶᗕⱘࡼᗕڣ㓪ⷕ㺙㕂ঞࡼ
ᗕڣ㾷ⷕ㺙㕂Ёˈৃҹҹڣഫऩԡᇍ⫼ ᐙখ✻ڣ䗮䖛ڣ㋴ᦦؐ⫳៤乘⌟ⱘڣᮍ⊩
ᐙখ✻ؐᦦڣ乘⌟ ǃॳᇕϡࡼഄᇚӏᛣ ᐙڣЁࣙⱘڣഫЎ乘⌟ⱘڣᮍ
⊩ǃҹঞ䗮䖛⬏䴶ݙ乘⌟⫳៤乘⌟ⱘڣᮍ⊩ㄝᮍ⊩䖯㸠ߛᤶDŽ
ࡼᗕڣ㓪ⷕ㺙㕂Ўᇚ䕧ܹⱘڣ᭄ ,PJ ߚࡆ៤ڣഫˈᇍߚࡆⱘ↣Ͼڣഫ䖯㸠
>@
㓪ⷕ໘⧚ⱘ㺙㕂ˈབ ᠔⼎ࣙᣀ䖤ࡼᅮ䚼 ǃ ڣ㋴ᦦؐ䚼 ǃ ఼⊩ޣǃڣ㓪ⷕ䚼
ǃڣ㾷ⷕ䚼 ǃࡴ⊩఼ ǃৃব䭓㓪ⷕ䚼 ǃᏻ㓧 ఼ކঞᓔ݇ DŽ
>@ 㸼⼎⫼ᐙখ✻ؐᦦڣ乘⌟䖯㸠㓪ⷕⱘڣഫ᠔Փ⫼ⱘϔᮍⱘখ✻ⱘڣ咬䅸
খ✻ڣ㓪ো 'HI5HI1Rˈ㹿䕧ܹࠄࡼᗕڣ㓪ⷕ㺙㕂ЁDŽᐙখ✻ؐᦦڣ乘⌟ᯊˈ 䖤ࡼ
ᅮ䚼 ᇚ ᐙখ✻ڣЁⱘ ᐙᅮЎ䕧ܹⱘ咬䅸খ✻ڣ㓪ো 'HI5HI1R ᠔ᣛ⼎ⱘ
খ✻ˈڣ䖯㸠䖤ࡼᅮDŽℸˈ䖤ࡼᅮ䚼 䕧ߎⱘখ✻ڣ㓪ো 5HI1R ⱘؐϢ咬䅸
খ✻ڣ㓪ো 'HI5HI1R ⱘؐЎৠϔؐDŽৃব䭓㓪ⷕ䚼 ᇍ⅟Ꮒ㓪ⷕ᭄ (5HVǃ乘⌟⾡
㉏ 3UHG7\SHǃখ✻ڣ㓪ো 5HI1Rǃ䖤ࡼⶶ䞣 09ǃ09ǃ
咬䅸খ✻ڣ㓪ো 'HI5HI1Rˈ
䕧ߎ
ࡼᗕڣ㓪ⷕ᭄ 6WU 䖯㸠ৃব䭓㓪ⷕDŽ
ϟ䴶ᇍϞ䗄䙷ḋᵘ៤ⱘࡼᗕڣ㓪ⷕ㺙㕂Ёˈ㓪ⷕᇍ䈵ڣഫⱘ乘⌟⾡㉏Ўᐙ
>@
খ✻ؐᦦڣ乘⌟ᯊⱘࡼ䖯㸠䇈ᯢDŽ
>@ 䕧ܹⱘڣ᭄ ,PJ ᣝڣഫऩԡ䕧ܹࠄ䖤ࡼᅮ䚼 ঞ ఼⊩ޣЁDŽ
>@ 䖤ࡼᅮ䚼 އᅮ䕧ܹⱘ㓪ⷕᇍ䈵ڣഫⱘ乘⌟⾡㉏ˈ ᇚ䆹乘⌟⾡㉏䕧ߎ㒭ᓔ݇
ঞৃব䭓㓪ⷕ䚼 DŽᑊϨˈᔧއᅮⱘ乘⌟⾡㉏ 3UHG7\SH Ўᐙখ✻ؐᦦڣ乘⌟ᯊˈ
䖤ࡼᅮ䚼 Փ ᐙখ✻ڣЁⱘ ᐙЎ䕧ܹⱘ咬䅸খ✻ڣ㓪ো 'HI5HI1R ᠔㸼⼎ⱘখ
✻އ߿ߚˈڣᅮϔᐙখ✻ڣঞᇍ䆹 ᐙখ✻ⱘڣ䖤ࡼⶶ䞣DŽ✊ৢˈ䖤ࡼᅮ䚼
ᇚখ✻ڣ㓪ো 5HI1R ঞ䖤ࡼⶶ䞣 09ǃ09 䕧ߎ㒭ᏻ㓧 ఼ކঞৃব䭓㓪ⷕ䚼 ˈ
ᇚখ✻ڣ㓪ো 5HI1R 䕧ߎ㒭ᏻ㓧 ఼ކDŽˈ咬䅸খ✻ڣ㓪ো 'HI5HI1R гৃҹ
Ң䖤ࡼᅮ䚼 䕧ߎࠄৃব䭓㓪ⷕ䚼 ЁDŽ
>@ ⴔˈᏻ㓧 ఼ކᇚϢখ✻ڣ㓪ো 5HI1R 䖤ࡼⶶ䞣 09 Ⳍᇍᑨⱘখ✻
ڣഫ 5HI%ONǃҹঞϢখ✻ڣ㓪ো 5HI1R 䖤ࡼⶶ䞣 09 Ⳍᇍᑨⱘখ✻ڣഫ 5HI%ON 䕧
ߎ㒭ڣ㋴ᦦؐ䚼 DŽڣ㋴ᦦؐ䚼 ⫼ᑇഛؐㄝᇍ Ͼখ✻ڣഫ 5HI%ONǃ 5HI%ON ᠔ᇍ
ᑨⱘڣ㋴ؐ䖯㸠ᦦؐˈ䕧ߎᦦؐڣഫ 5HI3RDŽ䖭䞠ˈ⬅Ѣ䖤ࡼᅮ䚼 އᅮⱘ乘⌟⾡㉏
3UHG7\SH Ўᐙখ✻ؐᦦڣ乘⌟ˈℸᓔ݇ ߛᤶࠄĀāϔջˈᇚᦦؐڣഫ 5HI3R
Ў乘⌟ڣ᭄ 3UHG 䕧ߎ㒭 ఼⊩ޣঞࡴ⊩఼ DŽ
఼⊩ޣҢ䕧ܹⱘڣ᭄ ,PJ Ёޣএ乘⌟ڣ᭄ 3UHGˈЎ⅟Ꮒ᭄ 5HV
>@
䕧ߎ㒭ڣ㓪ⷕ䚼 DŽڣ㓪ⷕ䚼 ᇍ䕧ܹⱘ⅟Ꮒ᭄ 5HV 䖯㸠ℷѸবᤶǃ䞣ᄤ࣪ㄝ
ڣ㓪ⷕ໘⧚ˈЎࣙ䞣ᄤ࣪ᅠ↩ⱘℷѸবᤶ㋏᭄ㄝⱘ⅟Ꮒ㓪ⷕ᭄ (UHVˈ䕧ߎ㒭ڣ
&1% 䇈ᯢк 义
&1% 䇈ᯢк 义
&1% 䇈ᯢк 义
&1% 䇈ᯢк 义
ᑣ⾏㓪ⷕᇍ䈵ڣ᳔䖥ⱘڣЎ咬䅸খ✻ⱘڣᮍ⊩ⱘᷛ䆚ヺ䖯㸠㓪ⷕˈᇍ㸼⼎䗝ᢽ
ֱᄬᏻ㓧఼ކЁⱘᏆ㓪ⷕᅠ↩ⱘڣЁⱘᰒ⼎乎ᑣֵᙃ䍙ࠡѢ㓪ⷕᇍ䈵ڣԚ㓪ⷕ
乎ᑣ⾏᳔݊䖥ⱘڣЎ咬䅸খ✻ⱘڣᮍ⊩ⱘᷛ䆚ヺ䖯㸠㓪ⷕˈᇍ㸼⼎䗝ᢽֱᄬ
ᏻ㓧఼ކЁⱘᏆ㓪ⷕᅠ↩ⱘڣЁⱘᰒ⼎乎ᑣֵᙃ⒲ৢѢ㓪ⷕᇍ䈵ڣԚ㓪ⷕ乎ᑣ⾏݊
᳔䖥ⱘڣЎ咬䅸খ✻ⱘڣᮍ⊩ⱘᷛ䆚ヺ䖯㸠㓪ⷕDŽˈ⫼䆹ᮍ⊩៤ⱘࡼᗕڣ
㓪ⷕ᭄ৃҹ⫼᳝ҹϟ䇈ᯢⱘᅲᮑᔶᗕ ⱘᵘ៤ⱘ㾷ⷕᮍ⊩㾷ⷕDŽ
>@ ᑊϨˈгৃҹϡ㓪ⷕ㸼⼎䗝ᢽϞ䗄咬䅸খ✻ⱘڣᮍ⊩ⱘᷛ䆚ヺˈ㗠Ϣᅲᮑᔶᗕ
ৠḋഄབ ᠔⼎䙷ḋˈᇍ㸼⼎咬䅸খ✻ⱘڣڣ㓪ো 'HI5HI1R 䖯㸠㓪ⷕˈ㗙ᇍ㓪ⷕ
ᇍ䈵ڣ᠔᳝ⱘڣ㓪োϢЎ咬䅸খ✻ڣ㗠䗝ᢽⱘڣ᠔᳝ⱘڣ㓪োⱘⳌᇍ
ⱘᏂߚؐ䖯㸠㓪ⷕˈ㗙ᇍ㸼⼎ⳌᇍᏂߚؐⱘᣛҸㄝֵᙃ䖯㸠㓪ⷕDŽ
Ўℸᯊⱘࡼᗕڣ㓪ⷕ㺙㕂ⱘᮍḚDŽ咬䅸খ✻ڣ㓪ো⫳៤䚼 བ
>@
᠔⼎ᇚ咬䅸খ✻ڣ㓪ো 'HI5HI1R 䕧ߎ㒭ৃব䭓㓪ⷕ䚼 DŽৃব䭓㓪ⷕ䚼 ᇍ⅟
Ꮒ㓪ⷕ᭄ (5HVǃ乘⌟⾡㉏ 3UHG7\SHǃখ✻ڣ㓪ো 5HI1Rǃ䖤ࡼⶶ䞣 09ǃ09 ঞ咬䅸খ✻
ڣ㓪ো 'HI5HI1Rˈ䕧ߎࡼᗕڣ㓪ⷕ᭄ 6WU 䖯㸠ৃব䭓㓪ⷕDŽℸᯊⱘ᭄ḐᓣϢ
᠔⼎ⱘ᭄ḐᓣⳌৠDŽˈ⫼䆹ᮍ⊩៤ⱘࡼᗕڣ㓪ⷕ᭄ৃҹ⫼ᅲᮑᔶᗕ 䇈ᯢ䖛
ⱘᵘ៤ⱘ㾷ⷕᮍ⊩㾷ⷕDŽ
>@ ᅲᮑᔶᗕ
>@ Ўᴀথᯢⱘᅲᮑᔶᗕ ⱘࡼᗕڣ㾷ⷕ㺙㕂ⱘᮍḚDŽˈϢ Ёᅲᮑ
ᔶᗕ ⱘࡼᗕڣ㾷ⷕ㺙㕂ⱘᮍḚⳌخৠࡼⱘऩܗঞⳌৠⱘࡼ᭄⏏ࡴⳌৠⱘ䰘
ᷛ䆄ˈⳕ⬹݊䇈ᯢDŽ
>@ ᴀᅲᮑᔶᗕⱘࡼᗕڣ㾷ⷕ㺙㕂ϡࣙᣀᅲᮑᔶᗕ ⱘᵘ៤Ё᠔⼎ⱘ咬䅸খ✻
ڣ㓪ো㓧ˈ ఼ކপ㗠ҷПⱘᰃࣙᣀ咬䅸খ✻ڣ㓪ো⫳៤䚼 DŽৃব䭓ᑺ㾷ⷕ䚼
ᇍ䕧ܹⱘࡼᗕڣ㓪ⷕ᭄ 6WU 䖯㸠ৃব䭓ᑺ㾷ⷕˈ䕧ߎ⅟Ꮒ㓪ⷕ᭄ (5HVǃ乘⌟⾡
㉏ 3UHG7\SHǃখ✻ڣ㓪ো 5HI1Rǃ䖤ࡼⶶ䞣 09ǃ09DŽ咬䅸খ✻ڣ㓪ো⫳៤䚼 ⫼
Ϣᅲᮑᔶᗕ 䇈ᯢ䖛ⱘ咬䅸খ✻ڣ㓪ো⫳៤䚼 Ⳍৠⱘᮍ⊩⫳៤咬䅸খ✻ڣ㓪ো
'HI5HI1Rˈᇚ䆹咬䅸খ✻ڣ㓪ো 'HI5HI1R Ўখ✻ڣ㓪ো 5HI1R 䕧ߎ㒭䖤ࡼ㸹ٓ䚼
DŽ
>@ བϞ᠔䗄ˈབᵰ䞛⫼ᴀᅲᮑᔶᗕˈ㛑ℷ⹂ഄ㾷ⷕՓ⫼њᅲᮑᔶᗕ 䇈ᯢ䖛ⱘᴀ
থᯢⱘࡼᗕڣ㓪ⷕᮍ⊩ⱘࡼᗕڣ㓪ⷕ㺙㕂㓪ⷕ䖛ⱘࡼᗕڣ㓪ⷕ᭄ 6WUDŽ
>@ ˈᔧᇍ᳝ࣙ⫼Ѣ㸼⼎Ϟ䗄ᅲᮑᔶᗕ ⱘবᔶ՟᠔䆄䗄ⱘ䗝ᢽ咬䅸খ✻ڣ
ⱘᮍ⊩ⱘᷛ䆚ヺ ,GHQW ⱘࡼᗕڣ㓪ⷕ᭄ 6WU 䖯㸠㾷ⷕᯊˈ
ࡼᗕڣ㾷ⷕ㺙㕂ৃҹབϟ
䖭ḋഄᵘ៤DŽ
>@ Ўℸᯊࡼᗕڣ㾷ⷕ㺙㕂ⱘᮍḚˈৃব䭓ᑺ㾷ⷕ䚼 བ ᠔⼎䙷ḋ
ᇍ䕧ܹⱘࡼᗕڣ㓪ⷕ᭄ 6WU 䖯㸠ৃব䭓ᑺ㾷ⷕˈ䕧ߎ⅟Ꮒ㓪ⷕ᭄ (5HVǃ乘⌟⾡㉏
3UHG7\SHǃখ✻ڣ㓪ো 5HI1Rǃ䖤ࡼⶶ䞣 09ǃ09 ঞ⫼Ѣ㸼⼎䗝ᢽ咬䅸খ✻ⱘڣᮍ⊩
ⱘᷛ䆚ヺ ,GHQWDŽ咬䅸 GHIDXOW খ✻ڣ㓪ো⫳៤䚼 ⫼ৃব䭓ᑺ㾷ⷕ䚼 䕧ܹⱘ
ᷛ䆚ヺ ,GHQW ᠔ᣛ⼎ⱘ䗝ᢽ咬䅸খ✻ⱘڣᮍ⊩⫳៤咬䅸খ✻ڣ㓪ো 'HI5HI1Rˈᇚ䆹咬
䅸খ✻ڣ㓪ো 'HI5HI1R Ўখ✻ڣ㓪ো 5HI1R 䕧ߎ㒭䖤ࡼ㸹ٓ䚼 DŽ
>@ 䖭ḋഄˈ㛑ℷ⹂ഄ㾷ⷕ᳝ࣙ⫼Ѣ㸼⼎Ϟ䗄ᅲᮑᔶᗕ 䇈ᯢ䖛ⱘ䗝ᢽ咬䅸খ✻
&1% 䇈ᯢк 义
&1% 䇈ᯢк 义
&1% 䇈ᯢк 义
&1% 䇈ᯢк 义
&1% 䇈ᯢк 义
㕸ᵘ៤ⱘᴀԧ䚼ˈ䕧ߎໄ䷇ⱘᡀໄ఼ㄝໄ䷇䕧ߎ䚼 H[ˈ䕧ܹໄ䷇ⱘ呺ܟ亢ㄝໄ䷇䕧ܹ䚼
H[ˈֱᄬᢡᨘⱘࡼ⬏䴭ℶ⬏䴶ⱘ᭄ǃᬊⱘ䚂ӊⱘ᭄ǃࡼ⬏ⱘ᭄㗙䴭ℶ⬏䴶
ⱘ᭄ㄝǃ㓪ⷕ䖛ⱘ᭄㗙㾷ⷕњⱘ᭄ⱘᄬټၦԧ H[ˈՓᄬټၦԧ H[ 㛑ᅝ㺙
⿏ࡼ⬉䆱ᴎ H[ Ϟⱘষ䚼 H[DŽᄬټၦԧ H[ Ў 6' वㄝᇚ㛑⬉⇨Ϟᬍݭ
⍜䰸ⱘ䴲ᯧ༅ᗻᄬ(( े఼ټ3520 ⬉ৃ᪺䰸ৃ㓪া䇏ᄬⱘ ఼ټϔ⾡ेⶀᯊᄬܗ఼ټӊ
ֱᄬล᭭ԧⱘݙᵘӊDŽ
>@ ⫼ݡ 䇈ᯢ⿏ࡼ⬉䆱ᴎ H[DŽ⿏ࡼ⬉䆱ᴎ H[ 䗮䖛ৠℹᘏ㒓 H[ ᇚ⬉⑤
⬉䏃 H[ǃ᪡䕧ܹࠊ䚼 H[ǃڣ㓪ⷕ䚼 H[ǃⳌᴎষ䚼 H[ǃ/&' ⎆ᰒ⼎
఼ ࠊ䚼 H[ǃڣ㾷ⷕ䚼 H[ǃ⫼ߚ⾏䚼 H[ǃ䆄ᔩ⫳ݡ䚼 H[ǃ䇗ࠊ㾷䇗⬉䏃
H[ ঞໄ䷇໘⧚䚼 H[ Ϣ㒳ᣀഄࠊࣙᣀᰒ⼎䚼 H[ ᪡䬂 H[ ⱘᴀԧ䚼ⱘ
䚼ߚⱘЏࠊ䚼 H[ ѦⳌ䖲DŽ
>@ བᵰ᪡㗙䗮䖛᪡㒧ᴳ⬉䆱ঞՓ⬉⑤䬂໘Ѣᓔⴔ⢊ᗕˈ߭⬉⑤⬉䏃 H[ 䗮䖛
Ң⬉⑤㒭Ͼ䚼ߚᦤկ⬉ਃࡼᏺᨘڣ༈ⱘ᭄ⷕ⿏ࡼ⬉䆱ᴎ H[ˈՓ݊໘ѢᎹ⢊ᗕDŽ
>@ ⿏ࡼ⬉䆱ᴎ H[ ḍ &38ˈ520ˈ
ҹঞ 5$0 ㄝᵘ៤ⱘЏࠊ䚼 H[ ⱘࠊˈ⫼ໄ
䷇໘⧚䚼 H[ ᇚ䇁䷇䗮䆱ᓣᯊໄ䷇䕧ܹ䚼 H[ ᬊ䲚ⱘໄ䷇ֵো䕀ᤶ៤᭄ᄫໄ䷇᭄
ˈ⫼㾷䇗⬉䏃 H[ ᇚ݊䖯㸠ᠽ乥໘⧚ˈ⫼ᬊথ⬉䏃䚼 H[ ᅲᮑ᭄বᤶ໘⧚ঞ乥
⥛䕀ᤶ໘⧚ৢ䗮䖛㒓 H[ থ䗕DŽ㗙ˈ⿏ࡼ⬉䆱ᴎ H[ ᇚ䇁䷇䗮䆱ᓣᯊ⫼㒓
H[ ᬊⱘᬊ᭄ᬒˈᅲᮑ乥⥛䕀ᤶ໘⧚ঞ᭄䕀ᤶ໘⧚ˈ ⫼䇗ࠊ㾷䇗⬉䏃 H[ 䖯
㸠䗚ᠽ乥໘⧚ˈ⫼ໄ䷇໘⧚䚼 H[ 䕀ᤶ៤ᢳໄ䷇᭄Пৢˈ 䗮䖛ໄ䷇䕧ߎ䚼 H[
ᇚ݊䕧ߎDŽ
ˈབᵰ᭄䗮䆃ᓣᯊথ䗕⬉ᄤ䚂ӊˈᴀԧ䚼ⱘ᪡䬂 H[ ⱘ᪡䕧ܹ
>@
ⱘ⬉ᄤ䚂ӊⱘ᭛ᴀ᭄䗮䖛᪡䕧ܹࠊ䚼 H[ 䕧ߎࠄЏࠊ䚼 H[DŽЏࠊ䚼 H[
⫼䇗ࠊ㾷䇗⬉䏃 H[ ᠽ乥᭛ᴀ᭄ˈ⫼ᬊথ⬉䏃䚼 H[ ᅲᮑ᭄বᤶ໘⧚ঞ乥⥛ব
ᤶ໘⧚ৢ䗮䖛㒓 H[ থ䗕㒭キ H[DŽ
>@བᵰ᭄䗮䆃ᓣᯊথ䗕ڣ᭄ˈ߭䗮䖛Ⳍᴎষ䚼 H[ ᇚⳌᴎ䚼 H[
ᢡᨘⱘڣ᭄ᦤկ㒭ڣ㓪ⷕ䚼 H[DŽ㗠Ϩˈϡথ䗕ڣ᭄ⱘᯊˈгৃҹ䗮䖛Ⳍ
ᴎষ䚼 H[ /&' ࠊ䚼 H[ ᇚⳌᴎ䚼 H[ ᢡᨘⱘڣ᭄Ⳉᰒ⼎ᰒ⼎䚼
H[ ϞDŽ
>@ڣ㓪ⷕ䚼 H[ Ўᴀথᯢ䇈ᯢ䖛ⱘࡼᗕڣ㓪ⷕ㺙㕂ⱘᵘӊˈ䗮䖛⫼Ϟ䗄
ᅲᮑᔶᗕভ䗄䖛ⱘࡼᗕڣ㓪ⷕ㺙㕂ЁՓ⫼䖛ⱘ㓪ⷕᮍ⊩य़㓽㓪ⷕⳌᴎ䚼 H[ ᦤկⱘ
ڣ᭄ᇚ݊বᤶ៤㓪ⷕڣ᭄ˈᇚ݊থ䗕㒭⫼ߚ⾏䚼 H[DŽ㗠Ϩˈ䖭ᯊ⿏ࡼ⬉䆱ᴎ
H[ ৠᯊ䗮䖛ໄ䷇໘⧚䚼 H[ ᇚໄ䷇䕧ܹ䚼 H[ Ⳍᴎ䚼ߚ H[ ᢡᨘᯊᬊ䲚ࠄⱘ
ໄ䷇Ў᭄ᄫໄ䷇᭄থ䗕ࠄ⫼ߚ⾏䚼 H[ ЁDŽ
⫼ߚ⾏䚼 H[ ҹ乘ᅮⱘᮍᓣ⫼⫼ڣ㓪ⷕ䚼 H[ ᦤկⱘ㓪ⷕڣ᭄
>@
ໄ䷇໘⧚䚼 H[ ᠔ᦤկⱘໄ䷇᭄ˈ⫼䇗ࠊ㾷䇗⬉䏃 H[ ᠽ乥໘⧚݊㒧ᵰ㦋ᕫⱘ
⫼᭄ˈ⫼ᬊথ⬉䏃䚼 H[ ᅲᮑ᭄বᤶ໘⧚乥⥛বᤶ໘⧚ৢ䗮䖛㒓 H[ থ䗕DŽ
བᵰ᭄䗮䆃ᓣᯊᬊϢЏ义ㄝ䖲ⱘࡼᗕڣ᭛ӊⱘ᭄ˈ߭⫼䇗ࠊ㾷䇗
>@
⬉䏃 H[ 䗚ᠽ乥໘⧚䗮䖛㒓 H[ Ңキ H[ ᬊࠄⱘᬊ᭄ˈᇚ݊㒧ᵰ㦋ᕫⱘ
⫼⫼᭄থ䗕㒭⫼ߚ⾏䚼 H[DŽ
&1% 䇈ᯢк 义
&1% 䇈ᯢк 义
ⷕ䆹খ✻ˈڣℸ㛑ᅲ⦄᳝ᬜⱘ㓪ⷕᑊϨ㛑ࠞޣ໘⧚䞣DŽ
>@ ᑊϨˈབᵰ䞛⫼ᴀথᯢⱘࡼᗕڣ㾷ⷕᮍ⊩ˈ㾷ⷕ⫼݅ৠⱘখ✻ڣ↣Ͼڣ
ഫⱘখ✻ڣ㓪ⷕৢ䕧ߎⱘࡼᗕڣ㓪ⷕ᭄ᯊ㛑ℷ⹂ഄ㾷ⷕ໘⧚DŽ
>@ ᎹϮᑨ⫼ᗻ
>@ བϞ᠔䗄ˈᴀথᯢⱘࡼᗕڣ㓪ⷕᮍ⊩ঞࡼᗕڣ㾷ⷕᮍ⊩ৃҹЎ䗮䖛՟བ⿏
ࡼ⬉䆱ᴎǃ'9' 㺙㕂ঞϾҎ⬉㛥ㄝᇍᵘ៤䕧ܹⱘڣڣ䖯㸠㓪ⷕᑊ䕧ߎࡼᗕڣ㓪ⷕ
᭄ˈ㾷ⷕ䆹ࡼᗕڣ㓪ⷕ᭄ⱘᮍ⊩Փ⫼DŽ
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
&1% 䇈ᯢк䰘 义
INTERNATIONAL TELECOMMUNICATION UNION
ITU-T H.263
TELECOMMUNICATION (02/98)
STANDARDIZATION SECTOR
OF ITU
Characteristics of transmission channels used for other than telephone purposes H.10–H.19
Use of telephone-type circuits for voice-frequency telegraphy H.20–H.29
Telephone circuits or cables used for various types of telegraph transmission or H.30–H.39
simultaneous transmission
Telephone-type circuits used for facsimile telegraphy H.40–H.49
Characteristics of data signals H.50–H.99
CHARACTERISTICS OF VISUAL TELEPHONE SYSTEMS H.100–H.199
INFRASTRUCTURE OF AUDIOVISUAL SERVICES
General H.200–H.219
Transmission multiplexing and synchronization H.220–H.229
Systems aspects H.230–H.239
Communication procedures H.240–H.259
Coding of moving video H.260–H.279
Related systems aspects H.280–H.299
Systems and terminal equipment for audiovisual services H.300–H.399
Summary
This Recommendation specifies a coded representation that can be used for compressing the moving
picture component of audio-visual services at low bit rates. The basic configuration of the video
source coding algorithm is based on Recommendation H.261 and is a hybrid of inter-picture
prediction to utilize temporal redundancy and transform coding of the remaining signal to reduce
spatial redundancy. The source coder can operate on five standardized video source formats:
sub-QCIF, QCIF, CIF, 4CIF and 16CIF, and can also operate using a broad range of custom video
formats.
The decoder has motion compensation capability, allowing optional incorporation of this technique
in the coder. Half pixel precision is used for the motion compensation, as opposed to
Recommendation H.261 where full pixel precision and a loopfilter are used. Variable length coding
is used for the symbols to be transmitted.
In addition to the basic video source coding algorithm, sixteen negotiable coding options are
included for improved compression performance and the support of additional capabilities.
Additional supplemental information may also be included in the bitstream for enhanced display
capability and for external usage.
Source
ITU-T Recommendation H.263 was revised by ITU-T Study Group 16 (1997-2000) and was
approved under the WTSC Resolution No. 1 procedure on the 6th of February 1998.
FOREWORD
ITU (International Telecommunication Union) is the United Nations Specialized Agency in the field of
telecommunications. The ITU Telecommunication Standardization Sector (ITU-T) is a permanent organ of
the ITU. The ITU-T is responsible for studying technical, operating and tariff questions and issuing
Recommendations on them with a view to standardizing telecommunications on a worldwide basis.
The World Telecommunication Standardization Conference (WTSC), which meets every four years,
establishes the topics for study by the ITU-T Study Groups which, in their turn, produce Recommendations
on these topics.
The approval of Recommendations by the Members of the ITU-T is covered by the procedure laid down in
WTSC Resolution No. 1.
In some areas of information technology which fall within ITU-T’s purview, the necessary standards are
prepared on a collaborative basis with ISO and IEC.
NOTE
In this Recommendation, the expression "Administration" is used for conciseness to indicate both a
telecommunication administration and a recognized operating agency.
© ITU 1998
All rights reserved. No part of this publication may be reproduced or utilized in any form or by any means,
electronic or mechanical, including photocopying and microfilm, without permission in writing from the ITU.
2 References .................................................................................................................. 1
1 Scope
This Recommendation specifies a coded representation that can be used for compressing the moving
picture component of audio-visual services at low bit rates. The basic configuration of the video
source coding algorithm is based on Recommendation H.261. Sixteen negotiable coding options are
included for improved performance and increased functionality. This Recommendation contains
Version 2 of Recommendation H.263, which is fully compatible with the original Recommendation,
adding only optional features to the original Version 1 content of the Recommendation.
2 References
The following Recommendations and other references contain provisions which, through reference
in this text, constitute provisions of this Recommendation. At the time of publication, the editions
indicated were valid. All Recommendations and other references are subject to revision; all users of
this Recommendation are therefore encouraged to investigate the possibility of applying the most
recent edition of the Recommendations and other references listed below.
[1] ITU-R Recommendation BT.601-5 (1995), Studio encoding parameters of digital television
for standard 4:3 and wide-screen 16:9 aspect ratios.
Reference [1] is referenced herein to define the (Y, CB, CR) colour space and its 8-bit integer
representation for pictures used by video codecs designed according to this Recommendation.
(Reference [1] is not used to define any other aspects of this Recommendation.)
The following additional ITU-T Recommendations are mentioned for illustration purposes in this
text; however, they do not constitute provisions of this Recommendation. At the time of publication,
the editions indicated were valid. All Recommendations and other references are subject to revision;
all users of this Recommendation are therefore encouraged to investigate the possibility of applying
the most recent edition of the Recommendations and other references listed below. A list of the
currently valid ITU-T Recommendations is regularly published.
[2] ITU-T Recommendation H.223 (1996), Multiplexing protocol for low bit rate multimedia
communication.
[3] ITU-T Recommendation H.242 (1997), System for establishing communication between
audiovisual terminals using digital channels up to 2 Mbit/s.
[4] ITU-T Recommendation H.245 (1998), Control protocol for multimedia communication.
[5] ITU-T Recommendation H.261 (1993), Video codec for audiovisual services at p × 64 kbit/s.
[6] ITU-T Recommendation H.262 (1995) | ISO/IEC 13818-2:1996, Information technology –
Generic coding of moving pictures and associated audio information: Video.
[7] ITU-T Recommendation H.324 (1998), Terminal for low bit rate multimedia
communication.
External control
Coding control
b) Video decoder
T1602680-97
3.6 Buffering
The encoder shall control its output bitstream to comply with the requirements of the hypothetical
reference decoder defined in Annex B. Video data shall be provided on every valid clock cycle. This
can be ensured by MCBPC stuffing (see Tables 7 and 8) or, when forward error correction is used,
also by forward error correction stuffing frames (see Annex H).
The number of bits created by coding any single picture shall not exceed a maximum value specified
by the parameter BPPmaxKb which is measured in units of 1024 bits. The minimum allowable value
of the BPPmaxKb parameter depends on the largest picture size that has been negotiated for use in
the bitstream (see Table 1). The picture size is measured as the picture width times height for the
luminance (Y) component, measured in pixels. An encoder may use a larger value for BPPmaxKb
than as specified in Table 1, provided the larger value is first negotiated by external means, for
example Recommendation H.245.
When the Temporal, SNR, and Spatial Scalability mode (Annex O) is employed, the number of bits
sent for each picture in each enhancement layer shall not exceed the maximum value specified by
BPPmaxKb.
4 Source coder
Table 2/H.263 – Number of pixels per line and number of lines for each of the
standardized H.263 picture formats
Picture Number of pixels Number of lines Number of pixels for Number of lines for
format for luminance (dx) for luminance (dy) chrominance (dx/2) chrominance (dy/2)
sub-QCIF 128 96 64 48
QCIF 176 144 88 72
CIF 352 288 176 144
4CIF 704 576 352 288
16CIF 1408 1152 704 576
T1602690-97
Luminance sample
Chrominance sample
Block edge
Custom picture formats can have a custom pixel aspect ratio as described in Table 3, if the custom
pixel aspect ratio use is first negotiated by external means. Custom picture formats can have any
number of lines and any number of pixels per line, provided that the number of lines is divisible by
four and is in the range [4, ..., 1152], and provided that the number of pixels per line is also divisible
by four and is in the range [4, ..., 2048]. For picture formats having a width or height that is not
divisible by 16, the picture is decoded in the same manner as if the width or height had the next
larger size that would be divisible by 16 and then the picture is cropped at the right and the bottom to
the proper width and height for display purposes only.
All decoders and encoders shall be able to operate using the CIF picture clock frequency. Some
decoders and encoders may also support custom picture clock frequencies. All decoders shall be able
to operate using the sub-QCIF picture format. All decoders shall also be able to operate using the
QCIF picture format. Some decoders may also operate with CIF, 4CIF or 16CIF, or custom picture
formats. Encoders shall be able to operate with one of the picture formats sub-QCIF and QCIF. The
p
CC
t
qz
Video T Q q
in
To
Q–1
video
multiplex
coder
T–1
P v
T1602700-97
T Transform
Q Quantizer
P Picture Memory with motion compensated variable delay
CC Coding control
p Flag for INTRA/INTER
t Flag for transmitted or not
qz Quantizer indication
q Quantizing index for transform coefficients
v Motion vector
The Slice Structured mode is described in Annex K. Slices are similar to GOBs in that they are a
multi-macroblock layer of the syntax, but slices have a more flexible shape and usage than GOBs,
and slices may appear in the bitstream in any order, under certain conditions.
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
Figure 4/H.263 – Arrangement of Groups of Blocks in a CIF picture
1 2
5 6
3 4
T1602710-97
Y CB CR
When in RRU mode, a macroblock relates to 32 pixels by 32 lines of Y and the spatially
corresponding 16 pixels by 16 lines of CB and CR, and each luminance or chrominance block relates
to 16 pixels by 16 lines of Y, CB or CR. Furthermore, a GOB comprises one macroblock row for CIF
and 4CIF, and two macroblock rows for 16CIF.
The macroblock numbering is done by use of horizontal scan of the macroblock rows from left to
right, starting with the upper macroblock row and ending with the lower macroblock row. Data for
the macroblocks is transmitted per macroblock in increasing macroblock number. Data for the blocks
is transmitted per block in increasing block number (see Figure 5).
The criteria for choice of mode and transmitting a block are not subject to recommendation and may
be varied dynamically as part of the coding control strategy. Transmitted blocks are transformed and
resulting coefficients are quantized and entropy coded.
4.2.2 Prediction
The primary form of prediction is inter-picture and may be augmented by motion compensation
(see 4.2.3). The coding mode in which temporal prediction is applied is called INTER; the coding
mode is called INTRA if no temporal prediction is applied. The INTRA coding mode can be
signalled at the picture level (INTRA for I-pictures or INTER for P-pictures) or at the macroblock
level in P-pictures. In the optional PB-frames mode, B-pictures are always coded in INTER mode.
The B-pictures are partly predicted bidirectionally (refer to Annex G).
In total, H.263 has seven basic picture types (of which only the first two are mandatory) which are
defined primarily in terms of their prediction structure:
1) INTRA: A picture having no reference picture(s) for prediction (also called an I-picture);
2) INTER: A picture using a temporally previous reference picture (also called a P-picture);
3) PB: A frame representing two pictures and having a temporally previous reference picture
(see Annex G);
4) Improved PB: A frame functionally similar but normally better than a PB-frame
(see Annex M);
4.2.4 Quantization
Unless the optional Advanced INTRA Coding mode or Modified Quantization mode is in use, the
number of quantizers is 1 for the first coefficient of INTRA blocks and 31 for all other coefficients.
Within a macroblock the same quantizer is used for all coefficients except the first one of INTRA
PSC PEI
TR PSUPP
PTYPE
PQUANT
ESTUF
CPM
TRB ESBI
DBQUANT
PSTUF
T1602720-97
CPFMT TRPI
EPAR TRP
CPCFC BCI
UUI
RPRP LAYER
SSS
T1602730-97
GSTUF TRI
GBSC TR
GN TRPI
GSBI TRP
GFID BCI
GQUANT
BCM LAYER
MB LAYER
T1602740-97
SSTUF TRI
SSC TR
SEPB1 TRPI
SSBI TRP
MBA BCI
SQUANT
MB LAYER
SWI
SEPB3
GFID
T1602750-97
BT BCPM
External framing
URF BSBI
BCM LAYER
TR BEPB1
BSTUF
ELNUMI GN/MBA
ELNUM BEPB2
RTR
T1602760-97
W DA
EP spatial pictures
M VD
Y_FILL
C B _EPB
C B _FILL
C R _EPB
C R _FILL
T1602770-97
I and P pictures,
PB and Improved PB frames EI pictures
B and EP pictures
DQUANT MVDFW
MVD MVDBW
BLOCK LAYER
MVDB INTRA DC
T1602780-97
PSC TR PTYPE PQUANT CPM PSBI TRB DBQUANT PEI PSUPP PEI GOBs ESTUF EOS PSTUF
T1602790-97
... PLUSPTYPE CPM PSBI CPFMT EPAR CPCFC ETR UUI SSS
RPRP ...
T1605280-98
5.1.4.5 Mode restrictions for certain picture types and mode inference rules
Certain modes do not apply to certain types of pictures. Particularly, these restrictions apply:
1) The following modes do not apply within I (INTRA) pictures: Unrestricted Motion Vector
(see Annex D), Advanced Prediction (see Annex F), Alternative INTER VLC (see Annex S),
Reference Picture Resampling (see Annex P), and Reduced-Resolution Update
(see Annex Q).
2) The following modes do not apply within B-pictures (see Annex O): Syntax-based
Arithmetic Coding (see Annex E), Deblocking Filter (see Annex J), and Advanced
Prediction (see Annex F).
3) The following modes do not apply within EI-pictures (see Annex O): Unrestricted Motion
Vector (see Annex D), Syntax-based Arithmetic Coding (see Annex E), Advanced
Prediction (see Annex F), Reference Picture Resampling (see Annex P), Reduced-Resolution
Update (see Annex Q), and Alternative INTER VLC (see Annex S).
4) The following modes do not apply within EP-pictures (see Annex O): Syntax-based
Arithmetic Coding (see Annex E) and Advanced Prediction (see Annex F).
One or more of the modes listed in the above four-item list may have a mode flag having a value "1"
in the optional part of PLUSPTYPE within a picture of a type that is prohibited for that mode (types
I, B, EI, or EP). This condition is allowed and shall be interpreted subject to the mode inference rules
which follow in the next paragraph.
Mode states are subject to the following mode inference rules:
1) Once a mode flag has been set to "1" in the optional part of PLUSPTYPE, the current picture
and each subsequent picture in the bitstream shall be assigned a state of "on" for that mode.
2) An inferred state of "off" shall be assigned to any mode which does not apply within a
picture having the current picture type code. However, each subsequent picture in the
bitstream shall have an inferred state of "on" for that mode (unless this too results in an
obvious conflict – which shall be resolved in the same way). In the case of layered scalable
bitstreams (see Annex O), the mode state shall be inferred only from within the same layer
of the bitstream.
5.1.4.7 The picture header location of CPM (1 bit) and PSBI (2 bits)
The location of the CPM and PSBI fields in the picture header depends on whether or not
PLUSPTYPE is present (see 5.1.20 and 5.1.21). If PLUSPTYPE is present, then CPM follows
immediately after PLUSPTYPE in the picture header. If PLUSPTYPE is not present, then CPM
follows immediately after PQUANT in the picture header. PSBI always follows immediately after
CPM (if CPM = "1").
5.3.2 Macroblock type & Coded Block Pattern for Chrominance (MCBPC) (Variable length)
MCBPC is a variable length codeword giving information about the macroblock type and the coded
block pattern for chrominance. The codewords for MCBPC are given in Tables 7 and 8. MCBPC is
always included in coded macroblocks.
Table 10/H.263 – Macroblock types and included data elements for PB-frames
Picture MB type Name COD MCBPC MOD CBPY CBP DQUANT MVD MVDB MVD2-4
type B B
INTER Not coded – X
INTER 0 INTER X X X X (X) X (X)
INTER 1 INTER+Q X X X X (X) X X (X)
INTER 2 INTER4V X X X X (X) X (X) X
INTER 3 INTRA X X X X (X) X (X)
INTER 4 INTRA+Q X X X X (X) X X (X)
INTER 5 INTER4V+Q X X X X (X) X X (X) X
INTER Stuffing – X X
NOTE 1 – "X" means that the item is present in the macroblock.
NOTE 2 – CBPB and MVDB are only present if indicated by MODB.
NOTE 3 – B-blocks are always coded in INTER mode, even if the macroblock type of the PB-macroblock indicates INTRA.
The coded block pattern for chrominance signifies CB and/or CR blocks when at least one
non-INTRADC transform coefficient is transmitted (INTRADC is the dc-coefficient for INTRA
blocks, see 5.4.1), unless the optional Advanced INTRA Coding mode is in use. CBPCN = 1 if any
non-INTRADC coefficient is present for block N, else 0, for CBPC5 and CBPC6 in the coded block
pattern. If Advanced INTRA Coding is in use, the usage is similar, but the INTRADC coefficient is
indicated in the same way as the other coefficients (see Annex I). Block numbering is given in
Figure 5. When MCBPC = Stuffing, the remaining part of the macroblock layer is skipped. In this
case, the preceding COD = 0 is not related to any coded or not-coded macroblock and therefore the
macroblock number is not incremented. For P-pictures, multiple stuffings are accomplished by
multiple sets of COD = 0 and MCBPC = Stuffing. See Tables 7 and 8.
6 Decoding process
MV2 MV3
MV1 MV
T1602810-97
Advantage is taken of the fact that the range of motion vector component values is constrained. Each
VLC word for MVD represents a pair of difference values. Only one of the pair will yield a
macroblock vector component falling within the permitted range [−16, 15.5]. A positive value of the
horizontal or vertical component of the motion vector signifies that the prediction is formed from
pixels in the previous picture which are spatially to the right or below the pixels being predicted. If
the unrestricted motion vector mode (see Annex D) is used, the decoding of motion vectors shall be
performed as specified in D.2.
The motion vector is used for all pixels in all four luminance blocks in the macroblock. Motion
vectors for both chrominance blocks are derived by dividing the component values of the
macroblock vector by two, due to the lower chrominance format. The component values of the
resulting quarter pixel resolution vectors are modified towards the nearest half pixel position as
indicated in Table 18.
A B
a b
c d
C D
T1602820-97
1 2 6 7 15 16 28 29
3 5 8 14 17 27 30 43
4 9 13 18 26 31 42 44
10 12 19 25 32 41 45 54
11 20 24 33 40 46 53 55
21 23 34 39 47 52 56 61
22 35 38 48 51 57 60 62
36 37 49 50 58 59 63 64
with u, v, x, y = 0, 1, 2, ..., 7
where:
x, y = spatial coordinates in the pixel domain,
u, v = coordinates in the transform domain;
C(u) = 1 / 2 for u = 0, otherwise 1;
C(v) = 1 / 2 for v = 0, otherwise 1.
NOTE – Within the block being transformed, x = 0 and y = 0 refer to the pixel nearest the left and top edges
of the picture respectively.
The arithmetic procedures for computing the inverse transform are not defined, but should meet the
error tolerance specified in Annex A.
6.3.1 Summation
After motion compensation and coefficients decoding (inverse transform included), a reconstruction
is formed for each luminance and chrominance block. For INTRA blocks, the reconstruction is equal
to the result of the inverse transformation. For INTER blocks, the reconstruction is formed by
6.3.2 Clipping
To prevent quantization distortion of transform coefficient amplitudes causing arithmetic overflow in
the encoder and decoder loops, clipping functions are inserted. The clipper operates after the
summation of prediction and reconstructed prediction error on resulting pixel values less than 0 or
greater than 255, changing them to 0 and 255 respectively.
ANNEX A
Inverse transform accuracy specification
A.1 Generate random integer pixel data values in the range −L to +H according to the random
number generator given below ("C" version). Arrange into 8 by 8 blocks. Data set of 10 000 blocks
should each be generated for (L = 256, H = 255), (L = H = 5) and (L = H = 300).
A.2 For each 8 by 8 block, perform a separable, orthonormal, matrix multiply, forward discrete
cosine transform using at least 64-bit floating point accuracy.
7 7
1 ⎡ u⎤ ⎡ v⎤
F (u, v ) =
4
C(u)C(v )∑ ∑ f (x , y)cos⎢⎣π(2 x + 1) 16 ⎥⎦ cos⎢⎣π(2 y + 1)16 ⎥⎦
x =0 y =0
with u, v, x, y = 0, 1, 2, ..., 7
where:
x, y = spatial coordinates in the pixel domain;
u, v = coordinates in the transform domain;
C(u) = 1 / 2 for u = 0, otherwise 1;
C(v) = 1 / 2 for v = 0, otherwise 1.
A.3 For each block, round the 64 resulting transformed coefficients to the nearest integer values.
Then clip them to the range −2048 to +2047. This is the 12-bit input data to the inverse transform.
A.4 For each 8 by 8 block of 12-bit data produced by A.3, perform a separable, orthonormal,
matrix multiply, inverse discrete cosine transform (IDCT) using at least 64-bit floating point
accuracy. Round the resulting pixels to the nearest integer and clip to the range −256 to +255. These
blocks of 8 × 8 pixels are the reference IDCT output data.
A.5 For each 8 by 8 block produced by A.3, apply the IDCT under test and clip the output to the
range −256 to +255. These blocks of 8 × 8 pixels are the test IDCT output data.
A.6 For each of the 64 IDCT output pixels, and for each of the 10 000 block data sets generated
above, measure the peak, mean and mean square error between the reference and the test data.
A.7 • For any pixel, the peak error should not exceed 1 in magnitude.
• For any pixel, the mean square error should not exceed 0.06.
• Overall, the mean square error should not exceed 0.02.
• For any pixel, the mean error should not exceed 0.015 in magnitude.
• Overall, the mean error should not exceed 0.0015 in magnitude.
A.8 All zeros in shall produce all zeros out.
ANNEX B
Hypothetical Reference Decoder
where:
bn is the buffer occupancy just after time tn;
tn is the time the nth coded picture is removed from the HRD buffer;
R(t) is the video bit rate at time t.
HRD buffer
occupancy
(bit) t n +1
∫ R(t )dt
tn
dn+1
B
bn
bn+1
Time
(CIF interval)
tn tn+1 T1602830-97
NOTE – Time (tn+1– tn) is an integer number of CIF picture periods (1/29.97, 2/29.97, 3/29.97,...).
ANNEX D
Unrestricted Motion Vector mode
This annex describes the optional Unrestricted Motion Vector mode of this Recommendation. The
capability of this H.263 mode is signalled by external means (for example, Recommendation H.245).
The use of this mode is indicated in PTYPE or PLUSPTYPE.
The range of motion vectors and the VLC table used for coding the motion vector differences for the
Unrestricted Motion Vector mode depend on whether the PLUSPTYPE field of the picture header is
present. When PLUSPTYPE is present, the range of motion vectors also depends on the picture size
and the value of the UUI field in the picture header.
⎧= 0 if y < 0;
⎪
y′ ⎨= 143 if y > 143;
⎪= y otherwise;
⎩
In the Reduced-Resolution Update mode, the specified range applies to the pseudo motion vectors.
This implies that the resulting actual motion vector range is enlarged to approximately double size
(see also Annex Q).
If UUI is set to "01", the motion vectors are not limited except by their distance to the coded area
border as explained in D.1.1. The same limitation applies to the actual motion vectors (not just the
pseudo motion vectors) in the Reduced-Resolution Update mode.
To encode the motion vectors when PLUSPTYPE is present, Table D.3 is used to encode the
difference between the motion vector and the motion vector prediction. Every entry in Table D.3 has
a single value (in contrast to Table 14). The motion vector range and the use of Table D.3 to encode
motion vector data apply to all picture types when PLUSPTYPE is present.
Motion vectors differences are always coded as a pair of a horizontal and a vertical component. If a
pair equals (0.5, 0.5) six consecutive zeros are produced. To prevent start code emulation, this
occurrence shall be followed by one bit set to "1". This corresponds to sending one additional zero
motion vector component.
Table D.3 is a regularly constructed reversible table. Each row represents an interval of motion
vector differences in half-pixel units. The bits "…x1x0" denote all bits following the leading "1" in
the binary representation of the absolute value of the motion vector difference. The bit "s" denotes
the sign of the motion vector difference, "0" for positive and "1" for negative. The binary
representation of the motion vector difference is interleaved with bits that indicate if the code
continues or ends. For example, the motion vector difference −13 has sign s = 1 and binary
representation 1x2x1x0 = 1101. It is encoded as 0 x21 x11 x01 s0 = 0 11 01 11 10. The 0 in the second
position of the last group of two bits indicates the end of the code.
ANNEX E
Syntax-based Arithmetic Coding mode
E.1 Introduction
In the Variable Length Coding/Decoding (VLC/VLD) as described in clause 5, a symbol is VLC
encoded using a specific table based on the syntax of the coder. This table typically stores lengths
and values of the VLC code words. The symbol is mapped to an entry of the table in a table look-up
operation, and then the binary code word specified by the entry is sent out normally to a buffer for
transmitting to the receiver. In VLD decoding, the received bitstream is matched entry by entry in a
specific table based on the syntax of the coder. This table must be the same as the one used in the
encoder for encoding the current symbol. The matched entry in the table is then mapped back to the
corresponding symbol which is the end result of the VLD decoder and is then used for recovering the
video pictures. This VLC/VLD process implies that each symbol must be encoded into a fixed
integral number of bits. Removing this restriction of fixed integral number of bits for symbols can
lead to reductions of resulting bit rates, which can be achieved by arithmetic coding.
#define q1 16384
#define q2 32768
#define q3 49152
#define top 65535
low *= 2;
high = 2 * high+1;
}
}
The values of low, high and opposite_bits are initialized to 0, top and 0, respectively.
PSC_FIFO is a FIFO for buffering the output bits from the arithmetic encoder. The model is
specified through cumul_freq[ ], and the symbol is specified using its index in the model.
low *= 2;
high = 2 * high + 1;
get bit from PSC_FIFO;
code_value = 2 * code_value + bit;
}
return (index–1);
}
Again the model is specified through cumul_freq[ ]. The decoded symbol is returned through
its index in the model. PSC_FIFO is a FIFO for buffering the incoming bitstream. The decoder is
initialized to start decoding an arithmetic coded bitstream by calling the following procedure.
void decoder_reset( )
{
code_value = 0;
low = 0;
high = top;
for (int i = 1; i <= 16; i++) {
get bit from PSC_FIFO;
code_value = 2 * code_value + bit;
}
}
E.4 Syntax
As in VLC table mode of this Recommendation, the syntax of the symbols is partitioned into four
layers: Picture, Group of Blocks, Macroblock and Block. The syntax of the top three layers remain
exactly the same. The syntax of the Block layer also remains quite similar, but is illustrated in
Figure E.1.
T1602840-97
In Figure E.1, TCOEF1, TCOEF2, TCOEF3 and TCOEFr are Last-Run-Level symbols as defined in
5.4.2, and are the possible 1st, 2nd, 3rd and rest of the symbols, respectively. TCOEF1, TCOEF2,
TCOEF3 and TCOEFr are only present when one, two, three or more coefficients are present in the
block layer, respectively.
E.5 PSC_FIFO
PSC_FIFO in encoder or decoder is a FIFO of size > 17 bits. In PSC_FIFO of encoder, illegal
emulations of PSC and GBSC are located and are avoided by stuffing a "1" after each successive
appearance of 14 "0"s (which are not part of PSC or GBSC). In PSC_FIFO of the decoder, the first
"1" after each string of 14 "0"s is deleted; if instead a string of 14 "0"s is followed by a "0", it
indicates that a legal PSC or GBSC is detected. The exact location of the PSC or GBSC is
determined by the next "1" following the string of zeros.
ANNEX F
Advanced Prediction mode
F.1 Introduction
This annex describes the optional Advanced Prediction mode of this Recommendation, including
overlapped block motion compensation and the possibility of four motion vectors per macroblock.
The capability of this mode is signalled by external means (for example, Recommendation H.245).
The use of this mode is indicated in PTYPE. In the Advanced Prediction mode, motion vectors are
allowed to cross picture boundaries as is the case in the Unrestricted Motion Vector mode (for the
description of this technique, refer to D.1). The extended motion vector range feature of the
Unrestricted Motion Vector mode is not automatically included in the Advanced Prediction mode,
and only is active if the Unrestricted Motion Vector mode is selected. If the Advanced Prediction
mode is used in combination with the PB-frames mode, overlapped motion compensation is only
used for prediction of the P-pictures, not for the B-pictures.
MV1 MV MV1 MV
MV1 MV MV1 MV
T1602850-97
Figure F.1/H.263 – Redefinition of the candidate predictors MV1, MV2 and MV3
for each of the luminance blocks in a macroblock
If four vectors are used, each of the motion vectors is used for all pixels in one of the four luminance
blocks in the macroblock. The numbering of the motion vectors is equivalent to the numbering of the
four luminance blocks as given in Figure 5. Motion vector MVDCHR for both chrominance blocks is
derived by calculating the sum of the four luminance vectors and dividing this sum by 8; the
component values of the resulting sixteenth pixel resolution vectors are modified towards the nearest
half-pixel position as indicated in Table F.1.
Resulting position 0 0 0 1 1 1 1 1 1 1 1 1 1 1 2 2 /2
Half-pixel values are found using bilinear interpolation as described in 6.1.2. In Advanced Prediction
mode, the prediction for luminance is obtained by overlapped motion compensation as described in
F.3. The prediction for chrominance is obtained by applying the motion vector MVDCHR to all pixels
in the two chrominance blocks (as it is done in the default prediction mode).
r ( x , y ) = p( x + MV 1 x , y + MV 1 y ),
s( x , y ) = p( x + MV 2 x , y + MV 2 y ),
4 5 5 5 5 5 5 4
5 5 5 5 5 5 5 5
5 5 6 6 6 6 5 5
5 5 6 6 6 6 5 5
5 5 6 6 6 6 5 5
5 5 6 6 6 6 5 5
5 5 5 5 5 5 5 5
4 5 5 5 5 5 5 4
2 2 2 2 2 2 2 2
1 1 2 2 2 2 1 1
1 1 1 1 1 1 1 1
1 1 1 1 1 1 1 1
1 1 1 1 1 1 1 1
1 1 1 1 1 1 1 1
1 1 2 2 2 2 1 1
2 2 2 2 2 2 2 2
Figure F.3/H.263 – Weighting values, H1, for prediction with motion vectors
of the luminance blocks on top or bottom of current luminance block
ANNEX G
PB-frames mode
G.1 Introduction
This annex describes the optional PB-frames mode of this Recommendation. The capability of this
mode is signalled by external means (for example, Recommendation H.245). The use of this mode is
indicated in PTYPE.
A PB-frame consists of two pictures being coded as one unit. The name PB comes from the name of
picture types in Recommendation H.262 where there are P-pictures and B-pictures. Thus, a PB-frame
consists of one P-picture which is predicted from the previous decoded P-picture and one B-picture
which is predicted both from the previous decoded P-picture and the P-picture currently being
decoded. The name B-picture was chosen because parts of B-pictures may be bidirectionally
predicted from the past and future pictures. The prediction process is illustrated in Figure G.1.
An improved version of the PB-frames mode, termed the "Improved PB-frames mode," is described
in Annex M. The PB-frames mode as described in this annex is retained herein only for purposes of
compatibility with systems designed prior to the adoption of the Improved PB-frames mode. For this
reason, the PB-frames mode as described in this annex cannot be used with the additional features of
the syntax which require the use of PLUSPTYPE.
P B P
T1602860-97
Forward
prediction
Bidirectional
prediction
B-block
T1602870-97
P-macroblock
Bidirectional prediction is used for pixels where the backward vector – MVB – points inside PREC.
These pixels are defined by the following procedures which are specified in C language.
Definitions:
nh Horizontal position of block within a macroblock (0 or 1).
nv Vertical position of block within a macroblock (0 or 1).
mh(nh,nv) Horizontal vector component of block (nh,nv) in half-pixel units.
mv(nh,nv) Vertical vector component of block (nh,nv) in half-pixel units.
mhc Horizontal chrominance vector component.
mvc Vertical chrominance vector component.
ANNEX H
Forward error correction for coded video signal
H.1 Introduction
This annex describes an optional forward error correction method (code and framing) for
transmission of H.263 encoded video data. This forward error correction may be used in situations
where no forward error correction is provided by external means, for example at the multiplex or
system level. It is not used for Recommendation H.324. Both the framing and the forward error
correction code are the same as in Recommendation H.261.
S1 S2 S3 S7 S8
S1 Data Parity
1 493 18
Fi
1 Coded data
T1602880-97
1 492
ANNEX I
Advanced INTRA Coding mode
This annex describes the optional Advanced INTRA Coding mode of this Recommendation. The
capability of this H.263 mode is signalled by external means (for example, Recommendation H.245).
The use of this mode is indicated in the PLUSPTYPE field of the picture header.
I.1 Introduction
This optional mode alters the decoding of macroblocks of type "INTRA" (macroblocks of other types
are not affected). The coding efficiency of INTRA macroblocks is improved by using:
1) INTRA-block prediction using neighbouring INTRA blocks for the same component (Y, C B,
or CR);
2) modified inverse quantization for INTRA coefficients; and
3) a separate VLC for INTRA coefficients.
A particular INTRA-coded block may be predicted from the block above the current block being
decoded, from the block to the left of the current block being decoded, or from both. Special cases
exist for situations in which the neighbouring blocks are not INTRA coded or are not in the same
video picture segment. Block prediction always uses data from the same luminance or color
difference component (Y, CB, or CR) as the block being decoded. In prediction, DC coefficients are
always predicted in some manner. The first row of AC coefficients may be predicted from those in
the block above, or the first column of AC coefficients may be predicted from those in the block to
the left, or only the DC coefficient may be predicted as an average from the block above and the
block to the left, as signalled on a macroblock-by-macroblock basis. The remaining AC coefficients
are never predicted. Inverse quantization of the INTRADC coefficient is modified to allow a varying
quantization step size, unlike in the main text of this Recommendation where a fixed step size of 8 is
used for INTRADC coefficients. Inverse quantization of all INTRA coefficients is performed
without a "dead-zone" in the quantizer reconstruction spacing.
Block
MVD MVD2 MVD3 MVD4 MVDB
data T1602890-97
1 2 3 4 11 12 13 14 1 5 7 21 23 37 39 53
5 6 9 10 18 17 16 15 2 6 8 22 24 38 40 54
7 8 20 19 27 28 29 30 3 9 20 25 35 41 51 55
21 22 25 26 31 32 33 34 4 10 19 26 36 42 52 56
23 24 35 36 43 44 45 46 11 18 27 31 43 47 57 61
37 38 41 42 47 48 49 50 12 17 28 32 44 48 58 62
39 40 51 52 57 58 59 60 13 16 29 33 45 49 59 63
53 54 55 56 61 62 63 64 14 15 30 34 46 50 60 64
Figure I.2/H.263 – Alternate DCT scanning patterns for Advanced INTRA coding
For INTRA-coded blocks, if Prediction Mode = 0, the zigzag scan shown in Figure 14 is selected for
all blocks in the macroblock, otherwise, the prediction direction is used to select a scan for the
macroblock.
Block A
RecA´(u,v)
0
• •
1
2
Block B Block C
RecB´(u,v) 3 RecC´(u,v)
7
T1602900-97
The definitions of MCBPC and CBPY are changed when Advanced INTRA coding is in use. When
Advanced INTRA coding is in use, INTRADC transform coefficients are no longer handled as a
separate case, but are instead treated in the same way as the AC coefficients in regard to MCBPC and
CBPY. This means that a zero INTRADC will not be coded as a LEVEL, but will simply increase
the run for the following AC coefficients.
The inverse quantization process for the B-part of an Improved PB frame (see Annex M) is not
altered by the use of the Advanced INTRA coding mode.
Define RecC(u,v) to be the reconstructed coefficient residuals of the current block. For all INTRA
coefficients, the reconstructed residual value is obtained by:
RecC(u,v) = 2 * QUANT * LEVEL(u,v) u = 0, …, 7, v = 0, …, 7.
NOTE – LEVEL(u,v) denotes a quantity having both a magnitude and a sign in the above equation.
Define RecC´(u,v) to be the final reconstructed coefficient values of the current block (after
adjustments for prediction, oddification as described below, and clipping). The final reconstructed
coefficient values RecC´(u,v) are recovered by adding RecC(u,v) to the appropriate prediction as
signalled in the INTRA_MODE field, altering the least-significant bit if needed for oddification of
the DC coefficient, and clipping.
If (block A is INTRA coded and is in the same video picture segment as block C) {
tempDC = RecC(0,0) + RecA´(0,0)
RecC´(u,0) = clipAC( RecC(u,0) + RecA´(u,0) ) u = 1, …, 7,
RecC´(u,v) = clipAC( RecC(u,v) ) u = 0, …, 7, v = 1, …, 7.
} else {
tempDC = RecC(0,0) + 1024
RecC´(u,v) = clipAC( RecC(u,v) ) (u,v) ≠ (0,0), u = 0, …,7, v = 0, …, 7
}
RecC´(0,0) = oddifyclipDC( tempDC )
If (block B is INTRA coded and is in the same video picture segment as block C) {
tempDC = RecC(0,0) + RecB´(0,0)
RecC´(0,v) = clipAC( RecC(0,v) + RecB´(0,v) ) v = 1, …, 7,
RecC´ (u,v) = clipAC( RecC(u,v) ) u = 1, …, 7, v = 0, …, 7.
} else {
tempDC = RecC(0,0) + 1024
RecC´(u,v) = clipAC( RecC(u,v) ) (u,v) ≠ (0,0), u = 0, …, 7, v = 0, …, 7
}
RecC'(0,0) = oddifyclipDC( tempDC )
ANNEX J
Deblocking Filter mode
J.1 Introduction
This annex describes the use of an optional block edge filter within the coding loop. The main
purpose of the block edge filter is to reduce blocking artifacts. The filtering is performed on 8 × 8
block edges. Motion vectors may have either 8 × 8 or 16 × 16 resolution (see J.2). The processing
described in this annex applies only for the P-, I-, EP-, or EI-pictures or the P-picture part of an
Improved PB-frame. (Possible filtering of B-pictures or the B-picture part of an Improved PB-frame
is not a matter for standardization; however, some type of filtering is recommended for improved
picture quality.) The capability of this mode is signalled by external means (for example,
Recommendation H.245). The use of this mode is indicated in the PLUSPTYPE field of the picture
header.
block1
A
B
Block boundary
C
Example for filtered pixels
A B C D D
on a vertical block edge
T1602910-97
One or both of the following conditions must be fulfilled in order to apply the filter across a
particular edge:
– Condition 1: block1 belongs to a coded macroblock (COD==0 || MB-type == INTRA); or
– Condition 2: block2 belongs to a coded macroblock (COD==0 || MB-type == INTRA).
If filtering is to be applied across the edge, A, B, C, and D shall be replaced by A1, B1, C1, D1
where:
B1 = clip(B + d1)
C1 = clip(C − d1)
A1 = A − d2
D1 = D + d2
d = (A−4B+4C−D) / 8
d1 = UpDownRamp(d, STRENGTH)
d2 = clipd1((A − D) / 4, d1/2)
UpDownRamp(x, STRENGTH) = SIGN(x) * (MAX(0, abs(x)−MAX(0, 2*(abs(x) −
STRENGTH))))
The function clip(x) is defined according to 6.3.2, and the function clipd1(x, lim) clips x to the range
± abs(lim). The symbol "/" denotes division by truncation toward zero.
Figure J.2 shows how the value of d1 varies as a function of d. As a result, the filter has an effect
only if d is smaller than 2*STRENGTH (and different from zero). This is to prevent the filtering of
strong true edges in the picture content. However, if the Reduced-Resolution Update mode is in use,
STRENGTH is set to infinity, and as a result, the value of d1 is always equal to the value of d
(see Q.7.2).
Strength 2*Strength d
T1602920-97
The definition of d1 is designed to ensure that small mismatches between the encoder and decoder
will remain small and will not build up over multiple pictures of a video sequence. This would be a
problem, for example, with a condition that simply switches the filter on or off, because a mismatch
of only ±1 for d could then cause the filter to be switched on at the encoder side and off at the
decoder side, or vice versa.
Due to rounding effects, the order of edges where filtering is performed must be specified.
Filtering across horizontal edges:
⎛ A⎞
⎜ ⎟
B
Basically this process is assumed to take place first. More precisely, the pixels ⎜⎜ ⎟⎟ that are used in
C
⎜ ⎟
⎝ D⎠
filtering across a horizontal edge shall not have been influenced by previous filtering across a vertical
edge.
Filtering across vertical edges:
Before filtering across a vertical edge using pixels (A, B, C, D), all modifications of pixels
(A, B, C, D) resulting from filtering across a horizontal edge shall have taken place.
Note that if one or more of the pixels (A, B, C, D) taking part in a filtering process are outside of a
picture, no filtering takes place. Also, if the Independent Segment Decoding mode is in use
(see Annex R) and one or more of the pixels (A, B, C, D) taking part in a filtering process are in
different video picture segments (see I.3 concerning when a block is considered to be in the same
video picture segment), no filtering is performed.
K.1 Introduction
This annex describes the optional Slice Structured mode of this Recommendation. The capability of
this H.263 mode is signalled by external means (for example, Recommendation H.245). The use of
this mode is indicated in the PLUSPTYPE field of the picture header. In order to facilitate optimal
usage in a number of environments, this mode contains two submodes which are also capable of
being signalled by external means (for example, Recommendation H.245). These two submodes are
used to indicate whether or not rectangular slices will be used and/or whether the slices will be
transmitted in sequential order or sent in an arbitrary order.
A slice is defined as a slice header followed by consecutive macroblocks in scanning order. An
exception is for the slice that immediately follows the picture start code in the bitstream for a picture
(which is not necessarily the slice starting with macroblock 0). In this case only part of the slice
header is transmitted as described in K.2. The slice layer defines a video picture segment, and is used
in place of the GOB layer in this optional mode. A slice video picture segment starts at a macroblock
boundary in the picture and contains a number of macroblocks. Different slices within the same
picture shall not overlap with each other, and every macroblock shall belong to one and only one
slice.
This mode contains two submodes, which are signalled in the SSS field of the picture header:
1) The Rectangular Slice submode (RS): When RS is in use, the slice shall occupy a
rectangular region of width specified by the SWI parameter of the slice header in units of
macroblocks, and contains a number of macroblocks in scanning order within the rectangular
region. When the Rectangular Slice submode is not in use, the SWI field is not present in the
slice header and a slice contains a number of macroblocks in scanning order within the
picture as a whole.
2) The Arbitrary Slice Ordering submode (ASO): When ASO is in use, the slices may appear in
any order within the bitstream. When ASO is not in use, the slices must be sent in the
(unique) order for which the MBA field of the slice header is strictly increasing from each
slice to each subsequent slice in the picture.
Slice boundaries are treated differently from simple macroblock boundaries in order to allow slice
header locations within the bitstream to act as resynchronization points for bit error and packet loss
recovery and in order to allow out-of-order slice decoding within a picture. Thus, no data
dependencies can cross the slice boundaries within the current picture, except for the Deblocking
Filter mode which, when in use without the Independent Segment Decoding mode, filters across the
boundaries of the blocks in the picture. However, motion vectors within a slice can cause data
dependencies which cross the slice boundaries in the reference picture used for prediction purposes,
unless the optional Independent Segment Decoding mode is in use.
The following rules are adopted to ensure that slice boundary locations can act as resynchronization
points and to ensure that slices can be sent out of order without causing additional decoding delays:
1) The prediction of motion vector values are the same as if a GOB header were present
(see 6.1.1), preventing the use of motion vectors of blocks outside the current slice for the
prediction of the values of motion vectors within the slice.
2) The Advanced INTRA Coding mode (see Annex I) treats the slice boundary as if it were a
picture boundary with respect to the prediction of INTRA block DCT coefficient values.
SSTUF SSC SEPB1 SSBI MBA SEPB2 SQUANT SWI SEPB3 GFID Macroblock Data
Refer to 5.2.5 for GFID, and to 5.3 for description of Macroblock layer.
ANNEX L
Supplemental Enhancement Information Specification
L.1 Introduction
This annex describes the format of the supplemental enhancement information sent in the PSUPP
field of the picture layer of this Recommendation. The capability of a decoder to provide any or all of
the enhanced capabilities described in this annex may be signalled by external means (for example
Recommendation H.245). Decoders which do not provide the enhanced capabilities may simply
discard any PSUPP information bits that appear in the bitstream. The presence of this supplemental
enhancement information is indicated in PEI, and an additional PEI bit is inserted between every
octet of PSUPP data, as described in 5.1.24 and 5.1.25.
In this annex, a distinction is made between the "decoded picture" and the "displayed picture". For
purposes of this annex, the "displayed picture" is a picture having the same picture format as
specified for the current picture by the picture layer of the video bitstream syntax. The "displayed
picture" is constructed as described in this annex from the decoded picture, the prior displayed
picture, the supplementary enhancement information described herein, and in some cases partly from
a background picture which is externally controlled.
L.3 Do Nothing
No action is requested by the Do Nothing function. This function is used to prevent start code
emulation. Whenever the last five or more bits of the final octet of the previous PSUPP octet are all
zero and no additional PSUPP function requests are to be sent, the Do Nothing function shall be
inserted into PSUPP to prevent the possibility of start code emulation. The Do Nothing function may
also be sent when it is not required by rule expressed in the previous sentence. DSIZE shall be zero
for the Do Nothing function.
ANNEX M
Improved PB-frames mode
M.1 Introduction
This annex describes an optional Improved PB-frames mode of this Recommendation. It is therefore
considered to be advantageous to use the present Improved PB-frames mode instead of the
PB-frames mode defined in Annex G. The capability of this mode is signalled by external means
(for example, Recommendation H.245). The use of this mode is indicated in the PLUSPTYPE field
of the picture header.
Most parts of this option are similar to the PB-frame option defined in Annex G.
To avoid confusion with B-pictures as defined in Annex O, the terms B-picture, B-macroblock, and
B-block will not be used in this annex. Instead, we will use the notation BPB to mean the "B-part" of
an Improved PB-frame. When reference is made to Annex G, B-picture and B-block shall be read as
BPB-picture and BPB-block.
The main difference between the PB-frames mode and the Improved PB-frames mode is that in the
Improved PB-frames mode, the BPB-macroblock has available a forward and a backward prediction
mode in addition to the bidirectional prediction mode. In this annex, MVDB (when present) refers to
a forward motion vector. (Note that, in Annex G, MVDB was used to refer to an enhancement of the
ANNEX N
Reference Picture Selection mode
N.1 Introduction
This annex describes the optional Reference Picture Selection mode of this Recommendation, which
operates using a modified interframe prediction method called "NEWPRED." The capability of this
H.263 mode is signalled by external means (for example, Recommendation H.245). The amount of
additional picture memory accommodated in the decoder may also be signalled by external means to
help the memory management at the encoder. This mode can use backward channel messages sent
from a decoder to an encoder to inform the encoder which part of which pictures have been correctly
decoded at the decoder. The use of this mode is indicated in the PLUSPTYPE field of the picture
header. This mode has two back-channel mode switches which define whether a backward channel is
used and what kind of messages are returned on that backward channel from the decoder, and has
another submode defined in terms of the channel for the backward channel messages.
The two back-channel mode switches of this mode determine the type of messages sent on the back-
channel, specifying whether ACK (Acknowledgment messages), or NACK (Non-Acknowledgment
messages) are sent. Together, the two switches define four basic methods of operation:
1) NEITHER, in which no back-channel data is returned from the decoder to the encoder;
2) ACK, in which the decoder returns only acknowledgment messages;
3) NACK, in which the decoder returns only non-acknowledgment messages; and
4) ACK+NACK, in which the decoder returns both acknowledgment and non-acknowledgment
messages.
The specific type of messages to be sent as outlined above is indicated in the picture header.
There are also two methods of operation in terms of the channel for backward channel messages:
1) Separate Logical Channel mode: This method of operation delivers back-channel data
through a separate logical channel in the multiplex layer of the system, and
2) VideoMux mode: This method of operation delivers back-channel data for received video
within the forward video data of a video stream of encoded data.
This annex specifies a syntax for the backward channel messages as well as for forward-channel
data.
p
CC
t
qz
Video T Q q
in
To
Q–1 video
multiplex
T–1 coder
P AP1
AP2
APn
T1602930-97
T Transform
Q Quantizer
P Picture Memory with motion compensated variable delay
AP Additional Picture Memory
CC Coding control
p Flag for INTRA/INTER
t Flag for transmitted or not
qz Quantizer indication
q Quantizing index for transform coefficients
v Motion vector
N.4 Syntax
T1602940-97
When the optional Slice Structured mode (see Annex K) is in use, the syntax of the slice layer is
modified in the same way as the GOB layer. The syntax is illustrated Figure N.3.
T1602950-97
BT URF TR ELNUMI ELNUM BCPM BSBI BEPB1 GN/MBA BEPB2 RTR BSTUF
ANNEX O
Temporal, SNR, and Spatial Scalability mode
This annex describes the optional mode of this Recommendation in support of Temporal, SNR, and
Spatial Scalability. This mode may also be used in conjunction with error control schemes. The
capability of this mode and the extent to which its features are supported is signaled by external
means (for example, Recommendation H.245). The use of this mode is indicated in PLUSPTYPE.
O.1 Overview
Scalability allows for the decoding of a sequence at more than one quality level. This is done by
using a hierarchy of pictures and enhancement pictures partitioned into one or more layers. There are
three types of pictures used for scalability: B-, EI-, and EP-pictures, as explained below. Each of
these has an enhancement layer number ELNUM that indicates to which layer it belongs, and a
reference layer number RLNUM that indicates which layer is used for its prediction. The lowest layer
is called the base layer, and has layer number 1.
Scalability is achieved by three basic methods: temporal, SNR, and spatial enhancement.
I1 B2 P3 B4 P5
T1602960-97
The location of B-pictures in the bitstream is in a data-dependence order rather than in strict temporal
order. (This rule is consistent with the ordering of other pictures in the bitstream, but for all picture
types other than the B-picture, no such conflict arises between the data-dependence order and the
temporal order.) For example, if the pictures of a video sequence were numbered 1, 2, 3, …, then the
bitstream order of the encoded pictures would be I1, P3, B2, P5, B4, …, where the subscript refers to
the original picture number (as illustrated in Figure O.1).
There is no limit to the number of B-pictures that may be inserted between pairs of reference pictures
in the reference layer (other than what is necessary to prevent temporal ambiguity from overflows of
the temporal reference field in the picture header). However, a maximum number of such pictures
may be signaled by external means (for example, Recommendation H.245).
The picture height, width, and pixel aspect ratio of a B-picture shall always be equal to those of its
temporally subsequent reference layer picture.
Motion vectors are allowed to extend beyond the picture boundaries of B-pictures.
Enhancement EI EP EP
layer
Base I P P
layer
T1602970-97
For both EI- and EP-pictures, the prediction from the reference layer uses no motion vectors.
However, as with normal P-pictures, EP-pictures use motion vectors when predicting from their
temporally prior reference picture in the same layer.
Base
I P P
layer
T1602980-97
Other than requiring an upsampling process to increase the size of the reference layer picture prior to
its use as a reference for the encoding process, the processing and syntax for a spatial scalability
picture is functionally identical to that for an SNR scalability picture.
Since there is very little syntactical distinction between pictures using SNR scalability and pictures
using spatial scalability, the pictures used for either purpose are called EI- and EP-pictures.
The picture in the base layer which is used for upward prediction in an EI- or EP-picture may be an I-
picture, a P-picture, or the P-part of a PB- or Improved PB-frame (but shall not be a B-picture or the
B-part of a PB- or Improved PB-frame).
Enhancement
EI EP EP EP
layer 1
Base
I P P P
layer
T1602990-97
In the case of multilayer scalability, the picture in a reference layer which is used for upward
prediction in an EI- or EP-picture may be a I-, P-, EI-, or EP-picture, or may be the P-part of a PB- or
Improved PB-frame in the base layer (but shall not be a B-picture or the B-part of a PB- or Improved
PB-frame).
As with the two-layer case, B-pictures may occur in any layer. However, any picture in an
enhancement layer which is temporally simultaneous with a B-picture in its reference layer must be a
B-picture or the B-picture part of an PB- or Improved PB-frame. This is to preserve the disposable
nature of B-pictures. Note, however, that B-pictures may occur in layers that have no corresponding
picture in lower layers. This allows an encoder to send enhancement video with a higher picture rate
than the lower layers.
The enhancement layer number and the reference layer number for each enhancement picture (B-, EI-
, or EP-) are indicated in the ELNUM and RLNUM fields, respectively, of the picture header (when
present). See the inference rules described in 5.1.4.4 for when these fields are not present. If a B-
picture appears in an enhancement layer in which temporally surrounding SNR or spatial scalability
pictures also appear, the Reference Layer Number (RLNUM) of the B-picture shall be the same as
the Enhancement Layer Number (ELNUM).
The picture height, width, and pixel aspect ratio of a B-picture shall always be equal to those of its
temporally subsequent reference layer picture.
5,6 7,8
Enhancement
B B
layer 2
Enhancement
EI B EP
layer 1
Base
layer I B P
T1603000-97
T1603010-97
The direct prediction mode is only available for B-pictures. It is a bidirectional prediction mode
similar to the bidirectional mode in Improved PB-frames mode (Annex M). The only difference is
that there is no restriction on which pixels can be backward predicted since the complete backward
prediction image is known at the decoder. Bidirectional mode uses separate motion vectors for
forward and backward prediction. In both direct and bidirectional mode, the prediction pixel values
are calculated by averaging the forward and backward prediction pixels. The average is calculated by
dividing the sum of the two predictions by two (division by truncation). In direct mode, when there
are four motion vectors in the reference macroblock, then all four motion vectors are used as in
Improved PB-frames mode (Annex M).
For B-pictures, forward prediction means prediction from a previous reference picture in the
reference layer. Backward prediction means prediction from a temporally subsequent reference
picture in the reference layer.
For EP-pictures, forward prediction means prediction from a previous EI- or EP-picture in the same
layer, while upward prediction is used to mean prediction from the temporally simultaneous
(possibly interpolated) reference picture in the reference layer. No motion vector is used for the
upward prediction (which is syntactically in the same place as backward prediction for B-pictures),
although one can be used for the forward prediction.
The macroblock syntax for EI-pictures is slightly different. As shown in Figure O.7, the MBTYPE
and CBPC are combined into an MCBPC field. No forward prediction is used, only upward
prediction from the temporally simultaneous reference picture in the reference layer. No motion
vector is used.
In B- and EP-pictures, motion vectors over picture boundaries may be used as described in D.1
(although extension of the motion vector range as described in D.2 is only active if the Unrestricted
Motion Vector mode is also in use).
c d
C D
T1603030-97
a b c
A B a=A
b = (3 * A + B + 2) / 4
d c = (A + 3 * B + 2) / 4
d = (3 * A + C + 2) / 4
e = (A + 3 * C + 2) / 4
C D
T1603040-97
Picture
boundary
a b B a = (3A + B + 2) / 4
A
b = (A + 3B + 2) / 4
T1603060-97
Picture
boundary
ANNEX P
Reference picture resampling
P.1 Introduction
This annex describes the use and syntax of a resampling process which can be applied to the previous
decoded reference picture in order to generate a "warped" picture for use in predicting the current
picture. This resampling syntax can specify the relationship of the current picture to a prior picture
having a different source format, and can also specify a "global motion" warping alteration of the
shape, size, and location of the prior picture with respect to the current picture. In particular, the
Reference Picture Resampling mode can be used to adaptively alter the resolution of pictures during
encoding. A fast algorithm is used to generate bilinear interpolation coefficients. The capability to
use this mode and the extent to which its features are supported is negotiated externally (for example,
Recommendation H.245). This mode may be used in restricted scenarios that may be defined during
capability negotiation (e.g. to support only factor-of-four picture resizing, to support only half-pixel
resolution picture warping, or to support arbitrary image size resizing and displacements).
NOTE – The default image transformation between pictures of differing resolutions is defined to maintain
the spatial alignment of the edges of the picture area and of the relative locations of the luminance and
chrominance samples. This may have implications on the design of any resampling operations used for
generating pictures of various resolutions at the encoder and for displaying pictures of various resolutions
after decoding (particularly regarding shifts in spatial location caused by phase shifts induced in resampling).
Also, since this mode can be used for dynamic adaptive picture resolution changes, operation with this mode
may benefit from external negotiation to display the decoded picture at a higher resolution than its encoded
picture size, in order to allow switching between encoded picture sizes without resizing the picture display.
If the Reference Picture Resampling bit in the PLUSPTYPE field is not set, but PLUSPTYPE is
present and the picture is an INTER, B, or EP-picture or an Improved PB-frame, and the picture size
differs from that of the temporally previous encoded picture, this condition invokes Reference
Picture Resampling with warping parameters (see P.2.2) set equal to zero, fill mode (see P.2.3) set to
1
clip, and displacement accuracy (see P.2.1) set to -pixel accuracy. This causes the resampling
16
process to easily act as a predictively-encoded change in picture resolution. In simple factor-of-four
v00 vH0
v0V
vHV
T1603070-97
The horizontal size H and vertical size V of the current picture, and the horizontal size H R and
vertical size V R of the reference picture are those indicated by the picture header, regardless of
r 0 = v 00
r x = v H 0 − v 00
r y = v 0V − v 00
r xy = v 00 − v H 0 − v 0V + v HV
Using this definition, the equation for bilinear interpolation is rewritten as:
⎛ x⎞ ⎛ y⎞ ⎛ x ⎞⎛ y ⎞
v ( x , y ) = r 0 + ⎜ ⎟ r x + ⎜ ⎟ r y + ⎜ ⎟ ⎜ ⎟ r xy
⎝ H⎠ ⎝V ⎠ ⎝ H ⎠⎝V ⎠
For warping purposes, it is assumed that the coordinates of the upper left corner of the picture area
are (x, y) = (0,0) and that each pixel has unit height and width, so that the centers of the pixels lie at
⎛ 1 1⎞
the points ( x , y ) = ⎜ i L + , j L + ⎟ for i L = 0, …, H – 1 and j L = 0, …, V – 1, where the L
⎝ 2 2⎠
subscript indicates that i L and j L pertain to the luminance field. (As the pixel aspect ratio is usually
constant or the conversion of the aspect ratio is performed by this resampling, there is no need to
consider the actual pixel aspect ratio for these purposes.) Using this convention, the x- and
y-displacements at the locations of interest in the luminance field of the reference picture are:
1 ⎡ 0 ⎛ 1⎞ x ⎛ 1⎞ y ⎛ 1⎞ ⎛ 1 ⎞ xy ⎤
v x (i L , j L ) = ⎢ HVrx + ⎜⎝ i L + 2 ⎟⎠Vrx + ⎜⎝ j L + 2 ⎟⎠ Hrx + ⎜⎝ i L + 2 ⎟⎠ ⎜⎝ j L + 2 ⎟⎠ rx ⎥
HV ⎣ ⎦
1 ⎡ 0 ⎛ 1⎞ x ⎛ 1⎞ y ⎛ 1⎞ ⎛ 1 ⎞ xy ⎤
v y (i L , j L ) = ⎢ HVry + ⎜⎝ i L + 2 ⎟⎠Vry + ⎜⎝ j L + 2 ⎟⎠ Hry + ⎜⎝ i L + 2 ⎟⎠ ⎜⎝ j L + 2 ⎟⎠ ry ⎥
HV ⎣ ⎦
Because all positions and phases must be computed relative to the center of the upper-left corner
⎛ 1 1⎞
pixel, which has the coordinate (x, y) = ⎜ , ⎟ , the quantities of primary interest are:
⎝ 2 2⎠
1 ⎛ 1⎞ 1
x R (i L , j L ) − = ⎜ i L + ⎟ + v x (i L , j L ) −
2 ⎝ 2⎠ 2
1 ⎡ 0 ⎛ 1⎞ x ⎛ 1⎞ y ⎛ 1⎞ ⎛ 1 ⎞ xy ⎤ 1
= ⎢ HVrx + ⎜⎝ i L + 2 ⎟⎠ ( HV + Vrx ) + ⎜⎝ j L + 2 ⎟⎠ Hrx + ⎜⎝ i L + 2 ⎟⎠ ⎜⎝ j L + 2 ⎟⎠ rx ⎥ − 2
HV ⎣ ⎦
1 ⎛ 1⎞ 1
y R (i L , j L ) − = ⎜ i L + ⎟ + v y (i L , j L ) −
2 ⎝ 2⎠ 2
1 ⎡ 0 ⎛ 1⎞ x ⎛ 1⎞ y ⎛ 1⎞ ⎛ 1 ⎞ xy ⎤ 1
= ⎢ HVry + ⎜⎝ j L + 2 ⎟⎠Vry + ⎜⎝ j L + 2 ⎟⎠ ( HV + Hry ) + ⎜⎝ i L + 2 ⎟⎠ ⎜⎝ j L + 2 ⎟⎠ ry ⎥ − 2
HV ⎣ ⎦
Once a location has been determined in the prior decoded reference picture using an approximation
of these equations, bilinear interpolation as specified later in this annex shall be used to generate a
value for the resampled pixel.
v 00 = v 00 00 00
warp + v size = v warp + ( 0,0)
v H 0 = v warp
H0 H0
+ v size H0
= v warp + ( H R − H ,0)
v 0V = v 0warp
V
+ v 0size
V
= v 0warp
V
+ (0,V R − V )
v HV = v warp
HV
+ v size
HV
= v warp
HV
+ ( H R − H ,V R − V )
P.2 Syntax
Whenever the reference picture resampling bit is set in the PLUSTPYPE field of the picture header,
the RPRP field of the picture header includes parameters which control the Reference Picture
Resampling process. This includes a two-bit Warping Displacement Accuracy (WDA) field, may
include eight warping parameters or one-bit warping parameter refinements, and includes a fill mode,
as described in this subclause.
When not a one-bit refinement, these parameters are sent in a similar manner as motion vector
differences in the Unrestricted Motion Vector mode when PLUSPTYPE is present, using Table D.3
with no restriction on the range of the parameters (i.e. a range of −4095 to +4095). As when
encoding motion vector difference pairs, an emulation prevention bit shall be added as needed after
each pair of warping parameters is sent, such that if the all-zero codeword (the value +1 in half-pixel
units) of Table D.3 is used for both warping parameters in the pair
w x0 = 2rx0 w 0y = 2ry0
(
w xx = 2 rxx − ( H R − H ) ) w yx = 2ryx
w xy = 2rxy (
w yy = 2 ryy − (V R − V ) )
w xy = 2rxy w yxy = 2ryxy
If the fill mode is clip, the coordinates of locations in the prior reference picture are independently
limited as in the Unrestricted Motion Vector mode so that pixel values outside of the prior reference
picture area are estimated by extrapolation from the values of pixels at the image border. If the fill
mode is black, luminance samples outside of the prior reference picture area are assigned a value of
Y = 16 and chrominance samples are assigned a value of C B = C R =128 . If the fill mode is gray,
luminance and chrominance values are assigned a value of Y = C B = C R =128 . If the fill mode is
color, then additional fields are sent to specify a fill color, as described in the next subclause.
P.2.4 Fill Color Specification (Y_FILL, CB_EPB, CB_FILL, CR_EPB, CR_FILL) (26 bits)
If the fill mode is color and the picture is not an EP-picture, then the fill-mode bits are followed in
the bitstream by three eight-bit integers, Y_fill, CB_ fill, and CR_ fill, which specify a fill color
precisely. Between these three eight-bit integers are two emulation prevention bits (CB_EPB and
CR_EPB) each of which is equal to 1. The format of this color specification, which is present only
when the fill mode is color, is shown in Figure P.2. Each eight-bit integer field is sent using its
Figure P.2/H.263 – Format of fill mode and fill color specification data
(
u xH 0 = 16 w x0 + w xx + 2(H R − H ) ) (
u yH 0 = 16 w 0y + w yx)
u x0V = 16(w x0 + w xy ) u 0yV = 16(w 0y + w yy + 2(V R − V ))
Next, H´ and V´, which denote the horizontal and vertical size of the virtual frame, are defined as the
smallest integers that satisfy the following condition:
H ′ ≥ H , V ′ ≥ V , H ′ = 2 m , V ′ = 2 n , m and n are positive integers
By applying bilinear extrapolation to the corner vectors of the luminance field, the integer parameters
u xLT , u yLT , u xRT , u yRT , u xLB , u yLB , u xRB , and u yRB which denote the x- and y-displacements of the
1
luminance field at the virtual points (x, y) = (0, 0), (H´, 0), (0, V´), and (H´, V´) in
-pixel accuracy
32
(the actual displacements are obtained by dividing these values by 32) are defined as:
u xLT = u x00 u yLT = u 00
y
( )
u xRT = ( H − H ′ )u x00 + H ′ u xH 0 / / H u yRT = (( H − H ′ )u 00
y + H′ uy ) / / H
H0
where "//" denotes integer division that rounds the quotient to the nearest integer, and rounds half
integer values away from 0.
0 1 2 3 4 5 0 1 2
x x
0 0
2 1
4 2
y y T1603080-97
Luminance Chrominance
Luminance sample
Chrominance sample
Picture boundary
( )
I R (i , j ) = Pi + ( SH ′ − 2i − 1)u xL ( j ) + (2i + 1)u xR ( j ) + 32 H ′ / P / / / (64 H ′ / P )
( )
J R (i , j ) = Pj + ( SH ′ − 2i − 1)u yL ( j ) + (2i + 1)u yR ( j ) + 32 H ′ / P / / / (64 H ′ / P)
i R (i , j ) = I R (i , j ) / / / P j R (i , j ) = J R (i , j ) / / / P
∅ x = I R (i , j ) − (I R (i , j ) / / / P )P ∅ y = J R (i , j ) − (J R (i , j ) / / / P )P
where
"///" integer division with truncation towards the negative infinity;
"/" integer division (in this case resulting in no loss of accuracy);
P accuracy of x- and y-displacements (P = 2 when WDA = "10" and P = 16 when WDA = "11"
or is absent, see P.2.1 for the definition of WDA);
⎛ I R (i , j ) 1 J R (i , j ) 1 ⎞
⎜ + , + ⎟ (x, y) location of the transformed position (both I R (i , j ) and J R (i , j )
⎝ P 2 P 2⎠
are integers);
⎛ 1 1⎞
⎜ i R (i , j ) + , j R (i , j ) + ⎟ (x, y) location of the sampling point near the transformed position
⎝ 2 2⎠
(both i R (i , j ) and j R (i , j ) are integers);
(
E P (i , j ) = ( P − ∅ y )(( P − ∅ x ) E R (i R , j R ) + ∅ x E R (i R +1, j R ))
+ ∅ y (( P − ∅ x ) E R (i R , j R + 1) + ∅ x E R (i R + 1, j R + 1)) + P 2 / 2 − 1 + RCRPR / P 2 )
where "/" denotes division by truncation. iR and jR are simplified notations for i R (i , j ) and j R (i , j ) ,
⎛ 1 1⎞
and E R (i R , j R ) denotes the sample value of the pixel located at ( x , y ) = ⎜ i R + , j R + ⎟ in the
⎝ 2 2⎠
reference picture after extrapolation using the proper fill mode if necessary. The value of parameter
RCRPR is defined as follows:
• For a B-picture (or the B-part of an Improved PB-frame) which has a P-picture as its
temporally subsequent anchor picture, RCRPR is equal to the rounding type (RTYPE) bit in
Since H, V, H´, and V´ are divisible by 4, the definition of u xRB can be rewritten as:
( ( ) (
u xRB = (VQ − VQ ′ ) ( H Q − HQ ′ )u x00 + H Q ′ u xH 0 + VQ ′ ( H Q − H Q ′ )u x0V + H Q ′ u xHV )) / / A
where H Q =H/4, VQ =V/4, HQ ′ =H´/4, VQ ′ =V´/4, A = HQVQ , and "//" denotes integer division that
rounds the quotient to the nearest integer, and rounds half integer values away from 0. Next,
parameters TT and TB are defined as:
TT = ( H Q − H Q ′ )u x00 + H Q ′ u xH 0
TB = ( H Q − H Q ′ )u x0V + H Q ′ u xHV
for simplicity of description. Using operator "///" which denotes integer division with truncation
towards the negative infinity, and operator "%" which is defined as a % b = a − (a /// b) b, the value
of u xRB can be obtained via the following pseudo-code:
q = (VQ − VQ´)*(TT /// A)+ VQ´*(TB /// A) + ((VQ − VQ´)*(TT % A)+ VQ´
*(TB % A)) /// A;
r = ((VQ − VQ´)*(TT % A)+ VQ´ *(TB % A)) % A;
if (q < 0)
u xRB = q+(r+(A-1)/2)/A;
else
u xRB = q+(r+A/2)/A;
A B a = (9A + 3B + 3C + D + 7 + RCRPR) / 16
b = (3A + 9B + C + 3D + 7 + RCRPR) / 16
c = (3A + B + 9C + 3D + 7 + RCRPR) / 16
a b d = (A + 3B + 3C + 9D + 7 + RCRPR) / 16
c d
C D
T1603090-97
a b c
A B a=A
b = (3A + B + 1 + RCRPR) / 4
c = (A + 3B + 1 + RCRPR) / 4
d d = (3A + C + 1 + RCRPR) / 4
e = (A + 3C + 1 + RCRPR) / 4
C D
Picture T1603100-97
boundary
a a = (A + B + C + D + 1 + RCRPR) / 4
C D
T1603110-97
ANNEX Q
Reduced-Resolution Update mode
Q.1 Introduction
This annex describes an optional Reduced-Resolution Update mode of this Recommendation. The
capability of this mode is signaled by external means (for example, Recommendation H.245). The
use of this mode is indicated in the PLUSPTYPE field of the picture header.
The Reduced-Resolution Update mode is expected to be used when encoding a highly active scene,
and provides the opportunity to increase the coding picture rate while maintaining sufficient
subjective quality. This mode allows the encoder to send update information for a picture that is
encoded at a reduced resolution, while preserving the detail in a higher resolution reference image to
create a final image at the higher resolution.
The syntax of the bitstream in this mode is identical to the syntax for coding without the mode, but
the semantics, or interpretation of the bitstream is somewhat different. In this mode, the portion of
the picture covered by a macroblock is twice as wide and twice as high. Thus, there is approximately
one-quarter the number of macroblocks as there would be without this mode. Motion vector data also
refers to blocks of twice the normal height and width, or 32 × 32 and 16 × 16 instead of the normal
16 × 16 and 8 × 8. On the other hand, the DCT or texture data should be thought of as describing
8 × 8 blocks on a reduced-resolution version of the picture. To produce the final picture, the texture
data is decoded at a reduced resolution and then up-sampled to the full resolution of the picture.
Scaling- Motion
up compensation
Pseudo- Reconstructed
vector vector
T1603120-97
16 * 16 prediction block
Q.2.4 Display
If both H and V are the same as HC and VC, the resulting picture in Q.2.2 is used for display purposes
as it is. Otherwise, this resulting picture is further cropped to the size H × V, and the cropped picture
is used for display purpose only.
192
= 32 * 6
176
144
160 referenced picture
= 32 * 5
T1603130-97
The extension of the referenced picture in QCIF is illustrated in Figure Q.2. The extended referenced
picture for luminance is given by the following formula:
RRRU(x, y) = R(x´, y´)
T1603140-97
For the macroblock using differential motion vectors in a B-picture, the motion vector for forward
and backward prediction are obtained independently. In the Reduced-Resolution Update mode, the
motion vector component MVc for a luminance block is reconstructed from MVD and MVD2-4 as
follows:
1) Pseudo-prediction vector component pseudo-PC is created from the prediction vector
component PC.
pseudo-PC = 0 if PC = 0
pseudo-PC = sign(PC) * (|PC| + 0.5) / 2.0 if PC ≠ 0
"/" indicates floating-point division (without loss of accuracy). The prediction vector
component PC is defined as the median value of the vector components MV1, MV2 and
MV3 as defined in 6.1.1 and F.2.
2) Pseudo-macroblock vector component pseudo-MVC is obtained by adding the motion vector
differences MVD (and MVD2-4) from Table 14 to the pseudo-PC.
In the default Reduced-Resolution Update mode, the value of pseudo-MVC is restricted to
the range [−16, 15.5]. Only one of the pair will yield a pseudo-MVC falling within the
permitted range. The procedure is performed in the similar way as defined in 6.1.1.
If the Unrestricted Motion Vector mode is also used with the Reduced-Resolution Update
mode, pseudo-MVC is obtained by adding the motion vector differences MVD (and MVD2-
4) from Table D.3.
If four motion vectors are present, the procedure is performed in the similar way as defined
in F.2.
2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2
2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2
1 1 1 1 2 2 2 2 2 2 2 2 1 1 1 1
1 1 1 1 2 2 2 2 2 2 2 2 1 1 1 1
1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
1 1 1 1 2 2 2 2 2 2 2 2 1 1 1 1
1 1 1 1 2 2 2 2 2 2 2 2 1 1 1 1
2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2
2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2
Figure Q.5/H.263 – Weighting values, H1, for prediction with motion vector of current 16 × 16
luminance blocks on top or bottom of current 16 × 16 luminance block
Block edge
A B a = (9A + 3B + 3C + D + 8) / 16
b = (3A + 9B + C + 3D + 8) / 16
c = (3A + B + 9C + 3D + 8) / 16
a b d = (A + 3B + 3C + 9D + 8) / 16
c d
C D
T1603160-97
Figure Q.8/H.263 – Creation of reconstructed prediction error for pixels inside block
Block
boundary
a b c
a=A
b = (3 * A + B + 2) / 4
a
c = (A + 3 * B + 2) / 4
d = (3 *A + C + 2) / 4
d e = (A + 3 * C + 2) / 4
b c
Block T1603170-97
boundary
block1
A
Block boundary
B
A B
Example for filtered pixels
on a vertical block edge
Example for filtered pixels
on a horizontal block edge
block1 block2
T1603180-97
One of the following conditions must be fulfilled in order to turn the filter on for a particular edge:
• block1 belongs to a coded macroblock (COD==0 || MB-type == INTRA); or
• block2 belongs to a coded macroblock (COD==0 || MB-type == INTRA).
A shall be replaced by A1 and B shall be replaced by B1. "/" indicates division by truncation.
A1 = (3 * A + B + 2) / 4
B1 = (A + 3 * B + 2) / 4
The order of edges where filtering is performed is identical to the description provided in J.3.
Q.7.2 Definition of the block boundary filter when Deblocking Filter mode is used
If the Deblocking Filter mode (see Annex J) is used with the Reduced-Resolution Update mode, the
filtering which is defined in Annex J with one modification is performed on the boundary pixels of
16 × 16 luminance and chrominance blocks, in place of the filtering described in Q.7.1. The one
modification of the filtering in Annex J is that the parameter STRENGTH is given the value of
positive infinity. This implies that the function UpDownRamp(x, STRENGTH) defined in J.3
becomes a linear function of x.
As a result, the procedure of the deblocking filter described in J.3 is redefined in the following
manner:
B1 = clip(B + d1)
C1 = clip(C − d1)
A1 = A − d2
D1 = D + d2
d1 = (A−4B+4C−D) / 8
d2 = clipd1((A − D) / 4, d1/2)
R.1 Introduction
This annex describes the optional Independent Segment Decoding mode of this Recommendation,
which allows a picture to be decoded without the presence of any data dependencies across slice
boundaries or across GOB boundaries having non-empty GOB headers. The use of this mode is
indicated in the PLUSPTYPE field of the picture header. The capability to use this optional mode is
negotiated by external means (for example, Recommendation H.245).
When the use of this mode is indicated, the video picture segment boundaries (as defined by the
boundaries of the slices or the upper boundaries of the GOBs for which GOB headers are sent, or the
boundaries of the picture, whichever bounds a region in the smallest way) are treated as picture
boundaries when decoding, including the treatment of motion vectors which cross those boundaries
(which result in boundary extrapolation when the Unrestricted Motion Vector mode, the Advanced
Prediction mode, the Deblocking Filter mode, or the Temporal, SNR, and Spatial Scalability mode
are in use, and which are prohibited when none of those optional modes are in use).
S.1 Introduction
This annex describes an optional Alternative INTER VLC mode of this Recommendation, which
improves the efficiency of inter-picture coding when significant changes are evident in the picture.
This efficiency improvement is obtained by allowing some VLC codes originally designed for
INTRA pictures to be used for some INTER picture coefficients and CBPY data as well. The use of
this mode is indicated in the PLUSPTYPE field of the picture header. The capability to use this
optional mode is negotiated by external means (for example, Recommendation H.245). The mode
contains two syntax alterations, one for the encoding of INTER coefficients and another for the
encoding of the INTER CBPY values.
ANNEX T
Modified Quantization mode
T.1 Introduction
This annex describes an optional Modified Quantization mode of this Recommendation, which
modifies quantizer operation. The use of this mode is indicated in the PLUSPTYPE field of the
picture header. The capability to use this optional mode is negotiated by external means
(for example, Recommendation H.245).
This mode includes four key features:
1) The bit-rate control ability for encoding is improved by altering the syntax for the DQUANT
field.
2) Chrominance fidelity is improved by specifying a smaller step size for chrominance than that
for luminance data.
3) The range of representable coefficient values is extended to allow the representation of any
possible true coefficient value to within the accuracy allowed by the quantization step size.
4) The range of quantized coefficient levels is restricted to those which can reasonably occur, to
improve the detectability of errors and minimize decoding complexity.
APPENDIX I
Error tracking
I.1 Introduction
This appendix describes a method to recover efficiently after transmission errors if erroneous MBs
are reported via a feedback channel to the encoder. The capability of sending and processing
feedback information is signalled via external means (for example, by Recommendation H.245).
Furthermore, format and content of the feedback message are defined externally (for example, by
Recommendation H.245).
For subsequent frames n, with nerr < n < nnext, the error may be estimated as:
N
d (i , mb, n)
E (mb, n) = ∑ E (i, n – 1)
256
i =1
where a uniformly distributed error in each macroblock is assumed after each iteration.
The estimated error E(mb, nnext – 1) is incorporated into the mode decision of the next frame. For
example, macroblock mb is coded in INTRA mode, if E(mb, nnext – 1) exceeds a threshold.
In practice, error tracking information will only be stored for the latest M frames. Then, if
nerr < nnext – M, no error-tracking information is available and the encoder must take special action.
For example, the next frame may be coded in INTRA mode. However, other procedures are possible
and may be more effective.
APPENDIX II
Recommended Optional Enhancement
II.1 Introduction
With the variety of optional modes available in this Recommendation (H.263 Version 2), it is crucial
that several preferred mode combinations for operation be defined, so that option-enhanced terminals
will have a high probability of connecting to each other using some syntax better than the "baseline."
This appendix contains a list of preferred mode combinations, structured into three "levels" of
support. Each level includes support for the levels below it, creating an "onion peel" structure of
capabilities. The primary objective of this appendix is to describe the order in which modes should
be supported in decoders, rather than to enforce a particular small set of mode combinations upon
encoders.
In determining which modes would be placed into this level structure, the primary criteria used were
performance-related: improvement in subjective quality, impact on delay, and impact on complexity
(this includes computational burden, data dependencies, and ease of implementation). However,
because this appendix seeks to address a wide variety of applications and transport layers, features
addressing error resilience and ease of packetization were also incorporated into the same level
structure. In keeping with these objectives, the levels of support described in this appendix are
intended to be universal, so that the means of transport will not be an issue and the need for
application-specific "profiles" will be minimized. This universality is seen as a key to promoting
good interoperability across different types of terminals and networks. To maintain universality,
several optional features have not been included in the level structure described herein. The absence
of these features is not meant to be a condemnation of their usage; instead, their absence reflects only
a collective opinion about the likelihood of widespread adoption of these features across a full
spectrum of terminals, networks, and applications.
原文:
[0019] Encoders can use this indicator to instruct decoders which pictures
resemble the current motion compensation reference picture so well that one
of them can be used as a spare reference picture if the actual reference picture
is lost during transmission. If a decoder lacks an actual reference picture but
can access a spare reference picture, preferably the decoder should not send
a request for an INTRA picture update. The indicator may be termed a spare
reference picture number since the indicator indicates to a decoder which
reference picture(s) resemble the default reference picture. This "spare"
reference picture may be used by a decoder to decode the current frame if the
default reference picture is lost for some reason.
[0020] The spare reference picture number may be in respect of the whole
picture or part of a picture. In the former case, typically the spare reference
picture number is included in a picture header. In the latter case the spare
reference picture number is included in the picture segment headers or
macroblock headers of the picture. In a preferred implementation of the
invention, the video signal is encoded according to the H.263 standard and the
indicator is included in the Supplemental Enhancement Information.
译文:
[0019] 编码器可用这种指示符来指示解码器哪些图像极为类似于当前运动补偿
参考图像,使得如果实际参考图像在传输过程中丢失,则这些图像之一可用作备
用参考图像。如果解码器缺少实际参考图像但是可以获得备用参考图像,则解码
器最好不要发出关于 INTRA 图像更新的请求。指示符可以称作备用参考图像编
号,因为指示符向解码器指出哪个(些)参考图像类似于缺省参考图像。如果缺省
参考图像由于某些原因而丢失,则这种“备用”参考图像可由解码器用来对当前
帧解码。
[0020] 备用参考图像编号可以与整个图像或图像的一部分有关。在前一种情况
下,备用参考图像编号通常包含在图像信头中。在后一种情况下,备用参考图像
编号包括在图像的图段信头或宏块信头中。在本发明的一种最佳实现中,视频信
号是根据 H.263 标准编码的,而指示符包含在附加增强信息中。
原文:
[0119] In a specific example where one spare reference picture is indicated
and the SRPN is represented with 10 bits, this message contains one data
byte, i.e., DSIZE is 3, CONT is 0, and EBIT is 6. It should be appreciated that
the values of DSIZE, CONT and EBIT will vary according to the number of
spare reference pictures indicated and the precision (number of bits) with
which the spare reference picture numbers are represented. If more than one
spare reference picture number is indicated, then preferably the message data
bytes contain the Spare Reference Picture Number(s) of the spare reference
pictures in preference order (the most preferred appearing first).
译文:
[0119] 在具体实例中,其中指明一个备用参考图像,并且用 10 位来表示 SRPN,
这个消息包含一个数据字节,也就是说,DSIZE 是 3 而 CONT 是 0,EBIT 是 6。
应当理解,DSIZE、CONT 和 EBIT 的值会根据所指明的备用参考图像的数目和
用以表示备用参考图像编号的精度(位数)而改变。如果指明一个以上的备用参考
图像编号,则最好是消息数据字节包括按照优先选择的次序(最佳选择排在第一)
排列的备用参考图像的备用参考图像编码。
二、证据 3
ITU-T H.263, 低比特率通信的视频编码(ITU-T H.263, Video coding for low
bit rate communication)
1、第 2 页 3.2
原文:
译文:
3.2 数字输出和输入
视频编码器提供自包容的数字比特流,它可以同其他多功能信号相结合(例如
ITU-T H.223 建议书中规定的)。视频译码器实施相反处理。
2、第 9 页 4.2
原文:
The source coder is shown in generalized form in Figure 3. The main elements
are prediction, block transformation and quantization.
译文:
4.2 视频源编码算法
源编码的一般形式如图 3 所示。主要的内容为预测、块变换和量化。
3、第 9 页最后一段
原文:
译文:
4.2.1 块组、切片、宏块和块
每个图像被分成块组(GOB,group of block)或切片。
4、第 10 页第一段后三行
原文:
Data for each GOB consists of a GOB header (may be empty) followed by data
for macroblocks. Data for GOBs is transmitted per GOB in increasing GOB
number.
译文:
5、第 11 页第 1 段
原文:
译文:
每个 GOB 被分成多个宏块。
6、第 11 页 4.2.2 第 1 段
原文:
The INTRA coding mode can be signaled at the picture level (INTRA for I‐pictures or
INTER for P‐pictures) or at the macroblock level in P‐pictures.
原文:
7、第 12 页第 3 段
原文:
译文:
解码器将接收每个宏块的一个矢量···
8、第 12 页最后一句
原文:
Within a macroblock the same quantizer is used for all coefficients except the
first one of intra blocks.
译文:
在宏块内,除了宏块内的第一个系数外,所有系数都使用相同的量化器。
9、第 13 页,5
原文:
5 Syntax and semantics
The video syntax is arranged in a hierarchical structure with four primary layers.
From top to bottom the layers are:
• Picture;
• Group of Blocks, or slice, or video picture segment;
• Macroblock;
• Block.
译文
视频句法排列成具有 4 个基本层的分级结构。从上到下,依次为:图像;块组;
宏块;块。
原文:
Data for each picture consists of a picture header followed by data for Group of
Blocks or for slices, eventually followed by an optional end-of-sequence code
and stuffing bits.
译文:
每个图像的数据包括一个图像信头以及跟随其后的像块组的数据或者片段的数
据。
11、第 96 页 N.2
原文:
译文:
N.2 视频源编码算法
原文:
N.4.1 前向信道
传递压缩视频信号的前向信道数据的句法仅在块组(GOB)层或切片层中更改。
GOB层的句法在图N.2中说明。添加TRI、TR、TRPI、TRP、BCI及BCM场到图9。