You are on page 1of 40

4/13/24, 10:59 PM Espacenet – search results

Espacenet
CN101930779B Video commenting method and video player

Applicants: HUAWEI DEVICE CO LTD

Inventors: JIANFENG XU,JUAN LIU

Classifications:
IPC H04N5/278;

Priorities: CN201010241023A 2010-07-29

Application: CN201010241023A·2010-07-29

Publication: CN101930779B·2012-02-29

Published as: CN101930779A; CN101930779B

Video commenting method and video player

Abstract

The embodiment of the invention discloses a video commenting method. The method comprises the following steps of:
receiving a comment adding command; calling a comment adding tool according to the comment adding command; adding
comment information for a video by using the comment adding tool, wherein the comment information comprises the motion
trail of an object in the video; a method for addingthe motion trail of a target comprises the following steps of: marking the
target in the video by using a graphic shape, acquiring the screen coordinate of the target at a moment when the video is
played and the screen coordinate of the target from a next moment when the video is played to next N moments, wherein N
is a natural number of more than 1; and acquiring the motion trail of the target which is formed by the movement of the
screen coordinate at different moments and consists of the graphic shape marking the target according to a video playing
sequence. The embodiment of the inventionalso provides a video player. In the embodiment of the invention, independence
of comment data and video data is realized and the motion trail of the target in the video is marked in the video without
compression.

https://worldwide.espacenet.com/patent/search/family/043369891/publication/CN101930779B?q=CN101930779B 1/1
Patent Translate
Powered by EPO and Google

Notice
This translation is machine-generated. It cannot be guaranteed that it is intelligible, accurate, complete, reliable or
fit for specific purposes. Critical decisions, such as commercially relevant or financial decisions, should not be
based on machine-translation output.

CLAIMS CN101930779B

1.
13 A video annotation method, characterized by comprising:
14 Receive a comment adding command;

15 Invoke an annotation adding tool according to the annotation adding command;

16 The annotation adding tool is used to add annotation information to the video, and the annotation

information includes the motion trajectory of the target in the video, wherein the method for adding the
motion trajectory of the target includes: using a graphic shape to mark the target in the video, obtaining the
screen coordinates of the target at the current moment of video playback, as well as the next moment of video
playback, until the next N moments of the screen coordinates of the target; N is a natural number greater than
1; according to the order of video playback, obtaining the motion trajectory of the target composed of the
graphic shape of the marked target formed by the movement of the screen coordinates at different moments.

2.
26 The method according to claim 1 is characterized in that the annotation adding tool includes: a graphics
drawing tool and/or a text editing tool.

3.
31 The method according to claim 1, wherein the annotation information further comprises:
32 Annotation content and the following information: the video data identification information identifier

corresponding to the annotation, the video playback time when the target appears in the video data, the screen
coordinates of the target in the video recorded by the annotation, the graphic shape used to mark the target,
and the appearance duration of the target recorded by the annotation; the annotation content includes text,
pictures and/or audio and video content.

13-04-2024 1
4.
40 The method according to claim 3, further comprising:
41 The annotation information added by the annotation adding tool and the default annotation information are

stored, and the video data identification information corresponding to the annotation information is
additionally recorded, wherein the video data identification information includes: the video file identifier,
name and storage location corresponding to the annotation information, the modification time and video file
size of the corresponding video file, and the video file fingerprint information;
46 The video file fingerprint information is a check digit obtained by operating on all or part of the video data of

the video file;


48 The default annotation information is information added by the system when any of the following annotation

information is not received using the annotation adding tool, specifically including: the video data
identification information identifier corresponding to the annotation, the video playback time when the target
appears in the video data, the screen coordinates of the target in the video recorded by the annotation, the
graphic shape used to mark the target, and the appearance duration of the target recorded by the annotation.

5.
56 The method according to claim 4, further comprising:
57 When playing the video again, the annotation information of the video is loaded, and the video playback time

when the target in the annotation information appears is parsed; when the video playback time is reached, the
annotation content, the screen coordinates of the target in the video recorded by the annotation, the graphic
shape used to mark the target, and the appearance duration of the target recorded by the annotation are
obtained; the graphic shape used to mark the target is displayed on a layer above the video display layer
according to the screen coordinates, and the annotation content is displayed on a layer above the video display
layer; and the display duration of the graphic shape and the annotation content is controlled by the appearance
duration of the target.

6.
68 The method according to claim 5, further comprising:
69 Automatically adjust the display position of the annotation contents of two overlapping annotations,

including: while ensuring that the annotation contents are in the screen area, the annotation contents of the
two annotations are respectively moved in the opposite direction relative to the origin of the other annotation
with their respective center points as origins, until the display positions of the two annotation contents do not
overlap each other.

7.
77 A video player, comprising a video playing unit for playing a video, characterized in that it also comprises:
78 A command receiving unit, used for receiving an annotation adding command;

79 an annotation adding tool providing unit, configured to call out an annotation adding tool according to the

13-04-2024 2
annotation adding command after the command receiving unit receives the annotation adding command;
81 An annotation unit is used to add annotation information to a video, wherein the annotation information

includes a motion trajectory of a target in the video, wherein a method for adding the motion trajectory of the
target includes: using a graphic shape to mark the target in the video, obtaining the screen coordinates of the
target at the current moment of video playback, and the screen coordinates of the target at the next moment of
video playback until the next N moments; N is a natural number greater than 1; according to the order of
video playback, obtaining the motion trajectory of the target formed by the movement of the screen
coordinates at different moments composed of the graphic shape marking the target.

8.
91 The video player according to claim 7, further comprising:
92 A storage unit, used to store the annotation information added by the annotation adding tool and the default

annotation information, and additionally record the video data identification information corresponding to
the annotation information, wherein the video data identification information includes: the video file
identifier, name and storage location corresponding to the annotation information, the modification time and
video file size of the corresponding video file, and the video file fingerprint information;
97 The video file fingerprint information is a check digit obtained by operating on all or part of the video data of

the video file;


99 The default annotation information is information added by the system when any of the following annotation

information is not received using the annotation adding tool, specifically including: the video data
identification information identifier corresponding to the annotation, the video playback time when the target
appears in the video data, the screen coordinates of the target in the video recorded by the annotation, the
graphic shape used to mark the target, and the appearance duration of the target recorded by the annotation.

9.
107 The video player according to claim 8, further comprising:
108 An annotation display unit is used to read and load the annotation information stored in the storage unit in

the video playback scene, parse the video playback time when the target in the annotation information
appears, and when the video playback time is reached, obtain the annotation content of the annotation, the
screen coordinates of the target in the video recorded by the annotation, the graphic shape used to mark the
target, and the appearance duration of the target recorded by the annotation, display the graphic shape used
to mark the target on a layer above the video display layer according to the screen coordinates, and display
the annotation content on a layer above the video display layer, and the appearance duration of the target
controls the display duration of the graphic shape and the annotation content.

10.
119 The video player according to claim 9, further comprising:
120 An adjustment unit is used to automatically adjust the display position of the annotation contents of two

overlapping annotations when the annotation display unit displays the annotation information, including:

13-04-2024 3
while ensuring that the annotation contents are within the screen area, the annotation contents of the two
annotations are respectively moved in the opposite direction relative to the origin of the other annotation
with their respective center points as origins until the display positions of the two annotation contents do not
overlap with each other.

13-04-2024 4
Patent Translate
Powered by EPO and Google

Notice
This translation is machine-generated. It cannot be guaranteed that it is intelligible, accurate, complete, reliable or
fit for specific purposes. Critical decisions, such as commercially relevant or financial decisions, should not be
based on machine-translation output.

DESCRIPTION CN101930779B
10 Video annotation method and video player

[0001]
14 Technical Field

[0002]
18 The present invention relates to the field of information processing, and in particular to a video annotation
method and a video player.

[0003]
23 Background technique

[0004]
27 At present, there are some software tools that can annotate text and image information on electronic devices,
such as Word office software and PDF readers that can annotate text documents. When reading a text
document, these tools can select a certain paragraph in the document and add comments and notes to the
paragraph.
31 However, in the video field, there is still a lack of methods for marking and annotating video content.

[0005]
35 The existing methods for marking and annotating video content generally include: using video synthesis
technology, through relatively professional-level video editing software, using video synthesis to compress the
annotations on the video content (text, graphics, etc.) and the video together to generate a new video data; or

13-04-2024 1
in the home video recorder scenario, simply associating the text content with the time coordinates of the video
to mark a certain time point in the video.

[0006]
43 The inventors found in the process of implementing the present invention that the prior art has at least the
following disadvantages:

[0007]
48 The use of video synthesis technology to superimpose the annotation content and the video together destroys
the original video data, is not conducive to the reuse of the video, and cannot extract the annotation content,
and cannot perform operations such as separate editing, time positioning and keyword retrieval on the
annotation content; its solution of simply associating the text content with the time coordinates of the video
cannot record events that occur continuously in the video, and cannot identify moving targets in the video.

[0008]
56 Therefore, existing video annotation technology cannot mark moving objects in the video without using video
synthesis technology.

[0009]
61 Summary of the invention

[0010]
65 The embodiment of the present invention provides a video annotation method and a video player, which can
identify the motion trajectory of an object in a video.

[0011]
70 The embodiment of the present invention is achieved through the following technical solutions:

[0012]
74 An embodiment of the present invention provides a video annotation method, comprising:

[0013]
78 Receive a comment adding command;

13-04-2024 2
[0014]
82 Invoke an annotation adding tool according to the annotation adding command;

[0015]
86 The annotation adding tool is used to add annotation information to the video, and the annotation
information includes the motion trajectory of the target in the video, wherein the method for adding the
motion trajectory of the target includes: using a graphic shape to mark the target in the video, obtaining the
screen coordinates of the target at the current moment of video playback, as well as the next moment of video
playback, until the next N moments of the screen coordinates of the target; N is a natural number greater than
1; according to the order of video playback, obtaining the motion trajectory of the target composed of the
graphic shape of the marked target formed by the movement of the screen coordinates at different moments.

[0016]
96 An embodiment of the present invention provides a video player, including a video playing unit for playing a
video, and further including:

[0017]
101 A command receiving unit, used for receiving an annotation adding command;

[0018]
105 an annotation adding tool providing unit, configured to call out an annotation adding tool according to the
annotation adding command after the command receiving unit receives the annotation adding command;

[0019]
110 An annotation unit is used to add annotation information to a video, wherein the annotation information
includes a motion trajectory of a target in the video, wherein a method for adding the motion trajectory of the
target includes: using a graphic shape to mark the target in the video, obtaining the screen coordinates of the
target at the current moment of video playback, and the screen coordinates of the target at the next moment
of video playback until the next N moments; N is a natural number greater than 1; according to the order of
video playback, obtaining the motion trajectory of the target formed by the movement of the screen
coordinates at different moments composed of the graphic shape marking the target.

[0020]
120 It can be seen from the technical solution provided by the above-mentioned embodiment of the present
invention that the embodiment of the present invention realizes the convenient and fast addition of

13-04-2024 3
annotations to videos. The annotation data and the video data are independent and are not compressed into
the video to identify the motion trajectory of the target in the video. Therefore, there is no need to modify the
original video content and only a small storage space is occupied.

[0021]
128 BRIEF DESCRIPTION OF THE DRAWINGS

[0022]
132 FIG1 is a flow chart of a video annotation method according to an embodiment of the present invention;

[0023]
136 FIG2 is a flow chart of a video annotation method according to another embodiment of the present invention;

[0024]
140 FIG3 is a schematic diagram of an example of a process of adding annotations according to the present
invention;

[0025]
145 FIG4 is a schematic diagram of an example of adding a motion trajectory according to the present invention;

[0026]
149 FIG5 is a schematic diagram of the display effect of annotations in video playback of the present invention;

[0027]
153 FIG6 is a schematic diagram of the structure of a video player according to an embodiment of the present
invention;

[0028]
158 FIG7 is a schematic diagram of the structure of a video player according to another embodiment of the
present invention;

[0029]
163 FIG8 is a schematic diagram of the structure of a video player according to another embodiment of the

13-04-2024 4
present invention.

[0030]
168 Detailed ways

[0031]
172 The technical solutions in the embodiments of the present invention will be clearly and completely described
below in conjunction with the drawings in the embodiments of the present invention. It can be understood
that the described embodiments are only part of the embodiments of the present invention, rather than all the
embodiments.
176 Based on the embodiments of the present invention, all other embodiments obtained by ordinary technicians

in this field without making any creative work shall fall within the scope of protection of the present
invention.

[0032]
182 An embodiment of the present invention provides a video annotation method, taking FIG. 1 as an example,
comprising the following steps:

[0033]
187 Step 10: Receive the annotation adding command;

[0034]
191 Step 11: calling an annotation adding tool according to the annotation adding command;

[0035]
195 Step 12: using the annotation adding tool to add annotation information to the video, the annotation
information including the motion trajectory of the target in the video;

[0036]
200 The method for adding the motion trajectory of the target includes: using a graphic shape to mark the target in
the video, obtaining the screen coordinates of the target at the current moment of video playback, as well as
the screen coordinates of the target at the next moment of video playback until the next N moments; N is a
natural number greater than 1; according to the order of video playback, obtaining the motion trajectory of
the target composed of the graphic shape of the marked target formed by the movement of the screen
coordinates at different moments.

13-04-2024 5
[0037]
209 The embodiment of the present invention realizes that annotations can be added to videos conveniently and
quickly. The annotation data is independent of the video data and is not compressed into the video to
identify the motion trajectory of the target in the video. Therefore, there is no need to modify the original
video content and only a small storage space is occupied.

[0038]
216 Another embodiment of the present invention provides a video annotation method, taking FIG. 2 as an
example, comprising the following steps:

[0039]
221 Step 20: Receive an annotation adding command;

[0040]
225 The comment adding command includes: the comment adding command input by the user through a
keyboard, a key input device such as a remote control, a touch screen, a pointing input device such as a
mouse, etc.

[0041]
231 The video playback may be paused after receiving the annotation adding command. The pause of the video
playback may be that the player automatically pauses the video playback after receiving the annotation adding
command, or the user may operate the video player to pause the video playback before adding the
annotation.
235 This embodiment of the present invention does not limit this.

[0042]
239 Step 21: calling an annotation adding tool according to the annotation adding command;

[0043]
243 The annotation adding tool includes: a graphics drawing tool and/or a text editing tool.
244 The function of the graphics drawing tool is similar to that of an electronic drawing board, and is used by the

user to mark the screen area where the target in the video is located. The user can select any of the multiple
graphic shapes provided to mark the target, and the user can also select the brush tool in the graphics drawing
tool to draw the graphic shape in the video area. The embodiment of the present invention does not make

13-04-2024 6
any specific limitation on this.
249 It is understandable that when the graphic drawing tool is not set in the annotation adding tool, a default

graphic shape, such as an ellipse, can be set to mark the target, and the graphic shape can be scaled.

[0044]
254 The text editing tool is used to add annotation content to the video in the annotation editing box. For
example, the annotation content may include: text, icons (such as emoticons), network links, music video
files, etc.
257 The graphic shape of the marked object and the annotation editing box described in the embodiment of the

present invention are displayed in a layer above the player video display layer.
259 It is understandable that if the text editing tool is not provided in the annotation adding tool, the annotation

content may be directly copied or inserted text.

[0045]
264 The annotation adding tool is in a hidden state during video playback. When receiving an annotation adding
command, the video player can call and display the annotation adding tool for the user to add rich
annotation content.

[0046]
270 Step 22: using the annotation adding tool to add annotation information to the video, the annotation
information including the motion trajectory of the target in the video;

[0047]
275 Among them, the method for adding the motion trajectory of the target includes: using a graphic shape to
mark the target in the video, obtaining the screen coordinates of the target at the current moment of video
playback, as well as the next moment of video playback, until the next N moments of the screen coordinates
of the target; N is a natural number greater than 1; according to the sequence of video playback, obtaining the
motion trajectory of the target composed of the graphic shape of the marked target formed by the movement
of the screen coordinates at different moments, thereby achieving the following effect: when the annotation is
replayed, the position of the annotated graphic shape is basically consistent with the actual target position in
the video, and moves accordingly with the movement of the target.

[0048]
286 The smaller the time difference between obtaining the two screen coordinates, the closer the marked graphic
shape is to the moving position of the actual target. The embodiment of the present invention does not
specifically limit the time difference for obtaining the two plane coordinates of the target.

13-04-2024 7
[0049]
292 Among them, there are many methods for triggering the marking of the motion trajectory of a target in a
video, which is not limited by the present invention. For example, during the video playback, a user may find
a target of interest, thereby triggering the marking of the target and marking the motion trajectory of the
target. Alternatively, before the video is played, certain conditions may be set in advance so that during the
video playback, the video player recognizes the corresponding target according to the conditions, thereby
triggering the marking of the target and marking the motion trajectory.
298 The video player can use existing target recognition technology to identify the target, and further obtain the

screen coordinates of the target at different times, thereby assisting in completing the movement trajectory of
the marked target.

[0050]
304 The annotation information also includes the annotation content and the following information: the video
data identification information mark corresponding to the annotation, the video playback time when the
target appears, the screen coordinates of the target in the video recorded by the annotation, the graphic shape
used to mark the target, and the appearance duration of the target recorded by the annotation.

[0051]
311 The annotation content may be in a rich text format. In the computer field, the rich text format refers to
content including text, pictures, audio and video, etc., wherein the text content may include information such
as font, size, color, and text layout.
314 The video data identification information identifier corresponding to the annotation is used to determine the

video data identification information corresponding to the annotation according to the identifier. The video
data identification information will be introduced in the following step 23.
317 The video playback time at which the target appears may be the video frame sequence number of the

annotated target in the video.


319 The screen coordinates of the target in the video recorded by the annotation may be the coordinates of the

center position of the target identified by applying the target recognition technology in the display screen of
the video player.
322 The appearance duration of the target recorded in the annotation is the display duration of the target marked

by the annotation during the video playback process, which is determined by the start time position and the
end time position of the target.

[0052]
328 In the embodiment of the present invention, if the user needs to accurately mark the starting time position of a
target in the video in the annotation, the user can use general video positioning methods, such as fast forward,
fast rewind, slow forward, slow rewind, jump by time, etc., to adjust and pause to the precise video screen,
and then add the annotation; the same method can also be used to mark the ending screen of a target in the

13-04-2024 8
video, and the time length from the target starting time position to the target ending time position will be used
as the playback duration of the target recorded in the annotation.

[0053]
337 Optionally, the embodiment of the present invention can set a default display time for the annotation when
replaying the video, automatically close the annotation after the display time is reached, or receive the display
time set by the user when the user adds the annotation.

[0054]
343 The embodiment of the present invention realizes adding rich text annotations to a video, and the rich text
annotations can record objects that continuously occur in the video and can identify moving objects in the
video, so as to achieve the purpose of accurately marking objects in the video.

[0055]
349 Step 23: storing the annotation information;

[0056]
353 Among them, the annotation information can be stored independently of the video file, and the stored
annotation information includes the annotation information added using the annotation adding tool and the
default information of the video player. The default information of the video player is the information added
by the system when the video player does not receive any of the annotation information added using the
annotation adding tool, specifically including: the video data identification information identifier
corresponding to the annotation, the video playback time when the target appears, the screen coordinates of
the target in the video recorded by the annotation, the graphic shape used to mark the target, and the
appearance duration of the target recorded by the annotation. The system is the operating system of the video
player.

[0057]
365 While storing the annotation information, the video data identification information corresponding to the
annotation information is additionally recorded, and the video data identification information includes: the
video file identification ID, name and storage location corresponding to the annotation information, the
modification time of the corresponding video file and the video file size.
369 Through the video file identification ID, name and storage location, the video file corresponding to the

annotation file can be found after the video player reads the annotation file; through the modification target
and video file size of the video file, it is possible to verify whether the annotation file is consistent with the
currently found video file before playing the video and displaying the annotation file, that is, the modification
time and video file size of the corresponding video file recorded in the annotation file correspond to the

13-04-2024 9
modification time and file size of the found video file. If they are consistent, the verification passes; if they are
inconsistent, it means that the verification fails, and the user can be prompted that the annotation file does not
match the video, thereby realizing the matching detection and prompt function of the video file.
377 In order to more accurately match the corresponding video file, the video data identification information may

also include video file fingerprint information, where the video file fingerprint information refers to a 128-bit
(or longer) checksum obtained by operating on all or part of the video data of the video file, such as a CRC
checksum.
381 The fingerprint information can be used to more accurately identify a video, regardless of whether the file

name or modification time of the video has changed.

[0058]
386 The embodiment of the present invention stores the annotations in rich text format independently of the
video file, so that the annotations of the video can be easily edited, deleted, merged, etc.

[0059]
391 The embodiment of the present invention can store the annotations of the video in a file after an annotation
adding process is completed, or after the video playback is completed, or when the player is closed, and the
embodiment of the present invention does not limit this.

[0060]
397 Optionally, the embodiment of the present invention can summarize all annotations of the video and compile
a list of annotations that can be displayed to the user, so that the user can view all annotations added to the
video and can also modify, delete, add, search, etc. the annotations.

[0061]
403 Optionally, the embodiment of the present invention can compile and store annotation index information for
the annotation information, so as to facilitate the user to perform keyword search, search and locate the video
content of interest.

[0062]
409 The embodiment of the present invention realizes that after adding annotations to a video, the annotations
can be conveniently edited, deleted, merged, etc., and the content of the annotations can be searched by
keywords.

[0063]
415 After an embodiment of the present invention adds an annotation to a video, the annotation content and part

13-04-2024 10
of the annotation information can be displayed when the video is played back. For example, when the video
is played again, the annotation of the video is loaded, and the video playback time when each target in the
annotation appears is parsed. When the time is reached, the video player obtains the rich text of the
annotation, the screen coordinates of the target in the video recorded by the annotation, the graphic shape
used to mark the target, the playback time of the target recorded by the annotation, and other information,
and displays the graphic shape used to mark the target on a layer above the video display layer according to
the screen coordinates, and displays the annotation content of the annotation on a layer above the video
display layer, and the playback time controls the display time of the graphic shape and the annotation
content.

[0064]
428 The embodiment of the present invention can automatically adjust the display position of two overlapping
annotations when displaying annotations during video playback, including: while ensuring that the
annotation content is within the screen area, the annotation contents of the two annotations are respectively
moved in the opposite direction relative to the origin of the other annotation with their respective center
points as origins until the display positions of the two annotation contents do not overlap with each other.

[0065]
436 The embodiment of the present invention realizes that annotations can be added to videos conveniently and
quickly. The annotation data is independent of the video data and is not compressed into the video to
identify the motion trajectory of the target in the video. Therefore, there is no need to modify the original
video content and only a small storage space is occupied.

[0066]
443 In addition, when the embodiment of the present invention displays annotations during video playback, a
video annotation can accurately locate a target occurring in the video, and at least locate the following
information of the target: the video playback time when the target appears, the screen coordinates of the
target in the video recorded by the annotation, the graphic shape used to mark the target, the playback
duration of the target recorded by the annotation, and the movement trajectory points of the target.

[0067]
451 To further understand the present invention, a method for adding video annotations, a method for storing
annotations, and a method for displaying annotations during video playback are introduced below with
specific examples.

[0068]
457 1. Example of adding annotations, see Figure 3

13-04-2024 11
[0069]
461 When a video player plays a video, after the user finds content of interest (called a target in the video), the user
activates the annotation function of the video player by pressing buttons, clicking on the touch screen,
operating a pointing input device, etc., and displays a graphic drawing tool, such as the graphic shape
selection box 31 in Figure 3, in which one of the graphic shapes is set to a selected state, and uses the currently
selected graphic shape to select an area 32 in the video screen. At this time, the video player used in this
example will automatically pause video playback, display a graphic shape of the marked target in the area
selected by the user, such as the default circle, and display an annotation edit box 33, and then wait to receive
rich text information entered in the annotation edit box, including text, icons (such as emoticons), network
links, music video files, etc.
470 The graphic shape and annotation editing box of the marked target are displayed in the graphic display layer

(located above the video display layer) of the video player.


472 After receiving the end instruction of the annotation added by the user, the video player using this example

will automatically continue to play the video.


474 The annotation end instruction described in this embodiment can be that the user clicks a close button

provided by the player, or presses a keyboard key indicating end.

[0070]
479 2. Example of adding motion trajectory, see Figure 4, taking the motion trajectory of the ball bouncing process
as an example

[0071]
484 After entering the annotation state, first use a circle to mark the sphere and the screen position of the sphere.
You can then choose to add a track tool, move the video playback time coordinates backward, and mark the
position of the sphere after the screen changes.
487 Repeat this process to mark multiple locations.

488 At this point, the video player forms a motion trajectory of the marked graphic shape with the position points

marked multiple times.


490 The position point of the sphere after the picture changes can be determined by the user based on the picture.

In the embodiment of the present invention, the video player can also determine the sphere after the picture
changes and obtain the position point of the sphere based on target recognition technology.

[0072]
496 The above-mentioned annotation of the motion trajectory can also be applied to the annotation of the
player's motion trajectory, which is explained below by taking the annotation of players in a football match as
an example.

13-04-2024 12
[0073]
502 When the user wants to identify the movement trajectory of "Player No. 3" throughout the game in the video
playback of the game, and mark the player's movement trajectory in the corresponding image, the user can
pre-select an image containing "Player No. 3", and use the face recognition of the video playback tool to
obtain his facial parameters, or analyze his body contour from the image, and then use these parameters to
identify or match the images during the video playback. In this way, the image containing "Player No. 3" can
be identified, and the corresponding annotation "Player No. 3" can be automatically marked using the video
annotation tool. In this way, the target that the user needs to mark can be marked in all video images.

[0074]
512 The face recognition, target recognition, and image matching technologies used in the above implementation
process are prior arts and will not be described in detail here.

[0075]
517 3. Annotation Storage Method Example

[0076]
521 In this example, after adding video annotations, the video player saves the video annotations separately in a
storage area called an annotation file. The annotation file stores annotations of one or more videos in a fixed
file format, specifically the following:

[0077]
527 1. Video data identification information corresponding to at least one annotation, including: video file
identification ID, name and storage location, and can store information such as video file modification time,
video file size for proofreading, and video file fingerprint information.

[0078]
533 2. At least one video annotation.
534 Each annotation contains the annotation content and at least one of the following information: the video data

identification information ID corresponding to the annotation, the video playback time (number of frames
played) when the target appears, the screen coordinates of the target in the video recorded by the annotation,
the graphic shape used to mark the target, the playback duration (number of frames) of the target recorded by
the annotation, and other information. Any of the above information can be added by the user through the
video adding tool, or it can be the default information of the video player. For example, the screen
coordinates of the target in the video recorded by the annotation can be specified by the video player as a
fixed area by default, or it can be set by the user when adding the annotation.

13-04-2024 13
542 The annotation information also includes a set of trajectory points of the target movement.

[0079]
546 IV. Example of how to display annotations during video playback

[0080]
550 The video player in this example can start displaying annotations during video playback in two ways.
551 One method is: specify a video file, and then the video player searches for the video annotation file in the

corresponding directory with a fixed naming rule. For example, it can be specified to load a video annotation
file in the same folder as the specified video file, with the same file name as the video file but a different suffix.

[0081]
557 Another method is: specify a video annotation file, the video player loads and displays the annotation list and
video data identification information, the user selects an annotation, and selects a video file based on the
video data identification information corresponding to the annotation. Specifically, when starting playback,
the video player reads the annotation file stored on the storage medium, first checks whether the annotation
file format is a legal annotation file. The annotation file is stored in a fixed format, so an annotation file that
does not conform to the format can be considered an illegal annotation file. Then, the corresponding video is
found through the video file identification ID, name, and storage location information recorded in the
annotation file, and the legality of the video is checked based on the file modification time and file size
information. If the check fails, the viewer will be prompted that the video and video notes may not match, or
the corresponding video file is found based on the video file fingerprint information, and the legality of the
video is checked based on the file modification time and file size information.
568 The video player considers that the video annotations can be played only after the video file and the video

annotation file are correctly loaded by the player at the same time.

[0082]
573 After the video player finishes loading the annotation file, it will parse each entry in the annotation file, obtain
each annotation information, and form an annotation list inside the player.

[0083]
578 In a video playback scenario, when the time when a certain target in the annotation occurs in the video is
reached, the video player obtains the annotation content, the screen coordinates of the target in the video
recorded by the annotation, the graphic shape used to mark the target, the playback time of the target
recorded by the annotation and other information, and displays the graphic shape used to mark the target on
a layer above the video display layer according to the screen coordinates, and displays the annotation content
on a layer above the video display layer, so as to achieve the effect of superimposing the display on the video,

13-04-2024 14
and the display time of the graphic shape and the annotation content is controlled by the playback time.
585 The display of the annotation file moves over time according to the movement track recorded by the

annotation.
587 The video player using this solution uses the display duration information of the annotations to determine

when to close the display of the annotations.


589 The annotation display effect is shown in FIG5 .

[0084]
593 The video player using this solution provides a conflict handling strategy for annotation display. When
displaying annotations during video playback, if the display positions of the annotation contents of two
annotations conflict, you can choose to automatically shift the display (or overlay display, which method can
be configured).
597 The method of automatic displacement display is that the video player using this solution obtains the display

area marked with two (or more) annotation contents that need to be displayed. When it is found that the
display area overlaps, the center point of each annotation is used as the origin, and then the display area is
moved in the opposite direction relative to the other origin until the display positions of the two annotation
contents do not overlap with each other. This process must be ensured within the screen display area.

[0085]
605 The video player using this solution can directly jump to the video playback at the time coordinate by
obtaining the time information of the annotation.

[0086]
610 The video annotation method described in the embodiment of the present invention can be widely used in
various fields. For example, it can be used in video websites on the Internet, so that when netizens watch the
videos provided by the video websites, they can also add annotations related to the video content to the
videos in real time; it can be used in the field of education, where teachers can add explanations to the video
teaching materials according to actual needs to enhance the effect of video teaching; it can be used in the field
of public security, where public security officers can repeatedly replay the videos when analyzing surveillance
videos and cases, and then mark and annotate the suspicious points or key points in the videos; it can be used
in the rapid production and sharing of home video recordings by ordinary home users, etc.

[0087]
621 Another embodiment of the present invention provides a video player, as shown in FIG6 , comprising: a
video playing unit 60 , a command receiving unit 61 , an annotation adding tool providing unit 62 , and an
annotation unit 63 ;

13-04-2024 15
[0088]
627 The video playback unit 60 is used to play the video; the video playback unit 60 can optionally pause the video
playback after the command receiving unit 61 receives the annotation adding command; the pause of the
video playback can be that the video playback unit 61 automatically pauses the video playback after receiving
the annotation adding command, or can be paused under the control of the user.
631 This embodiment of the present invention does not limit this.

[0089]
635 The command receiving unit 61 is used to receive an annotation adding command; the annotation adding
command includes: an annotation adding command input by the user by pressing a button, clicking a touch
screen, or operating an attached pointing device such as a mouse.

[0090]
641 The annotation adding tool providing unit 62 is used to call out an annotation adding tool according to the
annotation adding command after the command receiving unit 61 receives the annotation adding command;
the annotation adding tool includes: a graphic drawing tool and/or a text editing tool.
644 The function of the graphic drawing tool is similar to an electronic drawing board, and is used by the user to

mark the screen area where the video target is located.


646 The text editing tool is used to add annotation content to the video in the annotation editing box.

[0091]
650 The annotation unit 63 is used to add annotation information to the video, and the annotation information
includes the motion trajectory of the target in the video, wherein the method of adding the motion trajectory
of the target includes: using a graphic shape to mark the target in the video, obtaining the screen coordinates
of the target at the current moment of video playback, as well as the next moment of video playback, until the
next N moments of the screen coordinates of the target; N is a natural number greater than 1; in
chronological order, the graphic shape marking the target is moved according to the obtained multiple screen
coordinates to form the motion trajectory of the target.

[0092]
660 As shown in FIG. 7 , the video player further includes:

[0093]
664 The storage unit 64 is used to store the annotation information, the stored annotation information includes
the annotation information added using the annotation adding tool and the default annotation information,
and additionally records the video data identification information corresponding to the annotation
information, the video data identification information includes: the video file identifier, name and storage

13-04-2024 16
location corresponding to the annotation information, the modification time and video file size of the
corresponding video file, and video file fingerprint information; the video file fingerprint information is a
check digit obtained by operating on all or part of the video data of the video file; the default information
includes the following information that is used by default when any of the following information of the
annotation added using the annotation adding tool is not received, including: the video data identification
information identifier corresponding to the annotation, the video playback time when the target appears, the
screen coordinates of the target in the video recorded by the annotation, the graphic shape used to mark the
target, and the playback time of the target recorded by the annotation.

[0094]
679 The storage unit 64 can be any of the following: internal memory of the video player; additional memory such
as TF card, MiniSD card, etc.; or network storage with network devices such as WiFi, infrared, Bluetooth,
mobile phone network, etc.;

[0095]
685 Annotation display unit 65, an annotation display unit, is used to read and load the annotation information
stored in the storage unit 64 in the video playback scene, parse the playback time of the target in the
annotation information in the video, and when the time is reached, obtain the annotation content of the
annotation, the screen coordinates of the target in the video recorded by the annotation, the graphic shape
used to mark the target, and the playback time of the target recorded by the annotation, display the graphic
shape used to mark the target on the layer above the video display layer according to the screen coordinates,
and display the annotation content on the layer above the video display layer, and the playback time controls
the display time of the graphic shape and the annotation content.

[0096]
696 As shown in FIG8 , the video player may optionally include any one or more of the following units:

[0097]
700 The adjustment unit 66 is used to automatically adjust the display position of the annotation contents of two
overlapping annotations when the annotation display unit displays the annotation information, including:
while ensuring that the annotation contents are within the screen area, the annotation contents of the two
annotations are respectively moved in the opposite direction relative to the origin of the other annotation
with their respective center points as origins until the display positions of the two annotation contents do not
overlap with each other.

[0098]
709 The video player described in the embodiment of the present invention may be a video playback device in a

13-04-2024 17
computer, or a handheld playback device, or a portable playback device such as a mobile phone.

[0099]
714 The embodiment of the present invention realizes that annotations can be added to videos conveniently and
quickly. The annotation data is independent of the video data and is not compressed into the video to
identify the motion trajectory of the target in the video. Therefore, there is no need to modify the original
video content and only a small storage space is occupied.

[0100]
721 In addition, the embodiment of the present invention can accurately locate a target occurring in a video
through a video annotation, and at least locate the following information of the target: the video playback
time when the target appears, the screen coordinates of the target in the video recorded by the annotation, the
graphic shape of the marked target, the playback duration of the target, and the movement trajectory of the
target.

[0101]
729 You can easily edit, delete, merge, and perform other operations on annotations, and you can also perform
keyword searches on the content of annotations.

[0102]
734 In summary, the embodiments of the present invention realize that annotations can be added to videos
conveniently and quickly. The annotation data is independent of the video data and is not compressed into
the video to identify the motion trajectory of the target in the video. Therefore, there is no need to modify the
original video content and only a small storage space is occupied.

[0103]
741 In addition, the embodiment of the present invention can accurately locate a target occurring in a video
through a video annotation, as well as the following information of the target: the video playback time when
the target appears, the screen coordinates of the target in the video recorded by the annotation, the graphic
shape of the marked target, the target playback duration, and the movement trajectory of the target.

[0104]
748 It is convenient to store, edit, delete, merge and other operations on annotations, and it is also possible to
perform keyword searches on the contents of annotations.

13-04-2024 18
[0105]
753 Those skilled in the art will appreciate that all or part of the steps in the above-described embodiment method
may be implemented by instructing related hardware through a program, and the program may be stored in a
computer-readable storage medium, such as a read-only memory (ROM), a random access memory
(RAM), a disk, an optical disk, etc.

[0106]
760 The above description is only a preferred specific implementation manner of the present invention, but the
protection scope of the present invention is not limited thereto. Any changes or substitutions that can be
easily conceived by any technician familiar with the technical field within the technical scope disclosed by the
present invention should be covered within the protection scope of the present invention.
764 Therefore, the protection scope of the present invention should be based on the protection scope of the

claims.

13-04-2024 19

You might also like