Kuaipedia: a Large-scale Multi-modal Short-video Encyclopedia

10/28/2022
by   Haojie Pan, et al.
0

Online encyclopedias, such as Wikipedia, have been well-developed and researched in the last two decades. One can find any attributes or other information of a wiki item on a wiki page edited by a community of volunteers. However, the traditional text, images and tables can hardly express some aspects of an wiki item. For example, when we talk about “Shiba Inu”, one may care more about “How to feed it” or “How to train it not to protect its food”. Currently, short-video platforms have become a hallmark in the online world. Whether you're on TikTok, Instagram, Kuaishou, or YouTube Shorts, short-video apps have changed how we consume and create content today. Except for producing short videos for entertainment, we can find more and more authors sharing insightful knowledge widely across all walks of life. These short videos, which we call knowledge videos, can easily express any aspects (e.g. hair or how-to-feed) consumers want to know about an item (e.g. Shiba Inu), and they can be systematically analyzed and organized like an online encyclopedia. In this paper, we propose Kuaipedia, a large-scale multi-modal encyclopedia consisting of items, aspects, and short videos lined to them, which was extracted from billions of videos of Kuaishou (Kwai), a well-known short-video platform in China. We first collected items from multiple sources and mined user-centered aspects from millions of users' queries to build an item-aspect tree. Then we propose a new task called “multi-modal item-aspect linking” as an expansion of “entity linking” to link short videos into item-aspect pairs and build the whole short-video encyclopedia. Intrinsic evaluations show that our encyclopedia is of large scale and highly accurate. We also conduct sufficient extrinsic experiments to show how Kuaipedia can help fundamental applications such as entity typing and entity linking.

READ FULL TEXT

page 1

page 2

page 5

page 6

research
08/09/2023

AspectMMKG: A Multi-modal Knowledge Graph with Aspect-aware Entities

Multi-modal knowledge graphs (MMKGs) combine different modal data (e.g.,...
research
02/17/2023

Vision, Deduction and Alignment: An Empirical Study on Multi-modal Knowledge Graph Alignment

Entity alignment (EA) for knowledge graphs (KGs) plays a critical role i...
research
11/03/2021

Slapping Cats, Bopping Heads, and Oreo Shakes: Understanding Indicators of Virality in TikTok Short Videos

Short videos have become one of the leading media used by younger genera...
research
02/09/2021

Fashion Focus: Multi-modal Retrieval System for Video Commodity Localization in E-commerce

Nowadays, live-stream and short video shopping in E-commerce have grown ...
research
11/19/2022

A Unified Model for Video Understanding and Knowledge Embedding with Heterogeneous Knowledge Graph Dataset

Video understanding is an important task in short video business platfor...
research
06/04/2023

Predicting Information Pathways Across Online Communities

The problem of community-level information pathway prediction (CLIPP) ai...
research
10/27/2019

SoulMate: Short-text author linking through Multi-aspect temporal-textual embedding

Linking authors of short-text contents has important usages in many appl...

Please sign up or login with your details

Forgot password? Click here to reset