Variable Length Embeddings

05/17/2023
by   Johnathan Chiu, et al.
0

In this work, we introduce a novel deep learning architecture, Variable Length Embeddings (VLEs), an autoregressive model that can produce a latent representation composed of an arbitrary number of tokens. As a proof of concept, we demonstrate the capabilities of VLEs on tasks that involve reconstruction and image decomposition. We evaluate our experiments on a mix of the iNaturalist and ImageNet datasets and find that VLEs achieve comparable reconstruction results to a state of the art VAE, using less than a tenth of the parameters.

READ FULL TEXT

page 5

page 9

page 10

page 11

page 12

research
09/12/2023

A Complete Proof of an Important Theorem for Variable-to-Variable Length Codes

Variable-to-variable length (VV) codes are a class of lossless source co...
research
09/05/2019

FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow

Most sequence-to-sequence (seq2seq) models are autoregressive; they gene...
research
03/28/2023

SnakeVoxFormer: Transformer-based Single Image Voxel Reconstruction with Run Length Encoding

Deep learning-based 3D object reconstruction has achieved unprecedented ...
research
10/05/2022

Phenaki: Variable Length Video Generation From Open Domain Textual Description

We present Phenaki, a model capable of realistic video synthesis, given ...
research
11/18/2022

CNeRV: Content-adaptive Neural Representation for Visual Data

Compression and reconstruction of visual data have been widely studied i...
research
07/13/2023

Domain-Agnostic Tuning-Encoder for Fast Personalization of Text-To-Image Models

Text-to-image (T2I) personalization allows users to guide the creative i...
research
08/31/2020

Matrix Profile Goes MAD: Variable-Length Motif And Discord Discovery in Data Series

In the last fifteen years, data series motif and discord discovery have ...

Please sign up or login with your details

Forgot password? Click here to reset