Full-text search
+ 1,000 results
hustvl / lightningdit-xl-imagenet256-64ep
README.md
model
1 matches

changwh5 / Stylegan2-ada
README.md
model
1 matches
BooBooWu / Vec2Face
README.md
model
1 matches
tags:
unconditional-image-generation, arxiv:2409.02979, license:mit, region:us
6
7
8
9
10
# Vec2Face Model Card
<div align="center">
[**Project Page**](https://haiyuwu.github.io/vec2face.github.io/) **|** [**Paper**](https://arxiv.org/abs/2409.02979) **|** [**Code**](https://github.com/HaiyuWu/vec2face) **|** [🤗 **Gradio demo**](https://huggingface.co/spaces/BooBooWu/Vec2Face)

mit-han-lab / hart-0.7b-1024px
README.md
model
1 matches
tags:
unconditional-image-generation, arxiv:2410.10812, license:mit, region:us
5
6
7
8
9
# HART: Efficient Visual Generation with Hybrid Autoregressive Transformer
## Abstract
We introduce Hybrid Autoregressive Transformer (HART), an autoregressive (AR) visual generation model capable of directly generating 1024x1024 images, rivaling diffusion models in image generation quality. Existing AR models face limitations due to the poor image reconstruction quality of their discrete tokenizers and the prohibitive training costs associated with generating 1024px images. To address these challenges, we present the hybrid tokenizer, which decomposes the continuous latents from the autoencoder into two components: discrete tokens representing the big picture and continuous tokens representing the residual components that cannot be represented by the discrete tokens. The discrete component is modeled by a scalable-resolution discrete AR model, while the continuous component is learned with a lightweight residual diffusion module with only 37M parameters. Compared with the discrete-only VAR tokenizer, our hybrid approach improves reconstruction FID from 2.11 to 0.30 on MJHQ-30K, leading to a 31% generation FID improvement from 7.85 to 5.38. HART also outperforms state-of-the-art diffusion models in both FID and CLIP score, with 4.5-7.7x higher throughput and 6.9-13.4x lower MACs.

Gusanagy / UDBE-Unsupervised-Diffusion-based-Brightness-Enhancement-in-Underwater-Images
model
1 matches

brownvc / R3GAN-ImgNet-64x64
README.md
model
1 matches
tags:
unconditional-image-generation, dataset:ILSVRC/imagenet-1k, arxiv:2501.05441, region:us
7
8
9
10
11
# Model Card for ImageNet 64x64 R3GAN Model
This model card provides details about the R3GAN model trained on the ImageNet dataset found in the NeurIPS 2024 paper R3GAN: https://arxiv.org/abs/2501.05441
## Model Details

brownvc / R3GAN-FFHQ-256x256
README.md
model
1 matches

brownvc / R3GAN-FFHQ-64x64
README.md
model
1 matches

brownvc / R3GAN-CIFAR10
README.md
model
1 matches
tags:
unconditional-image-generation, dataset:uoft-cs/cifar10, arxiv:2501.05441, region:us
7
8
9
10
11
# Model Card for FFHQ 64x64 R3GAN Model
This model card provides details about the R3GAN model trained on the CIFAR10 dataset found in the NeurIPS 2024 paper: https://arxiv.org/abs/2501.05441
## Model Details

brownvc / R3GAN-ImgNet-32x32
README.md
model
1 matches
tags:
unconditional-image-generation, dataset:ILSVRC/imagenet-1k, arxiv:2501.05441, region:us
7
8
9
10
11
# Model Card for ImageNet 32x32 R3GAN Model
This model card provides details about the R3GAN model trained on the ImageNet dataset found in the NeurIPS 2024 paper: https://arxiv.org/abs/2501.05441
## Model Details
Epiphqny / PAR
README.md
model
1 matches
tags:
unconditional-image-generation, arxiv:2412.15119, region:us
5
6
7
8
9
# Parallelized Autoregressive Visual Generation
Autoregressive models have emerged as a powerful approach for visual generation but suffer from slow inference speed due to their sequential token-by-token prediction process. In this paper, we propose a simple yet effective approach for parallelized autoregressive visual generation that improves generation efficiency while preserving the advantages of autoregressive modeling. Our key insight is that parallel generation depends on visual token dependencies—tokens with weak dependencies can be generated in parallel, while strongly dependent adjacent tokens are difficult to generate together, as their independent sampling may lead to inconsistencies. Based on this observation, we develop a parallel generation strategy that generates distant tokens with weak dependencies in parallel while maintaining sequential generation for strongly dependent local tokens. Our approach can be seamlessly integrated into standard autoregressive models without modifying the architecture or tokenizer. Experiments on ImageNet and UCF-101 demonstrate that our method achieves a 3.6× speedup with comparable quality and up to 9.5× speedup with minimal quality degradation across both image and video generation tasks.
Abhiroop174 / galaxy_gen
README.md
model
1 matches

skytnt / fbanime-gan
README.md
model
1 matches

cmudrc / 2d-lattice-decoder
model
1 matches

utsavnandi / fashion-mnist-ddpm-32px-5000_steps
README.md
model
1 matches

changwh5 / BigBiGAN-MNIST-150epoch
README.md
model
1 matches
tags:
code, unconditional-image-generation, zh, dataset:mnist, region:us
11
12
13
14
15
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
This modelcard aims to be a base template for new models. It has been generated using (https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).