Skip to content

spaghettiSystems/tiny_diffusion

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 

Repository files navigation

# SMALL TRANSFORMERS, BIG RESULTS: EFFICIENT DIFFUSION WITH PARAMETER SHARING

This repository contains the code and pre-trained models for the paper "Small Transformers, Big Results: Efficient Diffusion with Parameter Sharing," published as a Tiny Paper at ICLR 2024

## Introduction

The interplay between model depth, computational complexity, and parameter count remains an intricate aspect of neural network design. We propose a novel block sharing mechanism for denoising diffusion generative models, enabling us to maintain or even improve model quality while reducing parameter count. Our approach leverages the architectural homogeneity of Vision Transformers and demonstrates enhanced performance with less computational overhead on various datasets.

## Installation

Git clone the repository. The code is essentially a fork of k-diffusion, and the instructions from there are applicable here. https://github.com/crowsonkb/k-diffusion

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages