Skip to content

VQGAN+CLIP with some additional tuning. For notebooks and the command line.

License

Notifications You must be signed in to change notification settings

dribnet/clipit_old

 
 

Repository files navigation

clipit

Yet Another VQGAN-CLIP Codebase

Alt text

This started as a fork of @nerdyrodent's VQGAN-CLIP code which was based on the notebooks of @RiversWithWings and @advadnoun. But it quickly morphed into a version of the code that had been tuned up with slightly different behavior and features. It also runs either at the command line or in a notebook or (soon) in batch mode.

Basically this is a verison of the notebook with opinionated defaults and slighly different internals. You are welcome to use it if you'd like.

For now, checkout THE DEMO NOTEBOOKS - especially the super simple "Start Here" colab.

Here's the curent default notebook (but warning - this will soon be obsolete): Colab

Citations

@misc{unpublished2021clip,
    title  = {CLIP: Connecting Text and Images},
    author = {Alec Radford, Ilya Sutskever, Jong Wook Kim, Gretchen Krueger, Sandhini Agarwal},
    year   = {2021}
}
@misc{esser2020taming,
      title={Taming Transformers for High-Resolution Image Synthesis}, 
      author={Patrick Esser and Robin Rombach and Björn Ommer},
      year={2020},
      eprint={2012.09841},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Katherine Crowson - https://github.com/crowsonkb Adverb https://twitter.com/advadnoun

About

VQGAN+CLIP with some additional tuning. For notebooks and the command line.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 92.2%
  • Shell 7.8%