@f Yup, a variant of it. I've been working on my own repo of it. He did implement pytorch mapping, so all 1.4/1.5 models work with it. I had to recode the tokenizer and UNET for 2.x and it works now! Unfortunately not for the 768 version, though. I guess it uses a different diffusion method?
But, thank you for the tips on CoreML! i will definitely look into it. I think I've maxed out the speed of TensorFlow on my Intel Mac.
@f I used Pytorch to create a dictionary variable of the models and then saved that has a .txt file. I didn't create the original mapping, but had to create this tool so I could implement different VAE's with the TensorFlow version.
I'm downloading the CoreML now and will run a test!
The only downside with divum's tensorflow implementation is that I don't know what scheduler it is. Would you happen to know?
@f Sorry to bother, but I keep getting this error:
"Fatal Python error: init_sys_streams: <stdin> is a directory, cannot continue
Python runtime state: core initialized
Current thread 0x00007ff85c75c680 (most recent call first):
<no Python frame>"
Any idea how I could troubleshoot it? Google has nothing...
@f Of course! This is the command I used for CoreML within my virtual environment:
python -m python_coreml_stable_diffusion.pipeline --prompt "a photo of an astronaut riding a horse on mars" -i models/coreml-stable-diffusion-v1-4_original_packages -o <creations/> --compute-unit CPU_ONLY --seed 93
I'm trying out Apple's CoreML with their base installation within a virtual environment.
@f Ah, that makes a lot of sense. For some reason I thought the CLI wanted me to point the directory within <>.
CoreML is certainly faster when it starts generating an image, but getting there takes waaaay longer than TensorFlow, even if I'm iterating another generation with the cached model.
@f I agree! On my Intel Mac I think it took about a minute to load the ORIGINAL models. But then the program just sits there trying to begin the first iteration. haha
@f Truth be told, Apple leaves the door open for software to fully utilize their hardware. Blender Cycles runs wicked fast on Silicon computers now because they adopted Metal. Iāve noticed the speed of Tensorflow is fantastic because it too has metal support (even for Intel Macās).
The real shame is PyTorchās lack of support for Intel Macās with metal. Their implementation only works with Silicon.