Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for Different Samplers and Schdulers? #45

Open
andyw-0612 opened this issue Sep 10, 2024 · 1 comment
Open

Support for Different Samplers and Schdulers? #45

andyw-0612 opened this issue Sep 10, 2024 · 1 comment

Comments

@andyw-0612
Copy link

First of all, thanks for an amazing project! This runs on average 30% faster than flux on ComfyUI. I was wondering if there's any planned support for different schedulers and samplers like how you can choose them in the sample and scheduler nodes in ComfyUI?

It would be great to test out different combination of schedulers as people have mentioned and shared that using different schedulers can improve/change generation style and quality.

@fblissjr
Copy link

Also love this project and would love to help contribute and collaborate where possible. Have been using flux on both comfy and the excellent https://github.com/aredden/flux-fp8-api project on my CUDA machine, and excited to see this in native MLX.

Generation speeds are a bit slower for me (at FP8 runtime quant, which is what I typically do in comfy).

Was looking into mlx compilation and trying a few quick things this morning on the code base, but seems like adding compilation will require some more time reading the code and documenting the data flow. Sounds like a job for Claude later on unless someone's already done something like this (I'm a data guy and it helps to see how it all flows).

Also highly interested in assisting with LoRA training if you get there. There's no real standards on this, but I like what kijai is doing in comfy (https://github.com/kijai/ComfyUI-FluxTrainer). Most are using ai-toolkit (github.com/ostris/ai-toolkit/), which I believe uses diffusers and kohya scripts (have tried to trace all the code here, and ostris did an awesome job opening the door on all this, but haven't been able to trace the flux-specific code end-to-end yet.

I've trained a few LoRAs just for tinkering purposes on single layers / full layers / several layers using a few different methods. If trained with runtime quant at FP8, the LoRA wont work unless it's also inferenced at runtime-quant FP8 (pre-quant doesn't work on this project or any others, which is good to see - consistency).

Looking forward to seeing this project grow, and using it as a reference educational tool for learning MLX deeper. Appreciate you open sourcing this, @filipstrand .

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants