𝙒𝙝𝙤 𝙣𝙚𝙚𝙙𝙨 𝙪𝙥𝙨𝙘𝙖𝙡𝙚𝙧𝙨?
The TempestV0.1 Initiative is a powerhouse in image generation, leveraging an unparalleled dataset of over 6 million images. The collection's vast scale, with resolutions from 1400x2100 to 4800x7200, encompasses 200GB of high-quality
i just finished what i think was the last finetune.... I am not one to say this by my god this might be dangerous to open source! Hands down MJV6 level. this is not a joke or exaggeration i feel at this point.
The new proteus mobius model has:
1. The most diverse outputs
2. The strongest prompt control
3. The best aesthetic
Of any model I've used to date. I mean just look at the perspective on these!
I have joined
@Corcel_X
as Head of Research! I will be working full time as a Applied ML Researcher/Head of their diffusion Research department. working on open sourcing state of the art Diffusion based models.
Thrilled to unveil that
@DataPlusEngine
has come aboard the Corcel Research Labs team, spearheading our Image Generation research.
Our mission: to craft the premier open source image generation model. Stay tuned for groundbreaking updates in the weeks ahead! 😍
I implemented me own "prompt injection",
Now I can... prompt SPECIFIC BLOCKS of the UNet in different ways, And I implemented three different kinds of nodes for different workflow usages Each of these on the screenshot does the same thing, inject a specific prompt in OUTPUT0
I am happy to present and announce ! a fine-tune of my OpenDalleV1.1 using 220k captioned gptV dataset to align and then dpo tuned on 10k dalle3 captioned image pairs. results are another step above OpenDalle in terms of prompt following and style.
We reverse engineered the ELLA training for 1.5 and successfully made a finetune of it. We are working on adapting the script to work with SDXL. major disappointment in them for not releasing it. So when in doubt do it yourself!
This cycle has repeated every release.
Step 1. Angry Anticipation: "WHERE THE F*CK IS THE MODEL!"
Step 2. Release Rage: "THIS MODEL F*CKING SUCKS"
Step 3. Reversion Resignation "I'm going back to 1.4, it make t*ddies good"
Step 4. Furries F*ck is released "This model is so
The PixArt Sigma 900M really helps fix the issue of grainy faces that the 600M base had! still undertrained right now. going to talk and cover information about it and Prometheus with
@HelloCivitai
tomorrow during their podcast/interview with me!
The 900M Pixart Sigma Model should be done cooking by this upcoming Wednesday. shout out to for making his adaptation to the base model possible. (it produces noised images since its not trained)
Just achieved full rank finetuning of Stable Diffusion 3 Medium 2B on a SINGLE consumer GPU!
Only months ago, this was thought possible only on 80GB VRAM cards.
Now:
✅ Model: SD3 2B
✅ Hardware: 24GB VRAM RTX 4090
✅ Dataset: First layer of Proteus (prompt-alignment subset)
Found the issue that was causing the lack of detail. retrained the FLUX-syntheticanime lora overnight and its a lot better. going to redo megalith next. the quality is so so so so much better now.
Appalling that
@StabilityAI
so ruthlessly censored SD3 (for fear that someone might draw—gasp!—a nipple) that it nearly lost the ability to draw women entirely. Behold, "A woman lying in the grass:"
Fire the entire "AI Safety" team and replace them with an anatomy textbook.
sorry for the delay regarding the 900M Pixart Sigma model. Been extremely sick, never got hit with covid all the way up until now. went to the docs yesterday because it got so bad. they have me on quick special dial because i also have a autoimmune condition.
I trained a new VAE with 16x depth and 42 channels (kl-f16-d42). I am now training SD1.5 to work with it, which will double the output size of SD1.5 without much additional compute overhead.
Every time I train a new latent space, it always starts out inverted. It's so odd.
lady's and gents we are about to work on the models and bot full time with 3 employees under my company DataPusle AI. this is all thanks to investment by a undisclosed party (will be revealed in a bit)
Novel flux-dev conditioning method in ComfyUI:
Independently condition CLIP and T5-XXL with architecture-specific prompts to influence style and content.
Example:
T5-XXL (content): "a cat sitting on a bench drinking a latte in anime style."
CLIP (style): "cat, anime, anime
Oh, also, btw I got Flux training working
First Lora will be a synthetic anime lora because I absolutely love the style! Btw, I found a way to even get a 32 rank lora to train on 24vram. instead of just 16 rank. I'll make a pull request
Does anyone want to collaborate on making a FLUX lora? I have some very, very well curated datasets. I have datasets ranging in size from 400K to just 120. The Flux training has been very finicky however. Please let me know!
🚀 Double release day! 🎉
1️⃣ PrometheusV1:
• First full rank finetune of Playground v2.5
• Optimized for open-source accessibility
• Custom CLIP integration (use clip skip 2!)
2️⃣ ProteusV0.5 is out too!
Pushing boundaries in
#AIArt
. Try them
We gave the
@FAL
early access to the upcoming Mobius model and its only been up on for 3 hours. its already the best stable diffusion based image model in the world based on human preference data
@DataPlusEngine
It's insane, my dude. Clear winner most of the times. If this thing is open sourced - you wll clearly be the best open source model provider on the market.
Our upcoming paper outlines and enables making entirely new base diffusion models without the need to extensively pretrain a new model from scratch. We can in a controlled way, break all the quality and composition associations without damaging the baseline understanding from the
my friend Mobi who helped created the prompt injection node just made a clip switch node for SDXL clip to SD3! no weird requirements.txt, super light, extra simple, but make sure to read the README of the repo,
Pro tip for using the Mobius model on
@huggingface
:
For realistic photos, set the CFG to 3.5.
For artistic images, crank it up to 7.0.
Experiment with these settings to get the best results!
cfg 3.5 image below!
working on getting a implementation of
#SD3
working with a sponsor. We are fairly confident that we can get one trained. wish us luck! more announcements coming soon.
samples from the current continued training run are looking great. 1B params, it should run on a potato. also MJ data was not used period. throw some prompts at me!
I put a lot of effort into making the playground 2.5 arch/model backwards compatible with as many existing SDXL tools, loras, samplers, and etc. as possible! that's one of the key aspects of the model!
Big News! We're so excited to unveil Corcel Research Labs' two new image models: Open Vision & Mobius! ✨
Both fully open sourced under the apache-2.0 licence!
Check them out here:
💖Open Vision: Finetuned from S19's MJ
playing around with a custom sampler that has very fine control and i am loving it! choose the noise map and everything custom selection of dpm or eular per step
two phase sampling
etc
ProteusV0.4: The Style Update
This update enhances stylistic capabilities, similar to Midjourney's approach, rather than advancing prompt comprehension. Methods used do not infringe on any copyrighted material.
and here is what i assume people actually want,
PrometheusV2-aesthetic. I would recommend using the refiner with it. also it has a trained trigger word of "aesthetic".
@qtnx_
i am rather agnostic regarding timelines and tend to lean on the quote "When a distinguished but elderly scientist states that something is possible, he is almost certainly right. When he states that something is impossible, he is very probably wrong"
Exciting News! OpenDalle v1.1 is here!
We've revved up the realism & style, staying true to your prompts. Think of it as a step above SDXL, yet playfully chasing DALLE-3's tail.
Get it now 👉
🚀 Exciting news from Corcel! We're previewing our upcoming image models and need your feedback to fine-tune it with DPO. Your input will help make it better for everyone. Check it out now and let us know what you think:
#AI
#MachineLearning
#Feedback
from testing aswell shift completely fixes the comp issues. from 6 to 10. y'all need to chill out its been out for less than a day lmao. its like driving a new car you have to give it some time to get used to the new system. chill yourself out yall lolol
𝗣𝗿𝗼𝘁𝗲𝘂𝘀-𝗥𝘂𝗻𝗗𝗶𝗳𝗳𝘂𝘀𝗶𝗼𝗻'𝘀 "𝗦𝘁𝘆𝗹𝗲 𝗨𝗻𝗹𝗼𝗰𝗸𝗶𝗻𝗴" 𝗶𝗻 𝗮𝗰𝘁𝗶𝗼𝗻:
We applied our innovative CLIP retraining approach to Animagine XL V3, an anime-only model, with remarkable results!
Without any retraining of the model itself, our novel CLIP
Introducing Proteus v0.4 + RunDiffusion Photo 'Style Unlocking' 🚀🎨: A game-changer in AI art generation. Our retrained CLIP model breaks through stylistic boundaries, unlocking endless creative possibilities. This has the potential to unlock numerous SDXL models!
#AIart
#ai
trained a Anime Variant along with the 900M base. So there will be two models 900M-base and a 900M-anime!
prompts:
left: a woman smiling
right: a epic swords man