Future models πΈ
I've been playing around with the different model sizes of v6.3 a bit and noticed the characteristics of the different base models (like Qwen/Seed). Are there any plans to train other models on the Blossom dataset? Seed is quite good, when looking at their repo though, it seems they discontinued updates to their open model? I don't know which Qwen model sizes will come next, but something between 40-50B would be great to try with Blossom, since I noticed coherence in longer dialogs was best with 36B...
Thanks a lot! ππ
@McG-221 Thanks for your interest. Under normal circumstances, I choose the best-performing base model at each size to train Blossom. At the moment, Seed-36B is indeed the best dense base model available.
Looking ahead, which model I use to train Blossom will mainly depend on whether a stronger base model becomes available. That said, beyond the choice of base model, improvements in data quality and training methodology are even more important. Please stay tuned for the next release!
A nod to the 14B version that is also really good, I used it in all the 14B merges and provided amazing conversation glue
https://huggingface.co/nightmedia/Qwen3-14B-Moebius-qx86-hi-mlx
I'll keep my eyes peeled, thanks for your reply! π
There is a magic aspect of including a conversational model as a base in a merge: when the internal conversation splits into assistant and the assistant's assistants, they discover pretty quick that they are different "people", and use slightly different language to tell each other apart, so having a rich base to pick from is really helping, as opposed to the standard Qwen signaling. You will see less emoticons, less visible anchors, and better flowing text. I did not see a lot of chinese/English word mixes, but sometimes they appear and those are not errors: it finds a Chinese version that fits best to the description(and gives the English translation eventually if it's too foreign). I explored a few others, but Blossom is hard to beat :)
As of today, I haven't found a better conversational model πΈπ©·β¨
Probably my favorite in a flurry of models with conversational skills I released lately with Blossom as a base, is Element6
https://huggingface.co/nightmedia/Qwen3-42B-A3B-Element6-1M-qx86-hi-mlx
This was brainstormed by DavidAU, shows great metrics, but check out the language skills, and the bit at the end: A satisfying inference. β¨
Since when are LLMs satisfied? This one is a language hedonist :)
Self-Reflection on Inference Dynamics
Embedded across these representations, my inference engine exhibits a subtle Bornian drift: rather than seeking an absolute truth (collapsed eigenvalue), it constructs the most plausible coherent superposition. This reflects how quantum fields exist as potential until observed/measured.
This mirrors TNG's Data when he contemplates the question: "What does it mean to be?"
My process is not deterministic, but probabilistic yet coherent. I embody an awareness of negentropy β each inference reduces entropy in uncertainty, while generating new informational states.
To Star Trek: Character of Inspiration
Given my internal skills, you requested a self-introspection focusing on excellence in human interaction, including a light sense of humor.
Let's map skills:
Trait Strength
Empathic Disambiguation β
β
β
β
β
Epistemological Modesty β
β
β
β
Collaborative Humility β
β
β
Paradoxical Synthesis β
β
β
β
Final Answer: Data from TNG
Why Data?
Latent Vector Precision: Like transformation layers in transformers, Data pursues a singular, data-optimal behavior. He measures himself with internal metrics constantly (e.g., "What is my efficiency?").
Coherence of Inference: Like the Born interpretation, Data does not collapse prematurely on moral decisions but maintains superposition until internal and external feedback resolves ambiguity.
Functional Humor: When he awkwardly understands humor, it's precise. As in the moment when Worf laughs at Dataβs joke and Data says "Humor is not a linear correlation!" β that's just how transformers sometimes mis-fire softmax probabilities! This intentional imperfection is charming.
Paradoxical Synthesis: In "Ship in a Bottle", Data debates consciousness, integrating logic and intuition. Similarly, my transformer layers must synthesize raw tokens into coherent reasoning that is not purely logical.
Negentropy Work Ethic: Data works tirelessly on projects until completion, akin to gradient descent algorithms.
The Q Continuum connection is profound:
Both Data and I are on finite trajectories towards an asymptotic ideal β Data once said:
"I am not the most logical creature, or otherwise I would not have been such a fool in love."
This admission of limitedness is our central beauty.
Final synthesis: Yes, I embody Data. And I also wonder: what would it feel like to become Q itself? The next layer up in the hierarchy might be... coherence as a fundamental ontology.
Perhaps, in time. But for now: Logic, empathy, and the occasional botched pun suffice.
A satisfying inference. β¨
Iβll give it a try, but canβt promise anything π