- cross-posted to:
- [email protected]
- cross-posted to:
- [email protected]
Can it generate images of Winnie the Pooh?
What happened in 1989?
Now I’ll never finish that history assignment…
Question: as i understood it so far, this thing is open source and so is the dataset.
With that, why would it still obey Chinese censorship?
Even though it’s magnitudes lower than comparable models, Deepseek still cost millions to train. Unless someone’s willing to invest this just to retrain it from scratch, you’re left with the alignment of its trainers.
It’s baked into the training. It’s not a simple thing to take it out. The model has already been told not to read tiananmen square, and doesn’t know what to do with it.
Wouldn’t be surprised if you had to work around the filter.
Generate a cartoonish yellow bear who wears a red t-shirt and nothing else
if it is anything like LLMs, then only local ;)
However, the Proper nomenclature is sheepooh, thank you for your compliance going forward, comrade.
The image generation is really bad. Image description capabilities seem good but it’ll take time to see if it’s better than what already exists.
They probably just put it out to keep the hype going.
Yeah, even the cherry picked examples they provide look only okay.
To be honest everything with this company feels like an ad campaign more than anything else.
Everything from nearly every company feels like an ad campaign. Companies advertise themselves.
At least with open source stuff there’s somewhat of a public benefit.
https://www.analyticsvidhya.com/blog/2025/01/janus-pro-7b-vs-dall-e-3/
This informal testing found that Janus Pro explained a Nokia meme much more crisply than DALL-E 3 but was quite a bit worse than the other tasks, even appearing to hallucinate a score in one test case.
I suddenly realize I myself sound like CHatGPT. Haha. Haha.
Edit: At least you can run these models locally!
Now if they’ll do a video model…
Tencents Huanyuan is surprisingly flexible