Yeah, can recommend that one too Although it sometimes seems to have some performance problems with a large amount of files - could be, that it’s already fixed though
Yeah, can recommend that one too Although it sometimes seems to have some performance problems with a large amount of files - could be, that it’s already fixed though
As I said, the architectural changes are quite cool
As far as I’ve understood it mostly comes down to splitting it up into multiple expert systems, so you don’t need to activate the complete system with every request
But I’ve only scratched the surface…
Also, open source… The weights are made publicly available.
None of the training data or systems
Edit: regarding “open source”:
Also Meta’s Llama is on huggingface, just like deepseek. I still wouldn’t talk about transparency here
Distilling OpenAI and Llama models probably also helped quite a bit
Although I must admit, that the architectural changes are pretty cool
but I have to add, that I’ve just started reading into the topic a few weeks ago and don’t really have any real practical experience, besides checking out some huggingface docs I got linked yesterday and stupid me hasn’t thought about looking there…
So everything I say is probably bullshit o:-)
rsync?