You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The guy converted the format of the 3 smallest models (3b,7b,10b) to HF transformers. Given the severe lack in non english output a good translation model would be a gift.
I just tried the CPU demo of the 3B, it produced quite good output, if that gets better with 7B+ it would be a real solution for a huge amount of people.
It could be added as a 2nd stage into llama.cpp
Though the architecture is "T5ForConditionalGeneration" which isn't supported.
So far there was no urgent reason to add those T5 models, they did not stick out as special but the idea to output text in every single language worldwide .. that would be remarkable
malipetek, flatsiedatsie, SachiaLanlus, vladfaust, corysus and 4 more