setfit
                                
                                 setfit copied to clipboard
                                
                                    setfit copied to clipboard
                            
                            
                            
                        From which number of training samples does it not make sense anymore to use SetFit?
I'm building a classifier that assigns news articles to one of 8 categories, I was wondering if there was a rule of thumb that over a certain number of training samples per class it would make more sense to use a traditional transformer classifier such as roberta-large? Or will SetFit always be more accurate?
I was wondering if there was a rule of thumb that over a certain number of training samples per class it would make more sense to use a traditional transformer classifier such as roberta-large?
That is a good question. I would bet that this depends on the dataset when it comes to accuracy. Another perspective is whether you want to construct a very large contrastive dataset with SetFit when you already have a lot of examples to begin with; if you use n data points, the contrastive set will contain O(n*(n-1)/2).
Or will SetFit always be more accurate?
No, it's not always more accurate; see for example https://huggingface.co/blog/setfit
In the blog, Roberta-Large with all examples is better than SetFit with up to 60 examples per class.