LLMs-from-scratch
                                
                                
                                
                                    LLMs-from-scratch copied to clipboard
                            
                            
                            
                        Qwen3 0.6b - context length
Could you provide more details about how you determined the context length?
I found this information:
- The 0.6b model seems to support only 32k (
32,768) tokens https://qwenlm.github.io/blog/qwen3/#introduction https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/README.md