Hands-On-Large-Language-Models icon indicating copy to clipboard operation
Hands-On-Large-Language-Models copied to clipboard

Official code repo for the O'Reilly Book - "Hands-On Large Language Models"

Results 22 Hands-On-Large-Language-Models issues
Sort by recently updated
recently updated
newest added

So I tried links from both the book and the README.md of repository and they return 404 Error. are colabs still accessible?

I encountered a TypeError when indexing the dataset object using a NumPy integer. The error occurred while generating document visualizations with BERTopic using the following code snippet: `fig = topic_model.visualize_documents(...

Thank you for writing this excellent and extremely conscience yet informative book. I thoroughly enjoyed it. If at all helpful, I wondered if there might be a possible typo on...

Hi There, In Chapter 10 - `Creating Text Embedding Models` from Part III, in the section of `Fine-Tuning an Embedding Model` on Page 313, I think there is a typo...

I tried using https://hf-mirror.com, but it was also very unstable. It often stopped downloading in the middle of the process.

When using `SentenceTransformer.encode` with a Column object from a Hugging Face datasets dataset (e.g., `data["train"]["text"]`), a TypeError occurs because the method attempts to index the Column object with` numpy.int64` indices,...

If anyone runs this SentenceTransformerTrainer in Chapter 10 but stuck, you can add `report_to=['none']` ``` from sentence_transformers.training_args import SentenceTransformerTrainingArguments # Define the training arguments args = SentenceTransformerTrainingArguments( output_dir="base_embedding_model", num_train_epochs=1, #...

Hi, I'm running the code from the final chapter of the Hands-On LLM book, specifically this snippet that loads the ultrachat_200k dataset and formats it using the TinyLlama chat template:...

Running this part of the code in chapter 12, I get the following error. Although I have the latest version of bitsandbytes. `from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig model_name =...

Hello, From the very beginning of chapter 6, when trying to running the jupyter notebook locally with my 8GB VRAM gpu card: # Load model and tokenizer model = AutoModelForCausalLM.from_pretrained(...