KBLaM icon indicating copy to clipboard operation
KBLaM copied to clipboard

Official Implementation of "KBLaM: Knowledge Base augmented Language Model"

Results 48 KBLaM issues
Sort by recently updated
recently updated
newest added

Training as proposed in the README or executing the 3rd cell of the eval_acc.ipynb gives the following error: File ~/Thomas/KBLaM/src/kblam/models/llama3_model.py:43 38 from transformers.modeling_outputs import ( 39 BaseModelOutputWithPast, 40 CausalLMOutputWithPast, 41...

Bumps the dependency-packages group with 2 updates in the / directory: [datasets](https://github.com/huggingface/datasets) and [transformers](https://github.com/huggingface/transformers). Updates `datasets` from 3.2.0 to 3.3.2 Release notes Sourced from datasets's releases. 3.3.2 Bug fixes Attempt...

dependencies
python

It seems I'm missing some modules when I run the experiements: ``` ~/kblam/experiments$ python train.py --verbose --dataset synthetic_data --N 120000 --B 20 --total_steps 601 --encoder_spec OAI --use_oai_embd --key_embd_src key --use_data_aug...

The assumption each line is it's own DataPoint is untrue, and breaks when trying to use the example datasets.

This classes are deprecated at transformers==4.46: ```python LlamaDynamicNTKScalingRotaryEmbedding LlamaLinearScalingRotaryEmbedding "`LlamaLinearScalingRotaryEmbedding` is deprecated an will be removed in v4.46. Please use " ``` However, KBLaM depends on transformers==4.48.0 and use that...

Hi @ti250 🤗 Niels here from the open-source team at Hugging Face. I discovered your work on Arxiv and was wondering whether you would like to submit it to [hf.co/papers](https://hf.co/papers)...

Hello, Thank you for sharing KBLaM's source code. I'd like to know more about the reproducibility of this work on different knowledge bases. According to what I've read, these are...

Trying Phi3 instead of Llama3 for the training with python train.py --dataset_dir ../datasets --train_dataset enron --N 120000 --B 20 --total_steps 601 --encoder_spec OAI --use_oai_embd --key_embd_src key --use_data_aug --hf_token ... --llm_type...

It seems that the configuration information in 'meta-llama/Llama-3.2-1B-Instruct/resolve/main/config.json' has changed since the code was used the last time. Running the training on the enron dataset gives: ` File "/home/fokus/Thomas/KBLaM/src/kblam/models/llama3_model.py", line...