Muhammad Abdullah

Results 10 comments of Muhammad Abdullah

PeftModelForCausalLM( (base_model): LoraModel( (model): MistralForCausalLM( (model): MistralModel( (embed_tokens): Embedding(32000, 4096, padding_idx=2) (layers): ModuleList( (0-31): 32 x MistralDecoderLayer( (self_attn): MistralAttention( (q_proj): Linear4bit(in_features=4096, out_features=4096, bias=False) (k_proj): lora.Linear4bit( (base_layer): Linear4bit(in_features=4096, out_features=1024, bias=False) (lora_dropout):...

I have tested your updated code https://github.com/EricLBuehler/xlora/pull/25 currently quantized model are trained using xlora , it start working with quantized model but facing issue when I want to make inference...

You can review my notebook at : https://colab.research.google.com/drive/1_B1ualsMbRfYWy0gdjdMi9RSDU-qmPHf#scrollTo=I4UZaqDAnnB6

Also, Checkout this notebook : https://colab.research.google.com/drive/1Eyh-mBd0LpcJwyzBHjGKhwNLQ9R74eLl?usp=drive_open Verify that a few lines are being repeated in the output.

What adjustments should we make if we wish to upgrade XLora for IA^3?

I am using your branch [alexrs:multi-ia3](https://github.com/alexrs/peft/tree/multi-ia3) to test the addd_weighted_adapter for ia3 FIRST_ADAPTER_PATH = "/content/drive/MyDrive/FedMl_test_llm/TrainedModels/WestLake_SFT_IA3_1/Weights/Epoch_1/Job_1" SECOND_ADAPTER_PATH = "/content/drive/MyDrive/FedMl_test_llm/TrainedModels/WestLake_SFT_IA3_2/Weights/Epoch_1/Job_1" FIRST_ADAPTER_NAME = "first" SECOND_ADAPTER_NAME = "second" model = PeftModel.from_pretrained(quantized_model_4bit, FIRST_ADAPTER_PATH, FIRST_ADAPTER_NAME) _...

I have tested this, it is working ![Screenshot 2024-05-02 171231](https://github.com/huggingface/peft/assets/101694671/a3d389c8-e261-4595-91cb-0facadf800de) ![Screenshot 2024-05-02 171255](https://github.com/huggingface/peft/assets/101694671/0e3fe721-6bd0-45d4-8ffb-7f57f7b294b3) it is working now, the next step could be for svd like strategies

i am also facing some issue relate ia3 that, Cannot merge ia3 layers when the model is loaded in 4-bit mention in this https://github.com/huggingface/peft/issues/1704 ![image](https://github.com/huggingface/peft/assets/101694671/ab4336e9-bc43-43bd-85bd-a5fc204eee16) ![Screenshot 2024-05-02 153000](https://github.com/huggingface/peft/assets/101694671/4aa4da8a-3d3b-4d41-80f3-e668754d1b0e) @BenjaminBossan please...