farrokhsiar
farrokhsiar
In the `gpei_hartmann_service.ipynb` tutorial, before exploring using GPEI, a Sobol strategy has been used. Just curious if there is any reference on the best practices regarding the Sobol strategy: mainly...
I initiate a AxClient as it shown below: `_gs = GenerationStrategy( steps=[ GenerationStep( model=Models.SOBOL, num_trials=3, max_parallelism=max_parallelism, model_kwargs={"seed": 999}, model_gen_kwargs={}, ), GenerationStep( model=Models.GPEI, num_trials=-1, max_parallelism=self.max_parallelism, ), ] ) ax_client = AxClient(...
I have tried to train the algorithm using different order of tokens, and I have realized anytime the token order is random, it generates NaN in the embedding layer. I'm...
Is there any end to end example to show the library should be used to train/finetune a language model?
Hi This sounds like an interesting work, and I'm trying to replicate your work. Could you add the data format, or share the data that was used for this work?
Running modernbert within transformer Trainer class, with torch_compile=False, I get the following error: ``` raise AssertionError( torch._dynamo.exc.BackendCompilerFailed: backend='inductor' raised: AssertionError: Please convert all Tensors to FakeTensors first or instantiate FakeTensorMode...