Jiaxing Liu
Jiaxing Liu
rt,算出的结果和我用Numpy算出的结果不一致,且差异较大。
### Describe the bug There is a bug in the test_op function within the python/tutorials/06-fused-attention.py tutorial file. Under specific q,k input, the assertion assert torch.allclose(ref_dv, tri_dv, atol=1e-2, rtol=rtol) fails. Failing...
Loaded as API: https://bigcode-bigcodebench-evaluator.hf.space/ ✔ Traceback (most recent call last): File "", line 1, in File "/home/ma-user/anaconda3/envs/evalplus_env/lib/python3.9/site-packages/gradio_client/client.py", line 171, in __init__ self._info = self._get_api_info() File "/home/ma-user/anaconda3/envs/evalplus_env/lib/python3.9/site-packages/gradio_client/client.py", line 566, in _get_api_info...
Under this setting, my evaluation results on qwen2.5coder-instruct-3b is betther than results claimed from the officical techinique report.
Due to Qwen2.5 coder does not suppot functional call, I wonder the functional calling format of your model..