luduling

Results 3 comments of luduling

env: Tensorflow 1.12.0 Tensor2Tensor 1.7 modify g2p.py the line output_node_names = ["transformer/parallel_0_5/transformer/body/decoder/" "layer_0/self_attention/multihead_attention/dot_product_attention/" "Softmax"...] To output_node_names = ["transformer/parallel_0_4/transformer/transformer/body/encoder/" "layer_0/self_attention/multihead_attention/dot_product_attention/" "attention_weights", "transformer/parallel_0_4/transformer/transformer/body/encoder/" "layer_1/self_attention/multihead_attention/dot_product_attention/" "attention_weights", "transformer/parallel_0_4/transformer/transformer/body/encoder/" "layer_2/self_attention/multihead_attention/dot_product_attention/" "attention_weights", "transformer/parallel_0_4/transformer/transformer/body/decoder/" "layer_0/self_attention/multihead_attention/dot_product_attention/" "attention_weights", "transformer/parallel_0_4/transformer/transformer/body/decoder/"...

@sendrolon 问一下 本地ollama部署的qwen 3.0 如何关闭 thinking 功能?我没有找到 ollama 开关 thinking的方法,vLLM是有开关 thinking 参数的。