yyyyynz
yyyyynz
nni-manager log file: ``` [2022-07-20 07:49:27] ERROR (tuner_command_channel.WebSocketChannel) Error: Error: tuner_command_channel: Tuner closed connection at WebSocket. (/home/nzy/miniconda3/envs/pyg1/lib/python3.8/site-packages/nni_node/core/tuner_command_channel/websocket_channel.js:41:49) at WebSocket.emit (node:events:538:35) at WebSocket.emitClose (/home/nzy/miniconda3/envs/pyg1/lib/python3.8/site-packages/nni_node/node_modules/express-ws/node_modules/ws/lib/websocket.js:246:10) at Receiver.receiverOnFinish (/home/nzy/miniconda3/envs/pyg1/lib/python3.8/site-packages/nni_node/node_modules/express-ws/node_modules/ws/lib/websocket.js:1037:20) at Receiver.emit (node:events:526:28) at...
您好,十分欣赏您及其团队所提出来的PathCon模型。由于我自己实验的需要,关于TransE, DistMult等基于Embedding的baseline方法,有些实现细节需要向您请教: 1. APPENDIX D中给出了baseline的代码的链接,请问超参的话是直接采用它们之前所采用的最佳超参吗? 2. 之前的方法在训练的时候是对头实体和尾实体进行破坏进行负采样,而就像4.1 Evaluation Protocol中提到的一样,本文是对关系进行预测,所以是对关系进行破坏进行负采样。那这样的话,负采样率这个超参还是沿用之前代码链接的负采样率吗?如果不是的话,想了解一下本文baseline实现时的负采样率。 3. 由公式16可知,本文是通过交叉熵,将预测关系转换为一个分类的问题,所以按照我的理解PathCon模型是没有进行负采样的(当然也可能理解错了)。我好奇baseline中的loss是直接沿用之前代码各自的实现,还是也采用PathCon这种方式的实现,即过完打分函数之后直接softmax再交叉熵(我的判断是直接采用的之前的实现,不然也不会提到对关系进行负采样)。因为基于[Ruffinelli et al.](https://openreview.net/forum?id=BkxSmlBFvr),我觉得不同loss的选择对模型的结果影响很大。不知您是否有做相关的实验。 (当然了,如果您愿意开源baseline的实验细节那是最好不过的了 4. 很好奇PathCon为什么没有像之前知识图谱链接预测的工作一样,给出头(尾)实体和关系,来预测尾(头)实体,而是给出两个实体预测它们之间的关系。我目前的推测是因为模型对关系上下文进行传递并结合两节点之间路径表示这种范式导致,所以直接给出头实体和尾实体预测关系更为直接和方便。但毕竟之前的论文都是采用链接预测这种方式,不知您是否有将PathCon在链接预测这种实验条件下进行实验?或者对于将PathCon用于链接预测的loss的构建有什么好的建议吗?
What is the purpose of testing and recording accuracy every 100 rounds during training? Isn't the pre-training process an unsupervised process? According to the DGI paper and code implementation, DGI...
As [this issues](https://github.com/thiviyanT/torch-rgcn/issues/12#issue-1042947158) mentioned, the link prediction code can't run. So I used a history commit `b1b6646`, it can run normally for link prediction task. But when I use the...
**Describe the issue**: After stopping a run of experiments at one time and opening a new one, all previous experiments that have ended their run are no longer available. Neither...
According to the implementation of other codes such as DGI, GRACE, etc., they all use the results of the last epoch or the minimum loss in the training process (early...
您好,看到您论文中transe,distmult,complex等baselines的效果都比原始论文的效果好很多。虽然您论文提到了所有模型超参选择的范围,不知是否方便公布一下各个baselines的最佳超参?十分感谢!