Recent text generation methods frequently learn node representations from graph‐based data via global or local aggregation,such as knowledge graphs.Since all nodes are connected directly,node global representation en...Recent text generation methods frequently learn node representations from graph‐based data via global or local aggregation,such as knowledge graphs.Since all nodes are connected directly,node global representation encoding enables direct communication between two distant nodes while disregarding graph topology.Node local representation encoding,which captures the graph structure,considers the connections between nearby nodes but misses out onlong‐range relations.A quantum‐like approach to learning bettercontextualised node embeddings is proposed using a fusion model that combines both encoding strategies.Our methods significantly improve on two graph‐to‐text datasets compared to state‐of‐the‐art models in various experiments.展开更多
基金supported by the National Natural Science Foundation of China under Grant(62077015)the Key Laboratory of Intelligent Education Technology and Application of Zhejiang Province,Zhejiang Normal University,Zhejiang,China,the Key Research and Development Program of Zhejiang Province(No.2021C03141)the National Key R&D Program of China under Grant(2022YFC3303600).
文摘Recent text generation methods frequently learn node representations from graph‐based data via global or local aggregation,such as knowledge graphs.Since all nodes are connected directly,node global representation encoding enables direct communication between two distant nodes while disregarding graph topology.Node local representation encoding,which captures the graph structure,considers the connections between nearby nodes but misses out onlong‐range relations.A quantum‐like approach to learning bettercontextualised node embeddings is proposed using a fusion model that combines both encoding strategies.Our methods significantly improve on two graph‐to‐text datasets compared to state‐of‐the‐art models in various experiments.