把 50 年方法演化的因果拓扑,转化为面向人与智能体的六个产品。
百万级 AI 论文构成的方法演化因果图谱,节点级查询、子图导出、祖先链追溯。
基于因果图谱生成科研创意——以拓扑约束代替组合幻觉,给出可追溯的研究路径。
5 维度科研创意评估:新颖性、可行性、显著性、有效性、清晰度,所有打分都基于图谱拓扑。
为 AI 智能体提供结构化科学记忆。/v1/query、/v1/trace、/v1/node 三个端点,一切以 JSON 流转。
由演化链自动生成的方法学研报:领域瓶颈、研究前沿与可追溯证据链。
开源知识图谱与论文池,CC-BY 许可。可下载、可二次构建,为开放科学而存。
每一个自动科研系统都在临时拼凑知识表示。 没人真正理解方法为何演化、解决了什么约束、 以及真正的前沿位于何处。
LLM 参数只是有偏的快照。罕见但关键的方法迁移,被压在分布的长尾里丢失。
"没人试过"和"试过但失败了"这两件事,在参数空间里是不可区分的。
"A 以精度为代价优化了 B 的效率,C 想兼得但失败了"——这种关系,没有任何地方记录。
Google Scholar、Semantic Scholar、Connected Papers 都停在文档粒度,没有任何一个能追溯方法的演化。
一条 8 段式流水线,把原始引文数据淬炼成可计算的因果拓扑。
沿引文图行走。每篇论文约 40 条引用,是演化边的天然候选。
粗筛:基于引文上下文的启发式规则。精排:对两篇论文做完整 LLM 分析。
"LoRA"、"Low-Rank Adaptation"、"Hu 等人的方法"——合并为同一节点,保留权威身份。
池外论文先建占位,保留入边。日后入池升级为正式节点——信息零丢失。
父子层级:Attention → MHA → GQA。方法之间存在父子层级关系。
内部主键 + 外部 ID 集(arXiv、DOI、S2)。arXiv 升级为期刊版后增量合并,绝不产生重复。
三级准入:Tier 1 顶会期刊自动入池,Tier 2 高引 arXiv,Tier 3 仅在被引用时建占位。
定期同步只处理新论文,不重跑全量——图谱以增量方式生长。
POST /v1/query
Content-Type: application/json
{
"idea": "Improving long-sequence modeling
in Transformers",
"depth": 2,
"year_range": [2020, 2025],
"max_nodes": 50
} {
"subgraph": {
"nodes": [
{
"id": "meth_00417",
"canonical_name": "Linear Attention",
"aliases": ["LinAttn", "Efficient Attention"],
"year": 2020,
"bottleneck_solved": "Quadratic memory in
sequence length",
"bottleneck_remaining": "Degraded recall
on associative tasks",
"parent_method": "meth_00291"
},
{
"id": "meth_00583",
"canonical_name": "Mamba",
"aliases": ["Selective SSM", "S6"],
"year": 2023,
"bottleneck_solved": "Content-aware
selection in SSMs",
"bottleneck_remaining": "No in-context
learning via attention heads",
"parent_method": "meth_00417"
}
],
"edges": [
{
"source": "meth_00417",
"target": "meth_00583",
"mechanism_delta": "Replace static kernel
with input-dependent selection gate"
}
]
},
"meta": { "total_nodes": 12, "depth_reached": 2 }
} POST /v1/trace
Content-Type: application/json
{
"node_id": "meth_00583",
"direction": "both",
"max_hops": 3
} {
"origin": {
"id": "meth_00583",
"canonical_name": "Mamba"
},
"upstream": [
{
"id": "meth_00417",
"canonical_name": "Linear Attention",
"year": 2020,
"hop": 1,
"mechanism_delta": "Replace static kernel
with input-dependent selection gate"
},
{
"id": "meth_00291",
"canonical_name": "Scaled Dot-Product
Attention",
"year": 2017,
"hop": 2,
"mechanism_delta": "Approximate softmax
via random feature map"
}
],
"downstream": [
{
"id": "meth_00641",
"canonical_name": "Jamba",
"year": 2024,
"hop": 1,
"mechanism_delta": "Interleave Mamba
layers with grouped-query attention"
},
{
"id": "meth_00659",
"canonical_name": "Mamba-2",
"year": 2024,
"hop": 1,
"mechanism_delta": "Reformulate SSM as
structured masked attention (SMA)"
}
],
"total_hops_upstream": 2,
"total_hops_downstream": 1
} GET /v1/node/meth_00583
Accept: application/json {
"id": "meth_00583",
"canonical_name": "Mamba",
"aliases": ["Selective SSM", "S6",
"Selective State Space"],
"year": 2023,
"paper_id": "arxiv:2312.00752",
"bottleneck_solved": "Content-aware selection
in state space models",
"bottleneck_remaining": "No in-context learning
via attention heads",
"parent_method": "meth_00417",
"edges": {
"incoming": [
{
"source": "meth_00417",
"source_name": "Linear Attention",
"mechanism_delta": "Replace static kernel
with input-dependent selection gate",
"relation": "derives_from"
},
{
"source": "meth_00389",
"source_name": "S4",
"mechanism_delta": "Make A, B, C matrices
input-dependent (drop LTI)",
"relation": "derives_from"
}
],
"outgoing": [
{
"target": "meth_00641",
"target_name": "Jamba",
"mechanism_delta": "Interleave Mamba layers
with grouped-query attention",
"relation": "enables"
},
{
"target": "meth_00659",
"target_name": "Mamba-2",
"mechanism_delta": "Reformulate SSM as
structured masked attention (SMA)",
"relation": "evolves_to"
}
]
}
} 把架构搜索变成图谱上的路径优化——在投入算力之前,先剪掉不可行的分支。
跨领域同构关系自动识别:GNN 传播 ≈ 流体动力学求解——这种关联在关键词检索里看不见。
基于因果约束生成研究创意,而不是组合空间里的幻觉。