[综述笔记]Benchmarking Graph Neural Networks for FMRI analysis

论文网址:[2211.08927] Benchmarking Graph Neural Networks for FMRI analysis (arxiv.org)

⭐不是真正意义上的综述,应该是分析性质的文章

英文是纯手打的!论文原文的summarizing and paraphrasing。可能会出现难以避免的拼写错误和语法错误,若有发现欢迎评论指正!文章偏向于笔记,谨慎食用

目录

[1. 省流版](#1. 省流版)

[1.1. 心得](#1.1. 心得)

[1.2. 论文总结图](#1.2. 论文总结图)

[2. 论文逐段精读](#2. 论文逐段精读)

[2.1. Abstract](#2.1. Abstract)

[2.2. Introduction](#2.2. Introduction)

[2.3. Background](#2.3. Background)

[2.3.1. Graph convolutional networks](#2.3.1. Graph convolutional networks)

[2.3.2. Graph attention networks](#2.3.2. Graph attention networks)

[2.3.3. Graph isomorphic network](#2.3.3. Graph isomorphic network)

[2.3.4. Spatio-temporal graph convolution network](#2.3.4. Spatio-temporal graph convolution network)

[2.3.5. Adaptive spatio-temporal graph convolution network](#2.3.5. Adaptive spatio-temporal graph convolution network)

[2.4. Methods](#2.4. Methods)

[2.4.1. Graph creation](#2.4.1. Graph creation)

[2.4.2. Graph Neural Networks](#2.4.2. Graph Neural Networks)

[2.4.3. Graph Readout](#2.4.3. Graph Readout)

[2.5. Data & Experimental Setup](#2.5. Data & Experimental Setup)

[2.5.1. Datasets](#2.5.1. Datasets)

[2.5.2. Model selection & assessment](#2.5.2. Model selection & assessment)

[2.5.3. Non-graph baselines](#2.5.3. Non-graph baselines)

[2.6. Results](#2.6. Results)

[2.6.1. Disorder diagnosis](#2.6.1. Disorder diagnosis)

[2.6.2. Scaling performance on the UKBioBank](#2.6.2. Scaling performance on the UKBioBank)

[2.6.3. Graph diffusion improves learning on noisy graphs](#2.6.3. Graph diffusion improves learning on noisy graphs)

[2.7. Discussion](#2.7. Discussion)

[3. 知识补充](#3. 知识补充)

[3.1. AST-GCN](#3.1. AST-GCN)

[3.2. graph diffusion convolution (GDC)](#3.2. graph diffusion convolution (GDC))

[4. Reference List](#4. Reference List)


1. 省流版

1.1. 心得

(1)有人能把节点特征当人话说出来也是挺不容易的(抹泪)(说节点特征是节点表示的真的呵呵)

(2)虽然给我拓宽了视野了解了AST-GCN,但是这...看了看标题,这真的适合fMRI数据吗?我们脑子这么随机的吗,电信号也不至于random walk吧

(3)所以你的价值是把每个baseline搬过来测一下?

(4)⭐"超参数调优的一个常见缺陷是根据最终测试集的性能进行搜索。这导致对模型泛化能力的过于乐观的估计。"

(5)你很喜欢问问题吗?

(6)侧重确实是比基线的,然后作者认为GNN都没有很行,然后觉得FC阈值选择很重要。其实不用这么care基线行不行吧?感觉基线只是一个,前人证明某个东西能吃,鲫鱼能吃金鱼不能吃。但是生吃鲫鱼也不好吃吧?那不得看后面的清蒸红烧黄焖炸煮糖醋各种各样的后期加工吗?

1.2. 论文总结图

2. 论文逐段精读

2.1. Abstract

①Testing 5 baselines of GNN and comparing them with kernel based or pixel based CNN. Then, the authors found that the performance of GNN is worse than CNN

②They reckon that GNN is limited to graph construction

2.2. Introduction

①GNN is used for phenotype prediction, disease diagnosis, and task inference

②They focus on testing the baselines

2.3. Background

2.3.1. Graph convolutional networks

①Briefly introducing GCN(额.....)

2.3.2. Graph attention networks

①....

2.3.3. Graph isomorphic network

2.3.4. Spatio-temporal graph convolution network

①ST-GCNs extract temporal features by 1D Conv, LSTM or others and spatial features by GCN

2.3.5. Adaptive spatio-temporal graph convolution network

①AST-GCNs are able to randomly initialize graph structure

2.4. Methods

2.4.1. Graph creation

①The most popular way of setting node feature in static brain graph: row of Pearson correlation(不得不说概念上来说这样定义是真**奇怪)

②The most popular way of setting edge feature in static brain graph: Pearson correlation

③"Pooling": threshold ratio hyperparameter

④Node feature of dynamic graph: time series of each ROI

⑤Edge feature/pooling strategy of dynamic graph: same as static for ST-GCN and random initialization for AST-GCN

⑥Graph construction of 5 baselines:

2.4.2. Graph Neural Networks

(1)GCNs

①Introducing spectral convolution and how it evolve into spatial convolution

(2)GATs

①Introducing GAT

②In their training, they set M=1 and choose LeakyReLU

(3)GINs

①Introducing GIN

(4)ST-GCNs

①Node feature: whole time series

②似乎没说M是什么

(5)AST-GCNs

①⭐"图结构的自适应学习是gnn领域的一个日益增长的研究领域"

2.4.3. Graph Readout

①They adopt MEAN, MEAN||MAX, and SUM readout methods

2.5. Data & Experimental Setup

2.5.1. Datasets

(1)ABIDE I/II

①Site: ABIDE I (17) + ABIDE II (19) = 29 哈哈哈哈哈还有二次参与的啊

②Sample: 1207 with 558 TD and 649 ASD (the largest 9 sites)

③Preprocessing:CPAC

④Atlas: craddock-200

(2)Rest-meta-MDD

①Cohort(?): 25

②Sample: 1453 with 628 HC and 825 MDD (the largest 9 sites)

③Preprocessing: DPARSF

④Atlas: Harvard-Oxford 112

(3)UkBioBank

①Sample: 5500 with 2750 males and 2750 females

②Atlas: AAL 116

2.5.2. Model selection & assessment

①⭐Fining hyperparameters rely on testing performance reduces the generalization of model. Thus they search hyperparameters on valiadtion set

②Cross validation: 5 fold

③Data split: 85/15 training/validation split for ABIDE and MDD; training set = [500,1000,2000,5000] on UkBioBank

pitfall n. 隐患;陷阱

2.5.3. Non-graph baselines

①Structure-agnostic baselines: MLP and support vector machines with radial basis function kernel (SVM-rbf) to learn the static lower-triangular matrix of each subject; 1D CNN for dynamic time series training

2.6. Results

2.6.1. Disorder diagnosis

①Comparison table:

there is no outstanding model and static methods outperform dynamic

2.6.2. Scaling performance on the UKBioBank

①Discussing the results on UKBioBank:

2.6.3. Graph diffusion improves learning on noisy graphs

①Threshold of FC: 50%-95% sparsity with stride equaling to 5%

②⭐There is noise under sparsification, graph diffusion convolution (GDC) may solve this problem with:

where denotes weighting coefficients and denotes generalized transition matrix. There are several choices.

③The effect of GDC on GCN with heat kernel coefficient and k = 2:

model is sensitive to GDC under low sparsification

2.7. Discussion

①你不会分段吗?

②Discussions......

3. 知识补充

3.1. AST-GCN

AST-GCN(Attention Spatial-Temporal Graph Convolutional Networks)是一种结合了注意力机制的时空图卷积网络。它主要用于处理和分析具有空间和时间相关性的数据,如交通流量预测、航拍视频暴力行为识别等。

在交通流量预测中,AST-GCN包含三个独立分量,分别模拟交通流量的近期依赖性、日周期性和周周期性。其主要贡献在于通过空间注意力捕捉不同位置之间的空间相关性,通过时间注意力捕捉不同时间之间的时间相关性。此外,AST-GCN还设计了时空卷积模块,包括空间图卷积和时间卷积,在真实公路交通流量数据集上取得了显著的效果。

在航拍视频暴力行为识别中,AST-GCN被用于解决航拍成像中目标出现运动模糊、尺度变化大的问题。该方法首先利用关键帧检测网络完成初定位,然后通过AST-GCN网络提取序列特征以完成行为识别确认。

AST-GCN的出现为处理和分析具有复杂时空相关性的数据提供了新的思路和方法,具有广阔的应用前景。如需更深入地了解AST-GCN的原理、实现细节以及应用案例,建议查阅相关论文和资料。

3.2. graph diffusion convolution (GDC)

参考学习:图扩散卷积:Graph_Diffusion_Convolution-CSDN博客

论文原文:23c894276a2c5a16470e6a31f4618d73-Paper.pdf (neurips.cc)

4. Reference List

ElGazzar, A., Thomas, R. & Wingen, G. (2022) 'Benchmarking Graph Neural Networks for FMRI analysis', arXiv: 2211.08927.

相关推荐
wxin_VXbishe13 分钟前
springboot合肥师范学院实习实训管理系统-计算机毕业设计源码31290
java·spring boot·python·spring·servlet·django·php
NuyoahC18 分钟前
算法笔记(十一)——优先级队列(堆)
c++·笔记·算法·优先级队列
ITenderL19 分钟前
Python学习笔记-函数
python·学习笔记
Langchain21 分钟前
不可错过!CMU最新《生成式人工智能大模型》课程:从文本、图像到多模态大模型
人工智能·自然语言处理·langchain·大模型·llm·大语言模型·多模态大模型
zmjia11121 分钟前
全流程Python编程、机器学习与深度学习实践技术应用
python·深度学习·机器学习
龙的爹233332 分钟前
论文翻译 | Generated Knowledge Prompting for Commonsense Reasoning
人工智能·gpt·机器学习·语言模型·自然语言处理·nlp·prompt
龙的爹233333 分钟前
论文翻译 | Model-tuning Via Prompts Makes NLP Models Adversarially Robust
人工智能·gpt·语言模型·自然语言处理·nlp·prompt
方世恩1 小时前
【进阶OpenCV】 (5)--指纹验证
人工智能·opencv·目标检测·计算机视觉
计算机学姐1 小时前
基于SpringBoot+Vue的在线投票系统
java·vue.js·spring boot·后端·学习·intellij-idea·mybatis
_.Switch1 小时前
Python机器学习:自然语言处理、计算机视觉与强化学习
python·机器学习·计算机视觉·自然语言处理·架构·tensorflow·scikit-learn