8.16模型整理

文章目录

  • [Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation(ECCV2018)](#Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation(ECCV2018))
  • [Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning(2016)](#Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning(2016))
  • [Wide Residual Networks(2017)](#Wide Residual Networks(2017))
  • [mixup: Beyond Empirical Risk Minimization(ICLR2018)](#mixup: Beyond Empirical Risk Minimization(ICLR2018))
  • [Swin Transformer: Hierarchical Vision Transformer using Shifted Windows](#Swin Transformer: Hierarchical Vision Transformer using Shifted Windows)
  • [Pyramid Scene Parsing Network(2017)](#Pyramid Scene Parsing Network(2017))
  • [Searching for MobileNetV3(2019)](#Searching for MobileNetV3(2019))
  • [SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size(2016)](#SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size(2016))
  • [Identity Mappings in Deep Residual Networks(2016)](#Identity Mappings in Deep Residual Networks(2016))
  • [Aggregated Residual Transformations for Deep Neural Networks](#Aggregated Residual Transformations for Deep Neural Networks)
  • [MLP-Mixer: An all-MLP Architecture for Vision(2021)](#MLP-Mixer: An all-MLP Architecture for Vision(2021))
  • [MOCO:Momentum Contrast for Unsupervised Visual Representation Learning](#MOCO:Momentum Contrast for Unsupervised Visual Representation Learning)
  • [A ConvNet for the 2020s](#A ConvNet for the 2020s)
  • [MAE:Masked Autoencoders Are Scalable Vision Learners](#MAE:Masked Autoencoders Are Scalable Vision Learners)
  • [Xception: Deep Learning with Depthwise Separable Convolutions](#Xception: Deep Learning with Depthwise Separable Convolutions)
  • [CLIP:Learning Transferable Visual Models From Natural Language Supervision](#CLIP:Learning Transferable Visual Models From Natural Language Supervision)
  • [ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices](#ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices)
  • [ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design](#ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design)
  • [ResNeSt: Split-Attention Networks](#ResNeSt: Split-Attention Networks)

Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation(ECCV2018)

方法

代码地址

DeepLabV3+结构

Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning(2016)

方法



















Wide Residual Networks(2017)

方法

代码地址

我感觉是没啥变化

mixup: Beyond Empirical Risk Minimization(ICLR2018)

方法

主要看代码里面得lam和alpha

Swin Transformer: Hierarchical Vision Transformer using Shifted Windows

方法



Vit的滑动窗口版本

Pyramid Scene Parsing Network(2017)


Searching for MobileNetV3(2019)

方法

这是一篇关于网络架构搜索的文章

SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size(2016)

方法

Identity Mappings in Deep Residual Networks(2016)

方法

讲了各种各样的跳跃连接分析




Aggregated Residual Transformations for Deep Neural Networks

方法



相当于就是参数减少

MLP-Mixer: An all-MLP Architecture for Vision(2021)

token混合和channel混合

MOCO:Momentum Contrast for Unsupervised Visual Representation Learning

采用不同存储结构,moco采用的是队列

A ConvNet for the 2020s

做到极致的卷积

MAE:Masked Autoencoders Are Scalable Vision Learners

类似于bert,预测mask部分,自监督学习

Xception: Deep Learning with Depthwise Separable Convolutions

方法


CLIP:Learning Transferable Visual Models From Natural Language Supervision

方法

ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices

方法

分组卷积并混合

ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design

方法

ResNeSt: Split-Attention Networks

本文方法



相关推荐
Yuleave15 分钟前
高效流式大语言模型(StreamingLLM)——基于“注意力汇聚点”的突破性研究
人工智能·语言模型·自然语言处理
cqbzcsq17 分钟前
ESMC-600M蛋白质语言模型本地部署攻略
人工智能·语言模型·自然语言处理
刀客1231 小时前
python3+TensorFlow 2.x(四)反向传播
人工智能·python·tensorflow
SpikeKing1 小时前
LLM - 大模型 ScallingLaws 的设计 100B 预训练方案(PLM) 教程(5)
人工智能·llm·预训练·scalinglaws·100b·deepnorm·egs
小枫@码2 小时前
免费GPU算力,不花钱部署DeepSeek-R1
人工智能·语言模型
liruiqiang052 小时前
机器学习 - 初学者需要弄懂的一些线性代数的概念
人工智能·线性代数·机器学习·线性回归
Icomi_2 小时前
【外文原版书阅读】《机器学习前置知识》1.线性代数的重要性,初识向量以及向量加法
c语言·c++·人工智能·深度学习·神经网络·机器学习·计算机视觉
微学AI2 小时前
GPU算力平台|在GPU算力平台部署可图大模型Kolors的应用实战教程
人工智能·大模型·llm·gpu算力
西猫雷婶2 小时前
python学opencv|读取图像(四十六)使用cv2.bitwise_or()函数实现图像按位或运算
人工智能·opencv·计算机视觉
IT古董2 小时前
【深度学习】常见模型-生成对抗网络(Generative Adversarial Network, GAN)
人工智能·深度学习·生成对抗网络