8.16模型整理

文章目录

  • [Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation(ECCV2018)](#Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation(ECCV2018))
  • [Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning(2016)](#Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning(2016))
  • [Wide Residual Networks(2017)](#Wide Residual Networks(2017))
  • [mixup: Beyond Empirical Risk Minimization(ICLR2018)](#mixup: Beyond Empirical Risk Minimization(ICLR2018))
  • [Swin Transformer: Hierarchical Vision Transformer using Shifted Windows](#Swin Transformer: Hierarchical Vision Transformer using Shifted Windows)
  • [Pyramid Scene Parsing Network(2017)](#Pyramid Scene Parsing Network(2017))
  • [Searching for MobileNetV3(2019)](#Searching for MobileNetV3(2019))
  • [SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size(2016)](#SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size(2016))
  • [Identity Mappings in Deep Residual Networks(2016)](#Identity Mappings in Deep Residual Networks(2016))
  • [Aggregated Residual Transformations for Deep Neural Networks](#Aggregated Residual Transformations for Deep Neural Networks)
  • [MLP-Mixer: An all-MLP Architecture for Vision(2021)](#MLP-Mixer: An all-MLP Architecture for Vision(2021))
  • [MOCO:Momentum Contrast for Unsupervised Visual Representation Learning](#MOCO:Momentum Contrast for Unsupervised Visual Representation Learning)
  • [A ConvNet for the 2020s](#A ConvNet for the 2020s)
  • [MAE:Masked Autoencoders Are Scalable Vision Learners](#MAE:Masked Autoencoders Are Scalable Vision Learners)
  • [Xception: Deep Learning with Depthwise Separable Convolutions](#Xception: Deep Learning with Depthwise Separable Convolutions)
  • [CLIP:Learning Transferable Visual Models From Natural Language Supervision](#CLIP:Learning Transferable Visual Models From Natural Language Supervision)
  • [ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices](#ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices)
  • [ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design](#ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design)
  • [ResNeSt: Split-Attention Networks](#ResNeSt: Split-Attention Networks)

Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation(ECCV2018)

方法

代码地址

DeepLabV3+结构

Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning(2016)

方法



















Wide Residual Networks(2017)

方法

代码地址

我感觉是没啥变化

mixup: Beyond Empirical Risk Minimization(ICLR2018)

方法

主要看代码里面得lam和alpha

Swin Transformer: Hierarchical Vision Transformer using Shifted Windows

方法



Vit的滑动窗口版本

Pyramid Scene Parsing Network(2017)


Searching for MobileNetV3(2019)

方法

这是一篇关于网络架构搜索的文章

SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size(2016)

方法

Identity Mappings in Deep Residual Networks(2016)

方法

讲了各种各样的跳跃连接分析




Aggregated Residual Transformations for Deep Neural Networks

方法



相当于就是参数减少

MLP-Mixer: An all-MLP Architecture for Vision(2021)

token混合和channel混合

MOCO:Momentum Contrast for Unsupervised Visual Representation Learning

采用不同存储结构,moco采用的是队列

A ConvNet for the 2020s

做到极致的卷积

MAE:Masked Autoencoders Are Scalable Vision Learners

类似于bert,预测mask部分,自监督学习

Xception: Deep Learning with Depthwise Separable Convolutions

方法


CLIP:Learning Transferable Visual Models From Natural Language Supervision

方法

ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices

方法

分组卷积并混合

ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design

方法

ResNeSt: Split-Attention Networks

本文方法



相关推荐
QBoson28 分钟前
量子计算+AI:特征选择与神经网络优化创新应用
人工智能·神经网络·量子计算·图像分类·特征选择·“五岳杯”量子计算挑战赛·相干光量子计算机
Juicedata1 小时前
JuiceFS v1.3-Beta2:集成 Apache Ranger,实现更精细化的权限控制
运维·人工智能·ai
Work(沉淀版)3 小时前
DAY 40
人工智能·深度学习·机器学习
蓦然回首却已人去楼空4 小时前
Build a Large Language Model (From Scratch) 序章
人工智能·语言模型·自然语言处理
CM莫问4 小时前
<论文>(微软)WINA:用于加速大语言模型推理的权重感知神经元激活
人工智能·算法·语言模型·自然语言处理·大模型·推理加速
拾忆-eleven4 小时前
NLP学习路线图(二十六):自注意力机制
人工智能·深度学习
MYH5165 小时前
在NLP文本处理中,将字符映射到阿拉伯数字(构建词汇表vocab)的核心目的和意义
人工智能·深度学习·自然语言处理
要努力啊啊啊6 小时前
KV Cache:大语言模型推理加速的核心机制详解
人工智能·语言模型·自然语言处理
mzlogin7 小时前
DIY|Mac 搭建 ESP-IDF 开发环境及编译小智 AI
人工智能
归去_来兮8 小时前
知识图谱技术概述
大数据·人工智能·知识图谱