LLM-Intro to Large Language Models

LLM

some LLM's model and weight are not opened to user

what is?

Llama 270b model

  • 2 files

    • parameters file
      • parameter or weight of neural network
      • parameter -- 2bytes, float number
    • code run parameters(inference)
      • c or python, etc
      • for c, 500 lines code without dependency to run
      • self contained package(no network need)
  • how to get parameters?

    • lossy compress large chunk of text (10TB) with 6000 GPU for 12 days (cost 200$) to 140G zip file(gestalt of the text, weights and parameters)
  • what neural do is trying to predict the next word in a sequence. parameters are dispersed throughout the neural network and neurons are connected to each other, fire in a certain way

  • prediction has strong relationship with compression

  • LLM create a correct form of text and fill it with its knowedge. not create a copy of text that was be trained.

  • how does it work?


training stage

  • pre-training

    • expensive
    • base model. get a document generator model
    • it's about knowledge
    • internet documents
  • fine tuning

    • cheaper
    • assistant model. get a assistant model
    • it's about alighment
    • Q&A document
    • training with high quality conversation(question and answer).write labeling instructions to specify how assistant should behave
    • focus on quality not amount
  • stage 3(optional)

    • use comparison label
    • reenforcement learning from human feedback
  • labeling is a human-machine collaboration
  • rank of LLM

LLM scaling laws:

  • more D and N will get better model
  • multimodality. now some LLM like GPT can use different tools to help it with answering questions. browser, calculator, python interpreter.

  • future directions of development in LLM

give LLM system 2 ablility


  • LLM now only have system one(instinctive)
  • convert time to accuracy

self-improvement

  • in narrow domain it is possible to self-improve

customization

experts in certain domain

future of LLM

相关推荐
PaperData3 分钟前
2000-2023年地级市数字基础设施评价指标体系
大数据·网络·数据库·人工智能·数据分析·经管
政安晨4 分钟前
政安晨【OpenClaw与Hermes指南】AI Coding Agent行为约束之道:Karpathy CLAUDE.md技能体系深度解读
人工智能·ai coding·karpathy·agent行为约束之道·karpathy claude·技能体系解读·agent技能
ZYH_Core4 分钟前
DeepSeek V4 实战测评
人工智能·ai·ai编程
70asunflower12 分钟前
从Token到芯片:AI推理时代的效率竞争与市场逻辑
人工智能
xrgs_shz16 分钟前
MATLAB 纹理特征提取:一文读懂 graycomatrix 与 graycoprops
人工智能·计算机视觉·matlab
BlockChain88820 分钟前
AI+区块链深度探索:算法与账本的共生时代
人工智能·算法·区块链
CoderJia程序员甲30 分钟前
GitHub 热榜项目 - 日榜(2026-05-03)
ai·大模型·llm·github·ai教程
生成论实验室30 分钟前
《源·觉·知·行·事·物:生成论视域下的统一认知语法》第一章 源:不可言说的生成之源
人工智能·科技·算法·生活·创业创新
jinglong.zha35 分钟前
AI萌宠短剧实战:从0孵化动物IP,用AI制作爆款短视频
人工智能·ai·音视频·网赚教程·萌宠
AI医影跨模态组学40 分钟前
如何将CT影像语义特征与肝癌术后辅助TACE获益相关的免疫抑制性肿瘤微环境建立关联,并进一步解释其与预后、PA-TACE治疗响应的机制联系
人工智能·深度学习·论文·医学·医学影像·影像组学