AXI4Arbiter object scala code reading

object AXI4Arbiter

{

def apply[T <: Data](policy: TLArbiter.Policy)(sink: IrrevocableIO[T], sources: IrrevocableIO[T]*): Unit = {

if (sources.isEmpty) {

sink.valid := false.B

} else {

returnWinner(policy)(sink, sources:_*)

}

}

def returnWinner[T <: Data](policy: TLArbiter.Policy)(sink: IrrevocableIO[T], sources: IrrevocableIO[T]*) = {

require (!sources.isEmpty)

// The arbiter is irrevocable; when !idle, repeat last request

val idle = RegInit(true.B)

// Who wants access to the sink?

val valids = sources.map(_.valid)

val anyValid = valids.reduce(_ || _)

// Arbitrate amongst the requests

val readys = VecInit(policy(valids.size, Cat(valids.reverse), idle).asBools)

// Which request wins arbitration?

val winner = VecInit((readys zip valids) map { case (r,v) => r&&v })

// Confirm the policy works properly

require (readys.size == valids.size)

// Never two winners

val prefixOR = winner.scanLeft(false.B)(||).init

assert((prefixOR zip winner) map { case (p,w) => !p || !w } reduce {_ && _})

// If there was any request, there is a winner

assert (!anyValid || winner.reduce(||))

// The one-hot source granted access in the previous cycle

val state = RegInit(VecInit.fill(sources.size)(false.B))

val muxState = Mux(idle, winner, state)

state := muxState

// Determine when we go idle

when (anyValid) { idle := false.B }

when (sink.fire) { idle := true.B }

if (sources.size > 1) {

val allowed = Mux(idle, readys, state)

(sources zip allowed) foreach { case (s, r) =>

s.ready := sink.ready && r

}

} else {

sources(0).ready := sink.ready

}

sink.valid := Mux(idle, anyValid, Mux1H(state, valids))

sink.bits :<= Mux1H(muxState, sources.map(_.bits))

muxState

}

}

相关推荐
隔着天花板看星星5 天前
SparkSQL-性能调优
大数据·分布式·sql·spark·scala
武子康6 天前
大数据-156 Apache Druid 案例实战 Scala Kafka 订单统计
java·大数据·clickhouse·flink·kafka·scala·apache
小亚文6 天前
scala 2.12 error: value foreach is not a member of Object
开发语言·后端·scala
pblh1236 天前
2023_Spark_实验九:编写WordCount程序(Scala版)
大数据·spark·scala
小黑036 天前
Scala第一天
开发语言·scala
武子康9 天前
大数据-150 Apache Druid 安装部署 单机启动 系统架构
java·大数据·hadoop·clickhouse·系统架构·scala·apache
武子康9 天前
大数据-146 Apache Kudu 安装运行 Dockerfile 模拟集群 启动测试
java·大数据·clickhouse·spark·scala·apache
小黑039 天前
Scala第二天
开发语言·后端·scala
lwprain10 天前
用idea编写并运行第一个spark scala处理程序
spark·scala·intellij-idea
隔着天花板看星星11 天前
SparkSQL-初识
大数据·分布式·sql·spark·scala