-
Notifications
You must be signed in to change notification settings - Fork 32
Expand file tree
/
Copy pathMainPipe.scala
More file actions
791 lines (699 loc) · 34.2 KB
/
MainPipe.scala
File metadata and controls
791 lines (699 loc) · 34.2 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
/** *************************************************************************************
* Copyright (c) 2020-2021 Institute of Computing Technology, Chinese Academy of Sciences
* Copyright (c) 2020-2021 Peng Cheng Laboratory
*
* XiangShan is licensed under Mulan PSL v2.
* You can use this software according to the terms and conditions of the Mulan PSL v2.
* You may obtain a copy of Mulan PSL v2 at:
* http://license.coscl.org.cn/MulanPSL2
*
* THIS SOFTWARE IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OF ANY KIND,
* EITHER EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO NON-INFRINGEMENT,
* MERCHANTABILITY OR FIT FOR A PARTICULAR PURPOSE.
*
* See the Mulan PSL v2 for more details.
* *************************************************************************************
*/
package coupledL2.tl2tl
import chisel3._
import chisel3.util._
import utility._
import coupledL2.MetaData._
import org.chipsalliance.cde.config.Parameters
import freechips.rocketchip.tilelink._
import freechips.rocketchip.tilelink.TLMessages._
import freechips.rocketchip.tilelink.TLPermissions._
import coupledL2._
import coupledL2.utils._
import coupledL2.debug._
import coupledL2.prefetch.{PfSource, PrefetchTrain}
class MainPipe(implicit p: Parameters) extends L2Module with HasPerfEvents {
val io = IO(new Bundle() {
/* receive task from arbiter at stage 2 */
val taskFromArb_s2 = Flipped(ValidIO(new TaskBundle()))
/* handle set conflict in req arb */
val fromReqArb = Input(new Bundle() {
val status_s1 = new PipeEntranceStatus
})
/* block B and C at Entrance */
val toReqArb = Output(new BlockInfo())
/* block A at Entrance */
val toReqBuf = Output(Vec(2, Bool()))
/* handle capacity conflict of GrantBuffer */
val status_vec_toD = Vec(3, ValidIO(new PipeStatus))
/* handle capacity conflict of SourceC */
val status_vec_toC = Vec(3, ValidIO(new PipeStatus))
/* get dir result at stage 3 */
val dirResp_s3 = Input(new DirResult)
val replResp = Flipped(ValidIO(new ReplacerResult))
/* send task to MSHRCtl at stage 3 */
val toMSHRCtl = new Bundle() {
val mshr_alloc_s3 = ValidIO(new MSHRRequest())
}
val fromMSHRCtl = new Bundle() {
val mshr_alloc_ptr = Input(UInt(mshrBits.W))
}
/* read C-channel Release Data and write into DS */
val bufResp = Input(new PipeBufferResp)
/* get ReleaseBuffer and RefillBuffer read result */
val refillBufResp_s3 = Flipped(ValidIO(new DSBlock))
val releaseBufResp_s3 = Flipped(ValidIO(new DSBlock))
/* read or write data storage */
val toDS = new Bundle() {
val en_s3 = Output(Bool())
val req_s3 = ValidIO(new DSRequest)
val rdata_s5 = Input(new DSBlock)
val wdata_s3 = Output(new DSBlock)
val error_s5 = Input(Bool())
}
/* send Release/Grant/ProbeAck via SourceC/D channels */
val toSourceC, toSourceD = DecoupledIO(new Bundle() {
val task = new TaskBundle
val data = new DSBlock
})
/* write dir, including reset dir */
val metaWReq = ValidIO(new MetaWrite)
val tagWReq = ValidIO(new TagWrite)
/* read DS and write data into ReleaseBuf when the task needs to replace */
val releaseBufWrite = ValidIO(new MSHRBufWrite())
val nestedwb = Output(new NestedWriteback)
val nestedwbData = Output(new DSBlock)
/* send Hint to L1 */
val l1Hint = DecoupledIO(new L2ToL1Hint())
/* receive s1 info for Hint */
val taskInfo_s1 = Flipped(ValidIO(new TaskBundle()))
/* send prefetchTrain to Prefetch to trigger a prefetch req */
val prefetchTrain = prefetchOpt.map(_ => DecoupledIO(new PrefetchTrain))
val toMonitor = Output(new MainpipeMoni())
/* ECC error*/
val error = ValidIO(new L2CacheErrorInfo)
})
val resetFinish = RegInit(false.B)
val resetIdx = RegInit((cacheParams.sets - 1).U)
/* block reqs when reset */
when(!resetFinish) {
resetIdx := resetIdx - 1.U
}
when(resetIdx === 0.U) {
resetFinish := true.B
}
val c_s3, c_s4, c_s5 = Wire(io.toSourceC.cloneType)
val d_s3, d_s4, d_s5 = Wire(io.toSourceD.cloneType)
/* ======== Stage 2 ======== */
// send out MSHR task if data is not needed
val task_s2 = io.taskFromArb_s2
val hasData_s2 = task_s2.bits.opcode(0)
/* ======== Stage 3 ======== */
val task_s3 = RegInit(0.U.asTypeOf(Valid(new TaskBundle())))
task_s3.valid := task_s2.valid
when(task_s2.valid) {
task_s3.bits := task_s2.bits
}
/* ======== Enchantment ======== */
val dirResult_s3 = io.dirResp_s3
val meta_s3 = dirResult_s3.meta
val req_s3 = task_s3.bits
val tagError_s3 = io.dirResp_s3.error || meta_s3.tagErr
val dataError_s3 = meta_s3.dataErr
val l2Error_s3 = io.dirResp_s3.error
val mshr_req_s3 = req_s3.mshrTask
val sink_req_s3 = !mshr_req_s3
val sinkA_req_s3 = !mshr_req_s3 && req_s3.fromA
val sinkB_req_s3 = !mshr_req_s3 && req_s3.fromB
val sinkC_req_s3 = !mshr_req_s3 && req_s3.fromC
val req_acquire_s3 = sinkA_req_s3 && (req_s3.opcode === AcquireBlock || req_s3.opcode === AcquirePerm)
val req_acquireBlock_s3 = sinkA_req_s3 && req_s3.opcode === AcquireBlock
val req_prefetch_s3 = sinkA_req_s3 && req_s3.opcode === Hint
val req_get_s3 = sinkA_req_s3 && req_s3.opcode === Get
val mshr_grant_s3 = mshr_req_s3 && req_s3.fromA && (req_s3.opcode === Grant || req_s3.opcode === GrantData) // Grant or GrantData from mshr
val mshr_grantdata_s3 = mshr_req_s3 && req_s3.fromA && req_s3.opcode === GrantData
val mshr_accessackdata_s3 = mshr_req_s3 && req_s3.fromA && req_s3.opcode === AccessAckData
val mshr_hintack_s3 = mshr_req_s3 && req_s3.fromA && req_s3.opcode === HintAck
val mshr_probeack_s3 = mshr_req_s3 && req_s3.fromB && (req_s3.opcode === ProbeAck || req_s3.opcode === ProbeAckData) // ProbeAck or ProbeAckData from mshr
val mshr_probeackdata_s3 = mshr_req_s3 && req_s3.fromB && req_s3.opcode === ProbeAckData
val mshr_release_s3 = mshr_req_s3 && (req_s3.opcode === Release || req_s3.opcode === ReleaseData) // voluntary Release or ReleaseData from mshr
val mshr_releasedata_s3 = mshr_req_s3 && req_s3.opcode === ReleaseData
val meta_has_clients_s3 = meta_s3.clients.orR
val req_needT_s3 = needT(req_s3.opcode, req_s3.param) // require T status to handle req
// val a_need_replacement = sinkA_req_s3 && !dirResult_s3.hit && meta_s3.state =/= INVALID // b and c do not need replacement
//[Alias] TODO: consider 1 client for now
val cache_alias = req_acquire_s3 && dirResult_s3.hit && meta_s3.clients(0) &&
meta_s3.alias.getOrElse(0.U) =/= req_s3.alias.getOrElse(0.U)
val mshr_refill_s3 = (mshr_accessackdata_s3 || mshr_hintack_s3 || mshr_grant_s3) // needs refill to L2 DS
val retry = io.replResp.bits.retry
val need_repl = io.replResp.bits.meta.state =/= INVALID && req_s3.replTask // Grant needs replacement
/* ======== Interact with MSHR ======== */
val acquire_on_miss_s3 = req_acquire_s3 || req_prefetch_s3 || req_get_s3 // TODO: remove this cause always acquire on miss?
val acquire_on_hit_s3 = meta_s3.state === BRANCH && req_needT_s3 && !req_prefetch_s3
// For channel A reqs, alloc mshr when: acquire downwards is needed || alias
val need_acquire_s3_a = req_s3.fromA && Mux(
dirResult_s3.hit,
acquire_on_hit_s3,
acquire_on_miss_s3
)
val need_probe_s3_a = req_get_s3 && dirResult_s3.hit && meta_s3.state === TRUNK
val need_mshr_s3_a = need_acquire_s3_a || need_probe_s3_a || cache_alias
// For channel B reqs, alloc mshr when Probe hits in both self and client dir
val need_mshr_s3_b = dirResult_s3.hit && req_s3.fromB &&
!((meta_s3.state === BRANCH || meta_s3.state === TIP) && req_s3.param === toB) &&
meta_has_clients_s3
// For channel C reqs, Release will always hit on MainPipe, no need for MSHR
val need_mshr_s3 = need_mshr_s3_a || need_mshr_s3_b
/* Signals to MSHR Ctl */
// Allocation of MSHR: new request only
val alloc_state = WireInit(0.U.asTypeOf(new FSMState()))
alloc_state.elements.foreach(_._2 := true.B)
io.toMSHRCtl.mshr_alloc_s3.valid := task_s3.valid && !mshr_req_s3 && need_mshr_s3
io.toMSHRCtl.mshr_alloc_s3.bits.dirResult := dirResult_s3
io.toMSHRCtl.mshr_alloc_s3.bits.state := alloc_state
val ms_task = io.toMSHRCtl.mshr_alloc_s3.bits.task
ms_task.channel := req_s3.channel
ms_task.set := req_s3.set
ms_task.tag := req_s3.tag
ms_task.off := req_s3.off
ms_task.alias.foreach(_ := req_s3.alias.getOrElse(0.U))
ms_task.vaddr.foreach(_ := req_s3.vaddr.getOrElse(0.U))
ms_task.isKeyword.foreach(_ := req_s3.isKeyword.get) //OrElse(false.B))
ms_task.opcode := req_s3.opcode
ms_task.param := req_s3.param
ms_task.size := req_s3.size
ms_task.sourceId := req_s3.sourceId
ms_task.bufIdx := 0.U(bufIdxBits.W)
ms_task.needProbeAckData := req_s3.needProbeAckData
ms_task.mshrTask := false.B
ms_task.mshrId := 0.U(mshrBits.W)
ms_task.aliasTask.foreach(_ := cache_alias)
ms_task.useProbeData := false.B
ms_task.readProbeDataDown := false.B
ms_task.mshrRetry := false.B
ms_task.fromL2pft.foreach(_ := req_s3.fromL2pft.get)
ms_task.needHint.foreach(_ := req_s3.needHint.get)
ms_task.dirty := false.B
ms_task.way := req_s3.way
ms_task.meta := 0.U.asTypeOf(new MetaEntry)
ms_task.metaWen := false.B
ms_task.tagWen := false.B
ms_task.dsWen := false.B
ms_task.wayMask := 0.U(cacheParams.ways.W)
ms_task.replTask := false.B
ms_task.cmoTask := false.B
ms_task.reqSource := req_s3.reqSource
ms_task.mergeA := req_s3.mergeA
ms_task.aMergeTask := req_s3.aMergeTask
ms_task.txChannel := 0.U
ms_task.snpHitRelease := false.B
ms_task.snpHitReleaseToInval := false.B
ms_task.snpHitReleaseToClean := false.B
ms_task.snpHitReleaseWithData := false.B
ms_task.snpHitReleaseIdx := 0.U
ms_task.snpHitReleaseMeta := MetaEntry()
ms_task.denied := false.B
ms_task.corrupt := false.B
ms_task.cmoAll := false.B
/* ======== Resps to SinkA/B/C Reqs ======== */
val sink_resp_s3 = WireInit(0.U.asTypeOf(Valid(new TaskBundle))) // resp for sinkA/B/C request that does not need to alloc mshr
val sink_resp_s3_a_promoteT = dirResult_s3.hit && isT(meta_s3.state)
sink_resp_s3.valid := task_s3.valid && !mshr_req_s3 && !need_mshr_s3
sink_resp_s3.bits := task_s3.bits
sink_resp_s3.bits.mshrId := (1 << (mshrBits-1)).U + sink_resp_s3.bits.sourceId // extra id for reqs that do not enter mshr
when(req_s3.fromA) {
sink_resp_s3.bits.opcode := odOpGen(req_s3.opcode)
sink_resp_s3.bits.param := Mux(
req_acquire_s3,
Mux(req_s3.param === NtoB && !sink_resp_s3_a_promoteT, toB, toT),
0.U // reserved
)
}.elsewhen(req_s3.fromB) {
sink_resp_s3.bits.opcode := Mux(
dirResult_s3.hit && (meta_s3.state === TIP && meta_s3.dirty || req_s3.needProbeAckData),
ProbeAckData,
ProbeAck
)
sink_resp_s3.bits.param := Mux(!dirResult_s3.hit, NtoN,
MuxLookup(Cat(req_s3.param, meta_s3.state), BtoB)(Seq(
Cat(toN, BRANCH) -> BtoN,
Cat(toN, TIP) -> TtoN,
Cat(toB, TIP) -> TtoB,
Cat(toT, TIP) -> TtoT
)) // other combinations should miss or have mshr allocated
)
}.otherwise { // req_s3.fromC
sink_resp_s3.bits.opcode := ReleaseAck
sink_resp_s3.bits.param := 0.U // param of ReleaseAck must be 0
}
val source_req_s3 = Wire(new TaskBundle)
source_req_s3 := Mux(sink_resp_s3.valid, sink_resp_s3.bits, req_s3)
source_req_s3.isKeyword.foreach(_ := req_s3.isKeyword.getOrElse(false.B))
/* ======== Interact with DS ======== */
val data_s3 = Mux(io.releaseBufResp_s3.valid, io.releaseBufResp_s3.bits.data, io.refillBufResp_s3.bits.data) // releaseBuf prior
val c_releaseData_s3 = io.bufResp.data.asUInt
val hasData_s3 = source_req_s3.opcode(0)
val need_data_a = dirResult_s3.hit && (req_get_s3 || req_acquireBlock_s3)
val need_data_b = sinkB_req_s3 && dirResult_s3.hit &&
(meta_s3.state === TRUNK || meta_s3.state === TIP && meta_s3.dirty || req_s3.needProbeAckData)
val need_data_mshr_repl = mshr_refill_s3 && need_repl && !retry
val ren = need_data_a || need_data_b || need_data_mshr_repl
val wen_c = sinkC_req_s3 && isParamFromT(req_s3.param) && req_s3.opcode(0) && dirResult_s3.hit
val wen_mshr = req_s3.dsWen && (
mshr_probeack_s3 || mshr_release_s3 ||
mshr_refill_s3 && !need_repl && !retry
)
val wen = wen_c || wen_mshr
// This is to let io.toDS.req_s3.valid hold for 2 cycles (see DataStorage for details)
val task_s3_valid_hold2 = RegInit(0.U(2.W))
when(task_s2.valid) {
task_s3_valid_hold2 := "b11".U
}.otherwise {
task_s3_valid_hold2 := task_s3_valid_hold2 >> 1.U
}
io.toDS.en_s3 := task_s3.valid && (ren || wen)
io.toDS.req_s3.valid := task_s3_valid_hold2(0) && (ren || wen)
io.toDS.req_s3.bits.way := Mux(mshr_refill_s3 && req_s3.replTask, io.replResp.bits.way,
Mux(mshr_req_s3, req_s3.way, dirResult_s3.way))
io.toDS.req_s3.bits.set := Mux(mshr_req_s3, req_s3.set, dirResult_s3.set)
io.toDS.req_s3.bits.wen := wen
io.toDS.wdata_s3.data := Mux(
!mshr_req_s3,
c_releaseData_s3, // Among all sinkTasks, only C-Release writes DS
Mux(
req_s3.useProbeData,
io.releaseBufResp_s3.bits.data,
io.refillBufResp_s3.bits.data
)
)
/* ======== Read DS and store data in Buffer ======== */
// A: need_write_releaseBuf indicates that DS should be read and the data will be written into ReleaseBuffer
// need_write_releaseBuf is assigned true when:
// inner clients' data is needed, but whether the client will ack data is uncertain, so DS data is also needed, or
val need_write_releaseBuf = need_probe_s3_a ||
cache_alias ||
need_data_b && need_mshr_s3_b ||
need_data_mshr_repl
// B: need_write_refillBuf indicates that DS should be read and the data will be written into RefillBuffer
// when L1 AcquireBlock but L2 AcquirePerm to L3, we need to prepare data for L1
// but this will no longer happen, cuz we always AcquireBlock for L1 AcquireBlock
val need_write_refillBuf = false.B
/* ======== Write Directory ======== */
val metaW_valid_s3_a = sinkA_req_s3 && !need_mshr_s3_a && !req_get_s3 && !req_prefetch_s3 // get & prefetch that hit will not write meta
val metaW_valid_s3_b = sinkB_req_s3 && !need_mshr_s3_b && dirResult_s3.hit && (meta_s3.state === TIP || meta_s3.state === BRANCH && req_s3.param === toN)
val metaW_valid_s3_c = sinkC_req_s3 && dirResult_s3.hit
val metaW_valid_s3_mshr = mshr_req_s3 && req_s3.metaWen && !(mshr_refill_s3 && retry)
require(clientBits == 1)
// Get and Prefetch should not change alias bit
val metaW_s3_a_alias = Mux(
req_get_s3 || req_prefetch_s3,
meta_s3.alias.getOrElse(0.U),
req_s3.alias.getOrElse(0.U)
)
val metaW_s3_a = MetaEntry(
dirty = meta_s3.dirty,
state = Mux(req_needT_s3 || sink_resp_s3_a_promoteT, TRUNK, meta_s3.state),
clients = Fill(clientBits, true.B),
alias = Some(metaW_s3_a_alias),
accessed = true.B,
tagErr = meta_s3.tagErr,
dataErr = meta_s3.dataErr
)
val metaW_s3_b = Mux(req_s3.param === toN, MetaEntry(),
MetaEntry(
dirty = false.B,
state = BRANCH,
clients = meta_s3.clients,
alias = meta_s3.alias,
tagErr = meta_s3.tagErr,
dataErr = meta_s3.dataErr
)
)
val metaW_s3_c = MetaEntry(
dirty = meta_s3.dirty || wen_c,
state = Mux(isParamFromT(req_s3.param), TIP, meta_s3.state),
clients = Fill(clientBits, !isToN(req_s3.param)),
alias = meta_s3.alias,
accessed = meta_s3.accessed,
tagErr = Mux(wen_c, req_s3.denied, meta_s3.tagErr),
dataErr = Mux(wen_c, req_s3.corrupt, meta_s3.dataErr) // update error when write DS
)
// use merge_meta if mergeA
val metaW_s3_mshr = WireInit(Mux(req_s3.mergeA, req_s3.aMergeTask.meta, req_s3.meta))
metaW_s3_mshr.tagErr := req_s3.denied
metaW_s3_mshr.dataErr := req_s3.corrupt
val metaW_way = Mux(mshr_refill_s3 && req_s3.replTask, io.replResp.bits.way, // grant always use replResp way
Mux(mshr_req_s3, req_s3.way, dirResult_s3.way))
// dir write signals in s3
val metaWReq_s3 = Wire(Valid(new MetaWrite()))
val tagWReq_s3 = Wire(Valid(new TagWrite()))
metaWReq_s3.valid := !resetFinish || task_s3.valid && (metaW_valid_s3_a || metaW_valid_s3_b || metaW_valid_s3_c || metaW_valid_s3_mshr)
metaWReq_s3.bits.set := Mux(resetFinish, req_s3.set, resetIdx)
metaWReq_s3.bits.wayOH := Mux(resetFinish, UIntToOH(metaW_way), Fill(cacheParams.ways, true.B))
metaWReq_s3.bits.wmeta := Mux(
resetFinish,
ParallelPriorityMux(
Seq(metaW_valid_s3_a, metaW_valid_s3_b, metaW_valid_s3_c, metaW_valid_s3_mshr),
Seq(metaW_s3_a, metaW_s3_b, metaW_s3_c, metaW_s3_mshr)
),
MetaEntry()
)
tagWReq_s3.valid := task_s3.valid && req_s3.tagWen && mshr_refill_s3 && !retry
tagWReq_s3.bits.set := req_s3.set
tagWReq_s3.bits.way := Mux(mshr_refill_s3 && req_s3.replTask, io.replResp.bits.way, req_s3.way)
tagWReq_s3.bits.wtag := req_s3.tag
/* ======== Interact with Channels (C & D) ======== */
// do not need s4 & s5
val chnl_fire_s3 = c_s3.fire || d_s3.fire
val req_drop_s3 = !need_write_releaseBuf && (
!mshr_req_s3 && need_mshr_s3 || chnl_fire_s3
) || (mshr_refill_s3 && retry)
val data_unready_s3 = hasData_s3 && !mshr_req_s3
val isC_s3 = Mux(
mshr_req_s3,
mshr_release_s3 || mshr_probeack_s3,
req_s3.fromB && !need_mshr_s3 && !data_unready_s3
)
val isD_s3 = Mux(
mshr_req_s3,
mshr_refill_s3 && !retry,
req_s3.fromC || req_s3.fromA && !need_mshr_s3 && !data_unready_s3 && req_s3.opcode =/= Hint
) // prefetch-hit will not generate response
c_s3.valid := task_s3.valid && isC_s3
d_s3.valid := task_s3.valid && isD_s3
c_s3.bits.task := source_req_s3
c_s3.bits.data.data := data_s3
d_s3.bits.task := source_req_s3
d_s3.bits.data.data := data_s3
/* ======== nested & prefetch ======== */
io.nestedwb.set := req_s3.set
io.nestedwb.tag := req_s3.tag
// This serves as VALID signal
// c_set_dirty is true iff Release has Data
io.nestedwb.c_set_dirty := task_s3.valid && task_s3.bits.fromC && task_s3.bits.opcode === ReleaseData
io.nestedwb.c_set_tip := false.B
io.nestedwb.b_inv_dirty := false.B
io.nestedwbData := c_releaseData_s3.asTypeOf(new DSBlock)
io.prefetchTrain.foreach {
train =>
// train on request(with needHint flag) miss or hit on prefetched block
// trigger train also in a_merge here
train.valid := task_s3.valid && (((req_acquire_s3 || req_get_s3) && req_s3.needHint.getOrElse(false.B) &&
(!dirResult_s3.hit || meta_s3.prefetch.get)) || req_s3.mergeA)
train.bits.tag := req_s3.tag
train.bits.set := req_s3.set
train.bits.needT := Mux(req_s3.mergeA, needT(req_s3.aMergeTask.opcode, req_s3.aMergeTask.param),req_needT_s3)
train.bits.source := Mux(req_s3.mergeA, req_s3.aMergeTask.sourceId, req_s3.sourceId)
train.bits.vaddr.foreach(_ := Mux(req_s3.mergeA, req_s3.aMergeTask.vaddr.getOrElse(0.U), req_s3.vaddr.getOrElse(0.U)))
train.bits.hit := Mux(req_s3.mergeA, true.B, dirResult_s3.hit)
train.bits.prefetched := Mux(req_s3.mergeA, true.B, meta_s3.prefetch.getOrElse(false.B))
train.bits.pfsource := meta_s3.prefetchSrc.getOrElse(PfSource.NoWhere.id.U) // TODO
train.bits.reqsource := req_s3.reqSource
}
/* ======== Stage 4 ======== */
val task_s4 = RegInit(0.U.asTypeOf(Valid(new TaskBundle())))
val taskWDir_s4 = RegInit(0.U.asTypeOf(Valid(new TaskBundle())))
val metaWReq_s4 = RegInit(0.U.asTypeOf(Valid(new MetaWrite())))
val tagWReq_s4 = RegInit(0.U.asTypeOf(Valid(new TagWrite())))
val data_unready_s4 = RegInit(false.B)
val data_s4 = Reg(UInt((blockBytes * 8).W))
val ren_s4 = RegInit(false.B)
val need_write_releaseBuf_s4 = RegInit(false.B)
val isC_s4, isD_s4 = RegInit(false.B)
val tagError_s4 = RegInit(false.B)
val dataError_s4 = RegInit(false.B)
val l2Error_s4 = RegInit(false.B)
task_s4.valid := task_s3.valid && !req_drop_s3
when (task_s3.valid && !req_drop_s3) {
task_s4.bits := source_req_s3
task_s4.bits.mshrId := Mux(!task_s3.bits.mshrTask && need_mshr_s3, io.fromMSHRCtl.mshr_alloc_ptr, source_req_s3.mshrId)
// task_s4.bits.isKeyword.foreach(_ :=source_req_s3.isKeyword.getOrElse(false.B))
data_unready_s4 := data_unready_s3
data_s4 := data_s3
ren_s4 := ren
need_write_releaseBuf_s4 := need_write_releaseBuf
isC_s4 := isC_s3
isD_s4 := isD_s3
tagError_s4 := tagError_s3
dataError_s4 := dataError_s3
l2Error_s4 := l2Error_s3
}
taskWDir_s4.valid := task_s3.valid && (metaWReq_s3.valid || tagWReq_s3.valid)
when (task_s3.valid || !resetFinish) {
taskWDir_s4.bits := source_req_s3
metaWReq_s4 := metaWReq_s3
tagWReq_s4 := tagWReq_s3
}
io.metaWReq.valid := metaWReq_s4.valid && (taskWDir_s4.valid || RegNext(!resetFinish))
io.metaWReq.bits := metaWReq_s4.bits
io.tagWReq.valid := tagWReq_s4.valid && (taskWDir_s4.valid || RegNext(!resetFinish))
io.tagWReq.bits := tagWReq_s4.bits
// A-alias-Acquire should send neither C nor D
// val isC_s4 = task_s4.bits.opcode(2, 1) === Release(2, 1) && task_s4.bits.fromA && !RegNext(cache_alias, false.B) ||
// task_s4.bits.opcode(2, 1) === ProbeAck(2, 1) && task_s4.bits.fromB
// val isD_s4 = task_s4.bits.fromC || task_s4.bits.fromA && (
// task_s4.bits.opcode(2, 1) === Grant(2, 1) ||
// task_s4.bits.opcode(2, 1) === AccessAck(2, 1) ||
// task_s4.bits.opcode === HintAck)
// for reqs that CANNOT give response in MainPipe, but needs to write releaseBuf/refillBuf
// we cannot drop them at s3, we must let them go to s4/s5
val chnl_fire_s4 = c_s4.fire || d_s4.fire
val req_drop_s4 = !need_write_releaseBuf_s4 && chnl_fire_s4
val c_d_valid_s4 = task_s4.valid && !RegNext(chnl_fire_s3, false.B)
c_s4.valid := c_d_valid_s4 && isC_s4
d_s4.valid := c_d_valid_s4 && isD_s4
c_s4.bits.task := task_s4.bits
c_s4.bits.data.data := data_s4
d_s4.bits.task := task_s4.bits
d_s4.bits.task.isKeyword.foreach(_ := task_s4.bits.isKeyword.getOrElse(false.B))
d_s4.bits.data.data := data_s4
/* ======== Stage 5 ======== */
val task_s5 = RegInit(0.U.asTypeOf(Valid(new TaskBundle())))
val ren_s5 = RegInit(false.B)
val data_s5 = Reg(UInt((blockBytes * 8).W))
val need_write_releaseBuf_s5 = RegInit(false.B)
val isC_s5, isD_s5 = RegInit(false.B)
val tagError_s5 = RegInit(false.B)
val dataMetaError_s5 = RegInit(false.B)
val l2TagError_s5 = RegInit(false.B)
// those hit@s3 and ready to fire@s5, and Now wait@s4
val pendingC_s4 = task_s4.bits.fromB && !task_s4.bits.mshrTask && task_s4.bits.opcode === ProbeAckData
val pendingD_s4 = task_s4.bits.fromA && !task_s4.bits.mshrTask &&
(task_s4.bits.opcode === GrantData || task_s4.bits.opcode === AccessAckData)
task_s5.valid := task_s4.valid && !req_drop_s4
when (task_s4.valid && !req_drop_s4) {
task_s5.bits := task_s4.bits
task_s5.bits.isKeyword.foreach(_ :=task_s4.bits.isKeyword.getOrElse(false.B))
ren_s5 := ren_s4
data_s5 := data_s4
need_write_releaseBuf_s5 := need_write_releaseBuf_s4
// except for those ready at s3/s4 (isC/D_s4), sink resps are also ready to fire at s5
isC_s5 := isC_s4 || pendingC_s4
isD_s5 := isD_s4 || pendingD_s4
tagError_s5 := tagError_s4
dataMetaError_s5 := dataError_s4
l2TagError_s5 := l2Error_s4
}
val rdata_s5 = io.toDS.rdata_s5.data
val dataError_s5 = io.toDS.error_s5 || dataMetaError_s5
val l2Error_s5 = l2TagError_s5 || io.toDS.error_s5
val out_data_s5 = Mux(!task_s5.bits.mshrTask, rdata_s5, data_s5)
val chnl_fire_s5 = c_s5.fire || d_s5.fire
val customL1Hint = Module(new CustomL1Hint)
customL1Hint.io.s1 := io.taskInfo_s1
customL1Hint.io.s3.task := task_s3
// overwrite opcode: if sinkReq can respond, use sink_resp_s3.bits.opcode = Grant/GrantData
customL1Hint.io.s3.task.bits.opcode := Mux(sink_resp_s3.valid, sink_resp_s3.bits.opcode, task_s3.bits.opcode)
customL1Hint.io.s3.need_mshr := need_mshr_s3
customL1Hint.io.l1Hint <> io.l1Hint
io.releaseBufWrite.valid := task_s5.valid && need_write_releaseBuf_s5
io.releaseBufWrite.bits.id := task_s5.bits.mshrId
io.releaseBufWrite.bits.data.data := rdata_s5
io.releaseBufWrite.bits.beatMask := Fill(beatSize, true.B)
val c_d_valid_s5 = task_s5.valid && !RegNext(chnl_fire_s4, false.B) && !RegNextN(chnl_fire_s3, 2, Some(false.B))
c_s5.valid := c_d_valid_s5 && isC_s5
d_s5.valid := c_d_valid_s5 && isD_s5
c_s5.bits.task := task_s5.bits
c_s5.bits.task.denied := Mux(!task_s5.bits.mshrTask, tagError_s5, task_s5.bits.denied)
c_s5.bits.task.corrupt := Mux(!task_s5.bits.mshrTask, dataError_s5, task_s5.bits.corrupt)
c_s5.bits.data.data := out_data_s5
d_s5.bits.task := task_s5.bits
d_s5.bits.task.denied := Mux(!task_s5.bits.mshrTask, tagError_s5, task_s5.bits.denied)
d_s5.bits.task.corrupt := Mux(!task_s5.bits.mshrTask, dataError_s5, task_s5.bits.corrupt)
d_s5.bits.data.data := out_data_s5
/* ======== BlockInfo ======== */
// if s2/s3 might write Dir, we must block s1 sink entrance
// TODO:[Check] it seems that s3 Dir write will naturally block all s1 by dirRead.ready
// (an even stronger blocking than set blocking)
// so we might not need s3 blocking here
def s23Block(chn: Char, s: TaskBundle): Bool = {
val s1 = io.fromReqArb.status_s1
val s1_set = chn match {
case 'a' => s1.a_set
case 'b' => s1.b_set
case 'c' => s1.c_set
case 'g' => s1.g_set
}
s.set === s1_set && !(s.mshrTask && !s.metaWen) // if guaranteed not to write meta, no blocking needed
}
def bBlock(s: TaskBundle, tag: Boolean = false): Bool = {
val s1 = io.fromReqArb.status_s1
// tag true: compare tag + set
s.set === s1.b_set && (if(tag) s.tag === s1.b_tag else true.B)
}
io.toReqBuf(0) := task_s2.valid && s23Block('a', task_s2.bits)
io.toReqBuf(1) := task_s3.valid && s23Block('a', task_s3.bits)
io.toReqArb.blockC_s1 := task_s2.valid && s23Block('c', task_s2.bits) ||
task_s3.valid && s23Block('c', task_s3.bits) && metaWReq_s3.valid
io.toReqArb.blockB_s1 :=
task_s2.valid && bBlock(task_s2.bits) ||
task_s3.valid && bBlock(task_s3.bits) ||
task_s4.valid && bBlock(task_s4.bits, tag = true) ||
taskWDir_s4.valid && bBlock(taskWDir_s4.bits, tag = true) ||
task_s5.valid && bBlock(task_s5.bits, tag = true)
io.toReqArb.blockA_s1 := false.B
io.toReqArb.blockG_s1 := task_s2.valid && s23Block('g', task_s2.bits) ||
task_s3.valid && s23Block('g', task_s3.bits) && metaWReq_s3.valid
/* ======== Pipeline Status ======== */
require(io.status_vec_toD.size == 3)
io.status_vec_toD(0).valid := task_s3.valid && Mux(
mshr_req_s3,
mshr_refill_s3 && !retry,
true.B
// TODO:
// To consider grantBuffer capacity conflict, only " req_s3.fromC || req_s3.fromA && !need_mshr_s3 " is needed
// But to consider mshrFull, all channel_reqs are needed
// so maybe it is excessive for grantBuf capacity conflict
)
io.status_vec_toD(0).bits.channel := task_s3.bits.channel
io.status_vec_toD(1).valid := task_s4.valid && (isD_s4 || pendingD_s4)
io.status_vec_toD(1).bits.channel := task_s4.bits.channel
io.status_vec_toD(2).valid := d_s5.valid
io.status_vec_toD(2).bits.channel := task_s5.bits.channel
require(io.status_vec_toC.size == 3)
io.status_vec_toC(0).valid := task_s3.valid && Mux(mshr_req_s3, mshr_release_s3 || mshr_probeack_s3, true.B)
io.status_vec_toC(0).bits.channel := task_s3.bits.channel
io.status_vec_toC(1).valid := task_s4.valid && (isC_s4 || pendingC_s4)
io.status_vec_toC(1).bits.channel := task_s4.bits.channel
io.status_vec_toC(2).valid := c_s5.valid
io.status_vec_toC(2).bits.channel := task_s5.bits.channel
/* ======== Other Signals Assignment ======== */
// Initial state assignment
// ! Caution: s_ and w_ are false-as-valid
when(req_s3.fromA) {
alloc_state.s_refill := false.B
alloc_state.w_replResp := dirResult_s3.hit // need replRead when NOT dirHit
// need Acquire downwards
when(need_acquire_s3_a) {
alloc_state.s_acquire := false.B
alloc_state.w_grantfirst := false.B
alloc_state.w_grantlast := false.B
alloc_state.w_grant := false.B
}
// need Probe for alias
// need Probe when Get hits on a TRUNK block
when(cache_alias || need_probe_s3_a) {
alloc_state.s_rprobe := false.B
alloc_state.w_rprobeackfirst := false.B
alloc_state.w_rprobeacklast := false.B
}
// need trigger a prefetch, send PrefetchTrain msg to Prefetcher
// prefetchOpt.foreach {_ =>
// when (req_s3.fromA && req_s3.needHint.getOrElse(false.B) && (!dirResult_s3.hit || meta_s3.prefetch.get)) {
// alloc_state.s_triggerprefetch.foreach(_ := false.B)
// }
// }
}
when(req_s3.fromB) {
// Only consider the situation when mshr needs to be allocated
alloc_state.s_pprobe := false.B
alloc_state.w_pprobeackfirst := false.B
alloc_state.w_pprobeacklast := false.B
alloc_state.s_probeack := false.B
}
val c = Seq(c_s5, c_s4, c_s3)
val d = Seq(d_s5, d_s4, d_s3)
val c_arb = Module(new Arbiter(io.toSourceC.bits.cloneType, c.size))
val d_arb = Module(new Arbiter(io.toSourceD.bits.cloneType, d.size))
c_arb.io.in <> c
d_arb.io.in <> d
io.toSourceC <> c_arb.io.out
io.toSourceD <> d_arb.io.out
io.error.valid := task_s5.valid
io.error.bits.valid := l2Error_s5 // if not enableECC, should be false
io.error.bits.address := Cat(task_s5.bits.tag, task_s5.bits.set, task_s5.bits.off)
/* ===== Performance counters ===== */
// num of mshr req
XSPerfAccumulate("mshr_grant_req", task_s3.valid && mshr_grant_s3 && !retry)
XSPerfAccumulate("mshr_grantdata_req", task_s3.valid && mshr_grantdata_s3 && !retry)
XSPerfAccumulate("mshr_accessackdata_req", task_s3.valid && mshr_accessackdata_s3 && !retry)
XSPerfAccumulate("mshr_hintack_req", task_s3.valid && mshr_hintack_s3 && !retry)
XSPerfAccumulate("mshr_probeack_req", task_s3.valid && mshr_probeack_s3)
XSPerfAccumulate("mshr_probeackdata_req", task_s3.valid && mshr_probeackdata_s3)
XSPerfAccumulate("mshr_release_req", task_s3.valid && mshr_release_s3)
// directory access result
val hit_s3 = task_s3.valid && !mshr_req_s3 && dirResult_s3.hit
val miss_s3 = task_s3.valid && !mshr_req_s3 && !dirResult_s3.hit
XSPerfAccumulate("a_req_hit", hit_s3 && req_s3.fromA)
XSPerfAccumulate("acquire_hit", hit_s3 && req_s3.fromA &&
(req_s3.opcode === AcquireBlock || req_s3.opcode === AcquirePerm))
XSPerfAccumulate("get_hit", hit_s3 && req_s3.fromA && req_s3.opcode === Get)
XSPerfAccumulate("retry", mshr_refill_s3 && retry)
XSPerfAccumulate("a_req_miss", miss_s3 && req_s3.fromA)
XSPerfAccumulate("acquire_miss", miss_s3 && req_s3.fromA &&
(req_s3.opcode === AcquireBlock || req_s3.opcode === AcquirePerm))
XSPerfAccumulate("get_miss", miss_s3 && req_s3.fromA && req_s3.opcode === Get)
XSPerfAccumulate("b_req_hit", hit_s3 && req_s3.fromB)
XSPerfAccumulate("b_req_miss", miss_s3 && req_s3.fromB)
XSPerfHistogram("a_req_access_way", perfCnt = dirResult_s3.way,
enable = task_s3.valid && !mshr_req_s3 && req_s3.fromA, start = 0, stop = cacheParams.ways, step = 1)
XSPerfHistogram("a_req_hit_way", perfCnt = dirResult_s3.way,
enable = hit_s3 && req_s3.fromA, start = 0, stop = cacheParams.ways, step = 1)
XSPerfHistogram("a_req_miss_way_choice", perfCnt = dirResult_s3.way,
enable = miss_s3 && req_s3.fromA, start = 0, stop = cacheParams.ways, step = 1)
// pipeline stages for sourceC and sourceD reqs
val sourceC_pipe_len = ParallelMux(Seq(
c_s5.fire -> 5.U,
c_s4.fire -> 4.U,
c_s3.fire -> 3.U
))
val sourceD_pipe_len = ParallelMux(Seq(
d_s5.fire -> 5.U,
d_s4.fire -> 4.U,
d_s3.fire -> 3.U
))
XSPerfHistogram("sourceC_pipeline_stages", sourceC_pipe_len,
enable = io.toSourceC.fire, start = 3, stop = 5+1, step = 1)
XSPerfHistogram("sourceD_pipeline_stages", sourceD_pipe_len,
enable = io.toSourceD.fire, start = 3, stop = 5+1, step = 1)
// XSPerfAccumulate("a_req_tigger_prefetch", io.prefetchTrain.)
prefetchOpt.foreach {
_ =>
XSPerfAccumulate("a_req_trigger_prefetch", io.prefetchTrain.get.fire)
XSPerfAccumulate("a_req_trigger_prefetch_not_ready", io.prefetchTrain.get.valid && !io.prefetchTrain.get.ready)
XSPerfAccumulate("acquire_trigger_prefetch_on_miss", io.prefetchTrain.get.fire && req_acquire_s3 && !dirResult_s3.hit)
XSPerfAccumulate("acquire_trigger_prefetch_on_hit_pft", io.prefetchTrain.get.fire && req_acquire_s3 && dirResult_s3.hit && meta_s3.prefetch.get)
XSPerfAccumulate("release_all", mshr_release_s3)
XSPerfAccumulate("release_prefetch_accessed", mshr_release_s3 && meta_s3.prefetch.get && meta_s3.accessed)
XSPerfAccumulate("release_prefetch_not_accessed", mshr_release_s3 && meta_s3.prefetch.get && !meta_s3.accessed)
XSPerfAccumulate("get_trigger_prefetch_on_miss", io.prefetchTrain.get.fire && req_get_s3 && !dirResult_s3.hit)
XSPerfAccumulate("get_trigger_prefetch_on_hit_pft", io.prefetchTrain.get.fire && req_get_s3 && dirResult_s3.hit && meta_s3.prefetch.get)
}
XSPerfAccumulate("early_prefetch", meta_s3.prefetch.getOrElse(false.B) && !meta_s3.accessed && !dirResult_s3.hit && task_s3.valid)
/* ===== Monitor ===== */
io.toMonitor.task_s2 := task_s2
io.toMonitor.task_s3 := task_s3
io.toMonitor.task_s4 := task_s4
io.toMonitor.task_s5 := task_s5
io.toMonitor.dirResult_s3 := dirResult_s3
io.toMonitor.allocMSHR_s3.valid := io.toMSHRCtl.mshr_alloc_s3.valid
io.toMonitor.allocMSHR_s3.bits := io.fromMSHRCtl.mshr_alloc_ptr
io.toMonitor.metaW_s3 := io.metaWReq
/* ===== Hardware Performance Monitor ===== */
val perfEvents = Seq(
("l2_cache_access", task_s3.valid && (sinkA_req_s3 && !req_prefetch_s3 || sinkC_req_s3)),
("l2_cache_l2wb", task_s3.valid && (mshr_releasedata_s3 || mshr_probeackdata_s3)),
("l2_cache_l1wb", task_s3.valid && sinkC_req_s3 && (req_s3.opcode === ReleaseData)),
("l2_cache_wb_victim", task_s3.valid && mshr_releasedata_s3),
("l2_cache_wb_cleaning_coh", task_s3.valid && mshr_probeackdata_s3),
("l2_cache_access_rd", task_s3.valid && sinkA_req_s3 && !req_prefetch_s3),
("l2_cache_access_wr", task_s3.valid && sinkC_req_s3),
("l2_cache_inv", task_s3.valid && sinkB_req_s3 && (req_s3.param === toN))
)
generatePerfEvent()
}