mirror of
https://github.com/vortexgpgpu/vortex.git
synced 2025-06-28 17:43:24 -04:00
164 lines
No EOL
4.7 KiB
C++
164 lines
No EOL
4.7 KiB
C++
// Copyright © 2019-2023
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
#include "cluster.h"
|
|
|
|
using namespace vortex;
|
|
|
|
Cluster::Cluster(const SimContext& ctx,
|
|
uint32_t cluster_id,
|
|
ProcessorImpl* processor,
|
|
const Arch &arch,
|
|
const DCRS &dcrs)
|
|
: SimObject(ctx, "cluster")
|
|
, mem_req_port(this)
|
|
, mem_rsp_port(this)
|
|
, cluster_id_(cluster_id)
|
|
, processor_(processor)
|
|
, sockets_(NUM_SOCKETS)
|
|
, barriers_(arch.num_barriers(), 0)
|
|
, cores_per_socket_(arch.socket_size())
|
|
{
|
|
char sname[100];
|
|
|
|
uint32_t sockets_per_cluster = sockets_.size();
|
|
|
|
// create sockets
|
|
|
|
snprintf(sname, 100, "cluster%d-icache-arb", cluster_id);
|
|
auto icache_switch = MemSwitch::Create(sname, ArbiterType::RoundRobin, sockets_per_cluster);
|
|
|
|
snprintf(sname, 100, "cluster%d-dcache-arb", cluster_id);
|
|
auto dcache_switch = MemSwitch::Create(sname, ArbiterType::RoundRobin, sockets_per_cluster);
|
|
|
|
for (uint32_t i = 0; i < sockets_per_cluster; ++i) {
|
|
uint32_t socket_id = cluster_id * sockets_per_cluster + i;
|
|
auto socket = Socket::Create(socket_id,
|
|
this,
|
|
arch,
|
|
dcrs);
|
|
|
|
socket->icache_mem_req_port.bind(&icache_switch->ReqIn.at(i));
|
|
icache_switch->RspIn.at(i).bind(&socket->icache_mem_rsp_port);
|
|
|
|
socket->dcache_mem_req_port.bind(&dcache_switch->ReqIn.at(i));
|
|
dcache_switch->RspIn.at(i).bind(&socket->dcache_mem_rsp_port);
|
|
|
|
sockets_.at(i) = socket;
|
|
}
|
|
|
|
// Create l2cache
|
|
|
|
snprintf(sname, 100, "cluster%d-l2cache", cluster_id);
|
|
l2cache_ = CacheSim::Create(sname, CacheSim::Config{
|
|
!L2_ENABLED,
|
|
log2ceil(L2_CACHE_SIZE),// C
|
|
log2ceil(MEM_BLOCK_SIZE),// L
|
|
log2ceil(L1_LINE_SIZE), // W
|
|
log2ceil(L2_NUM_WAYS), // A
|
|
log2ceil(L2_NUM_BANKS), // B
|
|
XLEN, // address bits
|
|
1, // number of ports
|
|
2, // request size
|
|
L2_WRITEBACK, // write-back
|
|
false, // write response
|
|
L2_MSHR_SIZE, // mshr size
|
|
2, // pipeline latency
|
|
});
|
|
|
|
l2cache_->MemReqPorts.at(0).bind(&this->mem_req_port);
|
|
this->mem_rsp_port.bind(&l2cache_->MemRspPorts.at(0));
|
|
|
|
icache_switch->ReqOut.at(0).bind(&l2cache_->CoreReqPorts.at(0));
|
|
l2cache_->CoreRspPorts.at(0).bind(&icache_switch->RspOut.at(0));
|
|
|
|
dcache_switch->ReqOut.at(0).bind(&l2cache_->CoreReqPorts.at(1));
|
|
l2cache_->CoreRspPorts.at(1).bind(&dcache_switch->RspOut.at(0));
|
|
}
|
|
|
|
Cluster::~Cluster() {
|
|
//--
|
|
}
|
|
|
|
void Cluster::reset() {
|
|
for (auto& barrier : barriers_) {
|
|
barrier.reset();
|
|
}
|
|
}
|
|
|
|
void Cluster::tick() {
|
|
//--
|
|
}
|
|
|
|
void Cluster::attach_ram(RAM* ram) {
|
|
for (auto& socket : sockets_) {
|
|
socket->attach_ram(ram);
|
|
}
|
|
}
|
|
|
|
#ifdef VM_ENABLE
|
|
void Cluster::set_satp(uint64_t satp) {
|
|
for (auto& socket : sockets_) {
|
|
socket->set_satp(satp);
|
|
}
|
|
}
|
|
#endif
|
|
|
|
bool Cluster::running() const {
|
|
for (auto& socket : sockets_) {
|
|
if (socket->running())
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
int Cluster::get_exitcode() const {
|
|
int exitcode = 0;
|
|
for (auto& socket : sockets_) {
|
|
exitcode |= socket->get_exitcode();
|
|
}
|
|
return exitcode;
|
|
}
|
|
|
|
void Cluster::barrier(uint32_t bar_id, uint32_t count, uint32_t core_id) {
|
|
auto& barrier = barriers_.at(bar_id);
|
|
|
|
auto sockets_per_cluster = sockets_.size();
|
|
auto cores_per_socket = cores_per_socket_;
|
|
|
|
uint32_t cores_per_cluster = sockets_per_cluster * cores_per_socket;
|
|
uint32_t local_core_id = core_id % cores_per_cluster;
|
|
barrier.set(local_core_id);
|
|
|
|
DP(3, "*** Suspend core #" << core_id << " at barrier #" << bar_id);
|
|
|
|
if (barrier.count() == (size_t)count) {
|
|
// resume all suspended cores
|
|
for (uint32_t s = 0; s < sockets_per_cluster; ++s) {
|
|
for (uint32_t c = 0; c < cores_per_socket; ++c) {
|
|
uint32_t i = s * cores_per_socket + c;
|
|
if (barrier.test(i)) {
|
|
DP(3, "*** Resume core #" << i << " at barrier #" << bar_id);
|
|
sockets_.at(s)->resume(c);
|
|
}
|
|
}
|
|
}
|
|
barrier.reset();
|
|
}
|
|
}
|
|
|
|
Cluster::PerfStats Cluster::perf_stats() const {
|
|
PerfStats perf_stats;
|
|
perf_stats.l2cache = l2cache_->perf_stats();
|
|
return perf_stats;
|
|
} |