blob: 32d267c7a3638695b25265093da21b4ef21d30cc [file] [log] [blame]
/*
* Copyright 2023 Google, Inc
* All rights reserved
*
* The license below extends only to copyright in the software and shall
* not be construed as granting a license to any other intellectual
* property including but not limited to intellectual property relating
* to a hardware implementation of the functionality of the software
* licensed hereunder. You may use the software subject to the license
* terms below provided that you ensure that this notice is replicated
* unmodified and in its entirety in all distributions of the software,
* modified or unmodified, in source code or in binary form.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are
* met: redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer;
* redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution;
* neither the name of the copyright holders nor the names of its
* contributors may be used to endorse or promote products derived from
* this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include <utility>
#include "base/logging.hh"
#include "mem/backdoor_manager.hh"
namespace gem5
{
BackdoorManager::BackdoorManager(const std::vector<AddrRange> &original_ranges,
const std::vector<AddrRange> &remapped_ranges)
: originalRanges(original_ranges),
remappedRanges(remapped_ranges),
backdoorLists(original_ranges.size())
{
}
MemBackdoorPtr
BackdoorManager::getRevertedBackdoor(MemBackdoorPtr backdoor,
const AddrRange &pkt_range)
{
MemBackdoorPtr reverted_backdoor = findBackdoor(pkt_range);
if (reverted_backdoor == nullptr) {
reverted_backdoor = createRevertedBackdoor(backdoor, pkt_range);
}
return reverted_backdoor;
}
MemBackdoorPtr
BackdoorManager::createRevertedBackdoor(MemBackdoorPtr backdoor,
const AddrRange &pkt_range)
{
std::unique_ptr<MemBackdoor> reverted_backdoor = std::make_unique<MemBackdoor>();
reverted_backdoor->flags(backdoor->flags());
reverted_backdoor->ptr(backdoor->ptr());
Addr addr = pkt_range.start();
for (int i = 0; i < originalRanges.size(); ++i) {
if (originalRanges[i].contains(addr)) {
/** Does not support interleaved range backdoors. */
if (originalRanges[i].interleaved() ||
remappedRanges[i].interleaved()) {
return nullptr;
}
/** Shrink the backdoor to fit inside address range. */
AddrRange shrinked_backdoor_range =
backdoor->range() & remappedRanges[i];
Addr backdoor_offset =
shrinked_backdoor_range.start() - remappedRanges[i].start();
Addr backdoor_size = shrinked_backdoor_range.size();
/** Create the backdoor in original address view. */
reverted_backdoor->range(AddrRange(
originalRanges[i].start() + backdoor_offset,
originalRanges[i].start() + backdoor_offset + backdoor_size));
/**
* The backdoor pointer also needs to be shrinked to point to the
* beginning of the range.
*/
Addr shrinked_offset =
shrinked_backdoor_range.start() - backdoor->range().start();
reverted_backdoor->ptr(backdoor->ptr() + shrinked_offset);
/**
* Bind the life cycle of the created backdoor with the target
* backdoor. Invalid and delete the created backdoor when the
* target backdoor is invalidated.
*/
MemBackdoorPtr reverted_backdoor_raw_ptr = reverted_backdoor.get();
auto it = backdoorLists[i].insert(backdoorLists[i].end(),
std::move(reverted_backdoor));
backdoor->addInvalidationCallback(
[this, i, it](const MemBackdoor &backdoor) {
(*it)->invalidate(); // *it is unique_ptr reverted_backdoor
this->backdoorLists[i].erase(it);
});
return reverted_backdoor_raw_ptr;
}
}
// Backdoor is not valid. Return an empty one.
panic("Target does not provide valid backdoor.");
}
MemBackdoorPtr
BackdoorManager::findBackdoor(const AddrRange &pkt_range) const
{
Addr addr = pkt_range.start();
Addr size = pkt_range.size();
for (int i = 0; i < originalRanges.size(); ++i) {
/** The original ranges should be disjoint, so at most one range
* contains the begin address.
*/
if (originalRanges[i].contains(addr)) {
if (!originalRanges[i].contains(addr + size - 1)) {
/** The request range doesn't fit in any address range. */
return nullptr;
}
for (const auto &backdoor : backdoorLists[i]) {
if (backdoor->range().contains(addr) &&
backdoor->range().contains(addr + size - 1)) {
return backdoor.get();
}
}
}
}
return nullptr;
}
} // namespace gem5