|
|
|
|
@ -10,6 +10,8 @@
|
|
|
|
|
#include "video_core/memory_manager.h"
|
|
|
|
|
#include "video_core/rasterizer_interface.h"
|
|
|
|
|
#include "video_core/renderer_base.h"
|
|
|
|
|
#include "video_core/textures/convert.h"
|
|
|
|
|
#include "video_core/textures/decoders.h"
|
|
|
|
|
|
|
|
|
|
namespace Tegra::Engines {
|
|
|
|
|
|
|
|
|
|
@ -27,30 +29,40 @@ void KeplerMemory::CallMethod(const GPU::MethodCall& method_call) {
|
|
|
|
|
|
|
|
|
|
switch (method_call.method) {
|
|
|
|
|
case KEPLERMEMORY_REG_INDEX(exec): {
|
|
|
|
|
state.write_offset = 0;
|
|
|
|
|
ProcessExec();
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
case KEPLERMEMORY_REG_INDEX(data): {
|
|
|
|
|
ProcessData(method_call.argument);
|
|
|
|
|
ProcessData(method_call.argument, method_call.IsLastCall());
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void KeplerMemory::ProcessData(u32 data) {
|
|
|
|
|
ASSERT_MSG(regs.exec.linear, "Non-linear uploads are not supported");
|
|
|
|
|
ASSERT(regs.dest.x == 0 && regs.dest.y == 0 && regs.dest.z == 0);
|
|
|
|
|
|
|
|
|
|
// We have to invalidate the destination region to evict any outdated surfaces from the cache.
|
|
|
|
|
// We do this before actually writing the new data because the destination address might
|
|
|
|
|
// contain a dirty surface that will have to be written back to memory.
|
|
|
|
|
const GPUVAddr address{regs.dest.Address() + state.write_offset * sizeof(u32)};
|
|
|
|
|
rasterizer.InvalidateRegion(ToCacheAddr(memory_manager.GetPointer(address)), sizeof(u32));
|
|
|
|
|
memory_manager.Write<u32>(address, data);
|
|
|
|
|
void KeplerMemory::ProcessExec() {
|
|
|
|
|
state.write_offset = 0;
|
|
|
|
|
state.copy_size = regs.line_length_in * regs.line_count;
|
|
|
|
|
state.inner_buffer.resize(state.copy_size);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
system.GPU().Maxwell3D().dirty_flags.OnMemoryWrite();
|
|
|
|
|
void KeplerMemory::ProcessData(u32 data, bool is_last_call) {
|
|
|
|
|
const u32 sub_copy_size = std::min(4U, state.copy_size - state.write_offset);
|
|
|
|
|
std::memcpy(&state.inner_buffer[state.write_offset], &data, sub_copy_size);
|
|
|
|
|
state.write_offset += sub_copy_size;
|
|
|
|
|
if (is_last_call) {
|
|
|
|
|
UNIMPLEMENTED_IF_MSG(regs.exec.linear == 0, "Block Linear Copy is not implemented");
|
|
|
|
|
if (regs.exec.linear != 0) {
|
|
|
|
|
const GPUVAddr address{regs.dest.Address()};
|
|
|
|
|
const auto host_ptr = memory_manager.GetPointer(address);
|
|
|
|
|
// We have to invalidate the destination region to evict any outdated surfaces from the
|
|
|
|
|
// cache. We do this before actually writing the new data because the destination
|
|
|
|
|
// address might contain a dirty surface that will have to be written back to memory.
|
|
|
|
|
|
|
|
|
|
state.write_offset++;
|
|
|
|
|
rasterizer.InvalidateRegion(ToCacheAddr(host_ptr), state.copy_size);
|
|
|
|
|
std::memcpy(host_ptr, state.inner_buffer.data(), state.copy_size);
|
|
|
|
|
system.GPU().Maxwell3D().dirty_flags.OnMemoryWrite();
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
} // namespace Tegra::Engines
|
|
|
|
|
|