mirror of
https://github.com/ml-explore/mlx.git
synced 2025-06-24 01:17:26 +08:00
787 lines
23 KiB
C++
787 lines
23 KiB
C++
// Copyright © 2023-2024 Apple Inc.
|
|
|
|
#include <cstdlib>
|
|
#include <filesystem>
|
|
#include <sstream>
|
|
|
|
#define NS_PRIVATE_IMPLEMENTATION
|
|
#define CA_PRIVATE_IMPLEMENTATION
|
|
#define MTL_PRIVATE_IMPLEMENTATION
|
|
|
|
#include "mlx/backend/metal/device.h"
|
|
#include "mlx/backend/metal/metal.h"
|
|
#include "mlx/backend/metal/utils.h"
|
|
#include "mlx/utils.h"
|
|
|
|
namespace fs = std::filesystem;
|
|
|
|
namespace mlx::core::metal {
|
|
|
|
namespace {
|
|
|
|
constexpr const char* default_mtllib_path = METAL_PATH;
|
|
|
|
auto get_metal_version() {
|
|
auto get_metal_version_ = []() {
|
|
if (__builtin_available(macOS 15, iOS 18, tvOS 18, visionOS 2, *)) {
|
|
return MTL::LanguageVersion3_2;
|
|
} else if (__builtin_available(macOS 14, iOS 17, tvOS 17, visionOS 1, *)) {
|
|
return MTL::LanguageVersion3_1;
|
|
} else {
|
|
return MTL::LanguageVersion3_0;
|
|
}
|
|
};
|
|
static auto metal_version_ = get_metal_version_();
|
|
return metal_version_;
|
|
}
|
|
|
|
auto load_device() {
|
|
auto devices = MTL::CopyAllDevices();
|
|
auto device = static_cast<MTL::Device*>(devices->object(0))
|
|
?: MTL::CreateSystemDefaultDevice();
|
|
if (!device) {
|
|
throw std::runtime_error("Failed to load device");
|
|
}
|
|
return device;
|
|
}
|
|
std::pair<MTL::Library*, NS::Error*> load_library_from_path(
|
|
MTL::Device* device,
|
|
const char* path) {
|
|
auto library = NS::String::string(path, NS::UTF8StringEncoding);
|
|
NS::Error* error;
|
|
auto lib = device->newLibrary(library, &error);
|
|
|
|
return std::make_pair(lib, error);
|
|
}
|
|
|
|
#ifdef SWIFTPM_BUNDLE
|
|
MTL::Library* try_load_bundle(
|
|
MTL::Device* device,
|
|
NS::URL* url,
|
|
const std::string& lib_name) {
|
|
std::string bundle_path = std::string(url->fileSystemRepresentation()) + "/" +
|
|
SWIFTPM_BUNDLE + ".bundle";
|
|
auto bundle = NS::Bundle::alloc()->init(
|
|
NS::String::string(bundle_path.c_str(), NS::UTF8StringEncoding));
|
|
if (bundle != nullptr) {
|
|
std::string resource_path =
|
|
std::string(bundle->resourceURL()->fileSystemRepresentation()) + "/" +
|
|
lib_name + ".metallib";
|
|
auto [lib, error] = load_library_from_path(device, resource_path.c_str());
|
|
if (lib) {
|
|
return lib;
|
|
}
|
|
}
|
|
return nullptr;
|
|
}
|
|
#endif
|
|
|
|
// Firstly, search for the metallib in the same path as this binary
|
|
std::pair<MTL::Library*, NS::Error*> load_colocated_library(
|
|
MTL::Device* device,
|
|
const std::string& relative_path) {
|
|
std::string binary_dir = get_binary_directory();
|
|
if (binary_dir.size() == 0) {
|
|
return {nullptr, nullptr};
|
|
}
|
|
|
|
auto path = fs::path(binary_dir) / relative_path;
|
|
if (!path.has_extension()) {
|
|
path.replace_extension(".metallib");
|
|
}
|
|
|
|
return load_library_from_path(device, path.c_str());
|
|
}
|
|
|
|
std::pair<MTL::Library*, NS::Error*> load_swiftpm_library(
|
|
MTL::Device* device,
|
|
const std::string& lib_name) {
|
|
#ifdef SWIFTPM_BUNDLE
|
|
MTL::Library* library =
|
|
try_load_bundle(device, NS::Bundle::mainBundle()->bundleURL(), lib_name);
|
|
if (library != nullptr) {
|
|
return {library, nullptr};
|
|
}
|
|
auto bundles = NS::Bundle::allBundles();
|
|
for (int i = 0, c = (int)bundles->count(); i < c; i++) {
|
|
auto bundle = reinterpret_cast<NS::Bundle*>(bundles->object(i));
|
|
library = try_load_bundle(device, bundle->resourceURL(), lib_name);
|
|
if (library != nullptr) {
|
|
return {library, nullptr};
|
|
}
|
|
}
|
|
#endif
|
|
return {nullptr, nullptr};
|
|
}
|
|
|
|
MTL::Library* load_default_library(MTL::Device* device) {
|
|
NS::Error* error[4];
|
|
MTL::Library* lib;
|
|
// First try the colocated mlx.metallib
|
|
std::tie(lib, error[0]) = load_colocated_library(device, "mlx");
|
|
if (lib) {
|
|
return lib;
|
|
}
|
|
|
|
std::tie(lib, error[1]) = load_colocated_library(device, "Resources/mlx");
|
|
if (lib) {
|
|
return lib;
|
|
}
|
|
|
|
// Then try default.metallib in a SwiftPM bundle if we have one
|
|
std::tie(lib, error[2]) = load_swiftpm_library(device, "default");
|
|
if (lib) {
|
|
return lib;
|
|
}
|
|
|
|
// Finally try default_mtllib_path
|
|
std::tie(lib, error[3]) = load_library_from_path(device, default_mtllib_path);
|
|
if (!lib) {
|
|
std::ostringstream msg;
|
|
msg << "Failed to load the default metallib. ";
|
|
for (int i = 0; i < 4; i++) {
|
|
if (error[i] != nullptr) {
|
|
msg << error[i]->localizedDescription()->utf8String() << " ";
|
|
}
|
|
}
|
|
throw std::runtime_error(msg.str());
|
|
}
|
|
return lib;
|
|
}
|
|
|
|
MTL::Library* load_library(
|
|
MTL::Device* device,
|
|
const std::string& lib_name,
|
|
const std::string& lib_path) {
|
|
// We have been given a path that ends in metallib so try to load it
|
|
if (lib_path.size() > 9 &&
|
|
std::equal(lib_path.end() - 9, lib_path.end(), ".metallib")) {
|
|
auto [lib, error] = load_library_from_path(device, lib_path.c_str());
|
|
if (!lib) {
|
|
std::ostringstream msg;
|
|
msg << "Failed to load the metallib from <" << lib_path << "> with error "
|
|
<< error->localizedDescription()->utf8String();
|
|
throw std::runtime_error(msg.str());
|
|
}
|
|
return lib;
|
|
}
|
|
|
|
// We have been given a path so try to load from lib_path / lib_name.metallib
|
|
if (lib_path.size() > 0) {
|
|
std::string full_path = lib_path + "/" + lib_name + ".metallib";
|
|
auto [lib, error] = load_library_from_path(device, full_path.c_str());
|
|
if (!lib) {
|
|
std::ostringstream msg;
|
|
msg << "Failed to load the metallib from <" << full_path
|
|
<< "> with error " << error->localizedDescription()->utf8String();
|
|
throw std::runtime_error(msg.str());
|
|
}
|
|
return lib;
|
|
}
|
|
|
|
// Try to load the colocated library
|
|
{
|
|
auto [lib, error] = load_colocated_library(device, lib_name);
|
|
if (lib) {
|
|
return lib;
|
|
}
|
|
}
|
|
|
|
// Try to load the library from swiftpm
|
|
{
|
|
auto [lib, error] = load_swiftpm_library(device, lib_name);
|
|
if (lib) {
|
|
return lib;
|
|
}
|
|
}
|
|
|
|
std::ostringstream msg;
|
|
msg << "Failed to load the metallib " << lib_name << ".metallib. "
|
|
<< "We attempted to load it from <" << get_binary_directory() << "/"
|
|
<< lib_name << ".metallib" << ">";
|
|
#ifdef SWIFTPM_BUNDLE
|
|
msg << " and from the Swift PM bundle.";
|
|
#endif
|
|
throw std::runtime_error(msg.str());
|
|
}
|
|
|
|
} // namespace
|
|
|
|
CommandEncoder::CommandEncoder(DeviceStream& stream) : stream_(stream) {
|
|
enc_ = stream_.buffer->computeCommandEncoder(MTL::DispatchTypeConcurrent);
|
|
enc_->retain();
|
|
}
|
|
|
|
CommandEncoder::~CommandEncoder() {
|
|
enc_->endEncoding();
|
|
enc_->release();
|
|
}
|
|
|
|
void CommandEncoder::set_buffer(
|
|
const MTL::Buffer* buf,
|
|
int idx,
|
|
int64_t offset /* = 0 */) {
|
|
enc_->setBuffer(buf, offset, idx);
|
|
}
|
|
|
|
void CommandEncoder::set_input_array(
|
|
const array& a,
|
|
int idx,
|
|
int64_t offset /* = 0 */) {
|
|
if (all_inputs_.insert(a.buffer().ptr()).second) {
|
|
stream_.buffer_sizes += a.data_size();
|
|
}
|
|
auto r_buf = static_cast<MTL::Resource*>(const_cast<void*>(a.buffer().ptr()));
|
|
needs_barrier_ =
|
|
needs_barrier_ | (prev_outputs_.find(r_buf) != prev_outputs_.end());
|
|
auto a_buf = static_cast<const MTL::Buffer*>(a.buffer().ptr());
|
|
auto base_offset = a.data<char>() -
|
|
static_cast<char*>(const_cast<MTL::Buffer*>(a_buf)->contents());
|
|
base_offset += offset;
|
|
enc_->setBuffer(a_buf, base_offset, idx);
|
|
}
|
|
|
|
void CommandEncoder::set_output_array(
|
|
array& a,
|
|
int idx,
|
|
int64_t offset /* = 0 */) {
|
|
// Add barriers before adding the output to the output set
|
|
set_input_array(a, idx, offset);
|
|
register_output_array(a);
|
|
}
|
|
|
|
void CommandEncoder::register_output_array(const array& a) {
|
|
all_outputs_.insert(a.buffer().ptr());
|
|
|
|
auto buf = static_cast<MTL::Resource*>(const_cast<void*>(a.buffer().ptr()));
|
|
if (concurrent_) {
|
|
concurrent_outputs_.insert(buf);
|
|
} else {
|
|
next_outputs_.insert(buf);
|
|
}
|
|
}
|
|
|
|
void CommandEncoder::maybeInsertBarrier() {
|
|
if (needs_barrier_) {
|
|
enc_->memoryBarrier(MTL::BarrierScopeBuffers);
|
|
needs_barrier_ = false;
|
|
prev_outputs_ = std::move(next_outputs_);
|
|
} else {
|
|
prev_outputs_.insert(next_outputs_.begin(), next_outputs_.end());
|
|
}
|
|
next_outputs_.clear();
|
|
}
|
|
|
|
void CommandEncoder::dispatch_threadgroups(
|
|
MTL::Size grid_dims,
|
|
MTL::Size group_dims) {
|
|
maybeInsertBarrier();
|
|
stream_.buffer_ops++;
|
|
enc_->dispatchThreadgroups(grid_dims, group_dims);
|
|
}
|
|
|
|
void CommandEncoder::dispatch_threads(
|
|
MTL::Size grid_dims,
|
|
MTL::Size group_dims) {
|
|
maybeInsertBarrier();
|
|
stream_.buffer_ops++;
|
|
enc_->dispatchThreads(grid_dims, group_dims);
|
|
}
|
|
|
|
void CommandEncoder::barrier() {
|
|
enc_->memoryBarrier(MTL::BarrierScopeBuffers);
|
|
}
|
|
|
|
Device::Device() {
|
|
auto pool = new_scoped_memory_pool();
|
|
device_ = load_device();
|
|
default_library_ = load_default_library(device_);
|
|
arch_ = std::string(device_->architecture()->name()->utf8String());
|
|
auto arch = arch_.back();
|
|
switch (arch) {
|
|
case 'p': // phone
|
|
max_ops_per_buffer_ = 20;
|
|
max_mb_per_buffer_ = 40;
|
|
break;
|
|
case 'g': // base, pro
|
|
max_ops_per_buffer_ = 40;
|
|
max_mb_per_buffer_ = 40;
|
|
break;
|
|
case 's': // max
|
|
max_ops_per_buffer_ = 50;
|
|
max_mb_per_buffer_ = 50;
|
|
break;
|
|
case 'd': // ultra
|
|
max_ops_per_buffer_ = 50;
|
|
max_mb_per_buffer_ = 50;
|
|
break;
|
|
default: // default to medium
|
|
max_ops_per_buffer_ = 40;
|
|
max_mb_per_buffer_ = 40;
|
|
break;
|
|
}
|
|
max_ops_per_buffer_ = env::max_ops_per_buffer(max_ops_per_buffer_);
|
|
max_mb_per_buffer_ = env::max_mb_per_buffer(max_mb_per_buffer_);
|
|
}
|
|
|
|
Device::~Device() {
|
|
auto pool = new_scoped_memory_pool();
|
|
for (auto& [l, kernel_map] : library_kernels_) {
|
|
l->release();
|
|
for (auto& [_, k] : kernel_map) {
|
|
k->release();
|
|
}
|
|
}
|
|
stream_map_.clear();
|
|
device_->release();
|
|
}
|
|
|
|
void Device::new_queue(int index) {
|
|
auto thread_pool = metal::new_scoped_memory_pool();
|
|
auto q = device_->newCommandQueue();
|
|
debug_set_stream_queue_label(q, index);
|
|
if (!q) {
|
|
throw std::runtime_error(
|
|
"[metal::Device] Failed to make new command queue.");
|
|
}
|
|
stream_map_.emplace(index, q);
|
|
if (residency_set_ != nullptr) {
|
|
q->addResidencySet(residency_set_);
|
|
}
|
|
}
|
|
|
|
MTL::CommandQueue* Device::get_queue(Stream stream) {
|
|
return get_stream_(stream.index).queue;
|
|
}
|
|
|
|
bool Device::command_buffer_needs_commit(int index) {
|
|
auto& stream = get_stream_(index);
|
|
if (stream.buffer_ops > max_ops_per_buffer_ ||
|
|
(stream.buffer_sizes >> 20) > max_mb_per_buffer_) {
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
MTL::CommandBuffer* Device::get_command_buffer(int index) {
|
|
auto& stream = get_stream_(index);
|
|
if (stream.buffer == nullptr) {
|
|
stream.buffer = stream.queue->commandBufferWithUnretainedReferences();
|
|
if (!stream.buffer) {
|
|
throw std::runtime_error(
|
|
"[metal::Device] Unable to create new command buffer");
|
|
}
|
|
// Increment ref count so the buffer is not garbage collected
|
|
stream.buffer->retain();
|
|
}
|
|
return stream.buffer;
|
|
}
|
|
|
|
void Device::commit_command_buffer(int index) {
|
|
auto& stream = get_stream_(index);
|
|
stream.buffer->commit();
|
|
stream.buffer->release();
|
|
stream.buffer = nullptr;
|
|
stream.buffer_ops = 0;
|
|
stream.buffer_sizes = 0;
|
|
}
|
|
|
|
void Device::add_temporary(array arr, int index) {
|
|
get_stream_(index).temporaries.push_back(std::move(arr));
|
|
}
|
|
|
|
void Device::add_temporaries(std::vector<array> arrays, int index) {
|
|
if (arrays.empty()) {
|
|
return;
|
|
}
|
|
auto& stream = get_stream_(index);
|
|
stream.temporaries.insert(
|
|
stream.temporaries.end(),
|
|
std::make_move_iterator(arrays.begin()),
|
|
std::make_move_iterator(arrays.end()));
|
|
}
|
|
|
|
void Device::end_encoding(int index) {
|
|
auto& stream = get_stream_(index);
|
|
if (stream.encoder != nullptr) {
|
|
// Each command encoder has a unique fence. We also store a map of
|
|
// all previous outputs of command encoders to their corresponding fence.
|
|
// - The command encoder records its inputs and outputs.
|
|
// - Wait on a fence if any inputs in the encoder are outputs of a previous
|
|
// encoder.
|
|
// - Update the map of outputs to include this command encoder's outputs.
|
|
// - Always signal this command encoders fence.
|
|
// - Add a completion handler for this command encoder that removes outputs
|
|
// from the map to limit the growth of the map and avoid unnecessary waits
|
|
// - Temporaries are a special case as they do not cross command encoder
|
|
// boundaries. These can be removed early from the encoders inputs and
|
|
// outputs since they don't need synchronization.
|
|
auto& enc = *stream.encoder;
|
|
// Remove temporaries from inputs and outputs
|
|
for (auto& t : stream.temporaries) {
|
|
if (t.data<void>() != nullptr) {
|
|
enc.outputs().erase(t.buffer().ptr());
|
|
enc.inputs().erase(t.buffer().ptr());
|
|
}
|
|
}
|
|
|
|
// Keep references to the fences we waited on and put them
|
|
// in the completion handler so they are not prematurely released
|
|
std::unordered_set<std::shared_ptr<Fence>> waiting_on;
|
|
{
|
|
std::lock_guard<std::mutex> lk(stream.fence_mtx);
|
|
for (auto in : enc.inputs()) {
|
|
if (auto it = stream.outputs.find(in); it != stream.outputs.end()) {
|
|
// If we've already waited on a fence, don't wait on it again.
|
|
if (waiting_on.find(it->second) == waiting_on.end()) {
|
|
enc.wait_for_fence(it->second->fence);
|
|
waiting_on.insert(it->second);
|
|
}
|
|
}
|
|
}
|
|
for (auto out : enc.outputs()) {
|
|
stream.outputs[out] = stream.fence;
|
|
}
|
|
}
|
|
enc.update_fence(stream.fence->fence);
|
|
stream.buffer->addCompletedHandler(
|
|
[&stream,
|
|
waiting_on = std::move(waiting_on),
|
|
fence = std::move(stream.fence),
|
|
outputs = std::move(enc.outputs()),
|
|
temporaries =
|
|
std::move(stream.temporaries)](MTL::CommandBuffer*) mutable {
|
|
temporaries.clear();
|
|
std::lock_guard<std::mutex> lk(stream.fence_mtx);
|
|
for (auto o : outputs) {
|
|
if (auto it = stream.outputs.find(o); it != stream.outputs.end()) {
|
|
if (it->second == fence) {
|
|
stream.outputs.erase(it);
|
|
}
|
|
}
|
|
}
|
|
});
|
|
}
|
|
stream.encoder = nullptr;
|
|
}
|
|
|
|
CommandEncoder& Device::get_command_encoder(int index) {
|
|
auto& stream = get_stream_(index);
|
|
if (stream.encoder == nullptr) {
|
|
stream.encoder = std::make_unique<CommandEncoder>(stream);
|
|
stream.fence = std::make_shared<Fence>(device_->newFence());
|
|
}
|
|
return *stream.encoder;
|
|
}
|
|
|
|
MTL::Library* Device::get_library(
|
|
const std::string& name,
|
|
const std::string& path /* = "" */) {
|
|
{
|
|
std::shared_lock rlock(library_mtx_);
|
|
if (auto it = library_map_.find(name); it != library_map_.end()) {
|
|
return it->second;
|
|
}
|
|
}
|
|
|
|
std::unique_lock wlock(library_mtx_);
|
|
if (auto it = library_map_.find(name); it != library_map_.end()) {
|
|
return it->second;
|
|
}
|
|
|
|
auto new_lib = load_library(device_, name, path.c_str());
|
|
library_map_.insert({name, new_lib});
|
|
return new_lib;
|
|
}
|
|
|
|
MTL::Library* Device::build_library_(const std::string& source_string) {
|
|
auto pool = new_scoped_memory_pool();
|
|
|
|
auto ns_code =
|
|
NS::String::string(source_string.c_str(), NS::ASCIIStringEncoding);
|
|
|
|
NS::Error* error = nullptr;
|
|
auto options = MTL::CompileOptions::alloc()->init();
|
|
options->setFastMathEnabled(false);
|
|
options->setLanguageVersion(get_metal_version());
|
|
auto mtl_lib = device_->newLibrary(ns_code, options, &error);
|
|
options->release();
|
|
|
|
// Throw error if unable to compile library
|
|
if (!mtl_lib) {
|
|
std::ostringstream msg;
|
|
msg << "[metal::Device] Unable to build metal library from source\n";
|
|
if (error) {
|
|
msg << error->localizedDescription()->utf8String() << "\n";
|
|
}
|
|
throw std::runtime_error(msg.str());
|
|
}
|
|
|
|
return mtl_lib;
|
|
}
|
|
|
|
MTL::Function* Device::get_function_(
|
|
const std::string& name,
|
|
MTL::Library* mtl_lib) {
|
|
// Pull kernel from library
|
|
auto ns_name = NS::String::string(name.c_str(), NS::ASCIIStringEncoding);
|
|
auto mtl_function = mtl_lib->newFunction(ns_name);
|
|
|
|
return mtl_function;
|
|
}
|
|
|
|
MTL::Function* Device::get_function_(
|
|
const std::string& name,
|
|
const std::string& specialized_name,
|
|
const MTLFCList& func_consts,
|
|
MTL::Library* mtl_lib) {
|
|
if (func_consts.empty() && (specialized_name == name)) {
|
|
return get_function_(name, mtl_lib);
|
|
}
|
|
|
|
// Prepare function constants
|
|
auto mtl_func_consts = MTL::FunctionConstantValues::alloc()->init();
|
|
|
|
for (auto [value, type, index] : func_consts) {
|
|
mtl_func_consts->setConstantValue(value, type, index);
|
|
}
|
|
|
|
// Prepare function desc
|
|
auto desc = MTL::FunctionDescriptor::functionDescriptor();
|
|
desc->setName(NS::String::string(name.c_str(), NS::ASCIIStringEncoding));
|
|
desc->setSpecializedName(
|
|
NS::String::string(specialized_name.c_str(), NS::ASCIIStringEncoding));
|
|
desc->setConstantValues(mtl_func_consts);
|
|
|
|
// Pull kernel from library
|
|
NS::Error* error = nullptr;
|
|
auto mtl_function = mtl_lib->newFunction(desc, &error);
|
|
|
|
// Throw error if unable to build metal function
|
|
if (!mtl_function) {
|
|
std::ostringstream msg;
|
|
msg << "[metal::Device] Unable to load function " << name << "\n";
|
|
if (error) {
|
|
msg << error->localizedDescription()->utf8String() << "\n";
|
|
}
|
|
throw std::runtime_error(msg.str());
|
|
}
|
|
|
|
mtl_func_consts->release();
|
|
|
|
return mtl_function;
|
|
}
|
|
|
|
MTL::ComputePipelineState* Device::get_kernel_(
|
|
const std::string& name,
|
|
const MTL::Function* mtl_function) {
|
|
// Compile kernel to compute pipeline
|
|
NS::Error* error = nullptr;
|
|
MTL::ComputePipelineState* kernel;
|
|
|
|
if (mtl_function) {
|
|
kernel = device_->newComputePipelineState(mtl_function, &error);
|
|
}
|
|
|
|
// Throw error if unable to compile metal function
|
|
if (!mtl_function || !kernel) {
|
|
std::ostringstream msg;
|
|
msg << "[metal::Device] Unable to load kernel " << name << "\n";
|
|
if (error) {
|
|
msg << error->localizedDescription()->utf8String() << "\n";
|
|
}
|
|
throw std::runtime_error(msg.str());
|
|
}
|
|
|
|
return kernel;
|
|
}
|
|
|
|
MTL::ComputePipelineState* Device::get_kernel_(
|
|
const std::string& name,
|
|
const MTL::Function* mtl_function,
|
|
const MTL::LinkedFunctions* linked_functions) {
|
|
// Check inputs
|
|
if (!linked_functions) {
|
|
return get_kernel_(name, mtl_function);
|
|
}
|
|
|
|
if (!mtl_function) {
|
|
std::ostringstream msg;
|
|
msg << "[metal::Device] Unable to load kernel " << name << "\n";
|
|
throw std::runtime_error(msg.str());
|
|
}
|
|
|
|
// Prepare compute pipeline state descriptor
|
|
auto desc = MTL::ComputePipelineDescriptor::alloc()->init();
|
|
desc->setComputeFunction(mtl_function);
|
|
desc->setLinkedFunctions(linked_functions);
|
|
|
|
// Compile kernel to compute pipeline
|
|
NS::Error* error = nullptr;
|
|
auto kernel = device_->newComputePipelineState(
|
|
desc, MTL::PipelineOptionNone, nullptr, &error);
|
|
|
|
// Throw error if unable to compile metal function
|
|
if (!kernel) {
|
|
std::ostringstream msg;
|
|
msg << "[metal::Device] Unable to load kernel " << name << "\n";
|
|
if (error) {
|
|
msg << error->localizedDescription()->utf8String() << "\n";
|
|
}
|
|
throw std::runtime_error(msg.str());
|
|
}
|
|
|
|
return kernel;
|
|
}
|
|
|
|
MTL::Library* Device::get_library_(const std::string& name) {
|
|
std::shared_lock lock(library_mtx_);
|
|
auto it = library_map_.find(name);
|
|
return (it != library_map_.end()) ? it->second : nullptr;
|
|
}
|
|
|
|
MTL::Library* Device::get_library(
|
|
const std::string& name,
|
|
const std::function<std::string(void)>& builder) {
|
|
{
|
|
std::shared_lock rlock(library_mtx_);
|
|
if (auto it = library_map_.find(name); it != library_map_.end()) {
|
|
return it->second;
|
|
}
|
|
}
|
|
|
|
std::unique_lock wlock(library_mtx_);
|
|
if (auto it = library_map_.find(name); it != library_map_.end()) {
|
|
return it->second;
|
|
}
|
|
|
|
auto mtl_lib = build_library_(builder());
|
|
library_map_.insert({name, mtl_lib});
|
|
return mtl_lib;
|
|
}
|
|
|
|
void Device::clear_library(const std::string& name) {
|
|
std::unique_lock wlock(library_mtx_);
|
|
if (auto it = library_map_.find(name); it != library_map_.end()) {
|
|
auto kernel_map_it = library_kernels_.find(it->second);
|
|
for (auto& [_, kernel] : kernel_map_it->second) {
|
|
kernel->release();
|
|
}
|
|
library_kernels_.erase(kernel_map_it);
|
|
it->second->release();
|
|
library_map_.erase(it);
|
|
}
|
|
}
|
|
|
|
MTL::LinkedFunctions* Device::get_linked_functions_(
|
|
const std::vector<MTL::Function*>& funcs) {
|
|
if (funcs.empty()) {
|
|
return nullptr;
|
|
}
|
|
|
|
auto lfuncs = MTL::LinkedFunctions::linkedFunctions();
|
|
|
|
std::vector<NS::Object*> objs(funcs.size());
|
|
for (int i = 0; i < funcs.size(); i++) {
|
|
objs[i] = funcs[i];
|
|
}
|
|
|
|
NS::Array* funcs_arr = NS::Array::array(objs.data(), funcs.size());
|
|
|
|
lfuncs->setPrivateFunctions(funcs_arr);
|
|
|
|
return lfuncs;
|
|
}
|
|
|
|
MTL::ComputePipelineState* Device::get_kernel_(
|
|
const std::string& base_name,
|
|
MTL::Library* mtl_lib,
|
|
const std::string& hash_name,
|
|
const MTLFCList& func_consts /* = {} */,
|
|
const std::vector<MTL::Function*>& linked_functions /* = {} */) {
|
|
// Single writer allowed
|
|
std::unique_lock wlock(kernel_mtx_);
|
|
|
|
// Try loading again to avoid loading twice
|
|
auto& kernel_map_ = library_kernels_[mtl_lib];
|
|
if (auto it = kernel_map_.find(hash_name); it != kernel_map_.end()) {
|
|
return it->second;
|
|
}
|
|
|
|
auto pool = new_scoped_memory_pool();
|
|
|
|
// Pull kernel from library
|
|
auto mtl_function = get_function_(base_name, hash_name, func_consts, mtl_lib);
|
|
|
|
// Compile kernel to compute pipeline
|
|
auto mtl_linked_funcs = get_linked_functions_(linked_functions);
|
|
auto kernel = get_kernel_(hash_name, mtl_function, mtl_linked_funcs);
|
|
|
|
mtl_function->release();
|
|
mtl_linked_funcs->release();
|
|
|
|
// Add kernel to cache
|
|
auto inserted = kernel_map_.insert({hash_name, kernel});
|
|
|
|
return kernel;
|
|
}
|
|
|
|
MTL::ComputePipelineState* Device::get_kernel(
|
|
const std::string& base_name,
|
|
MTL::Library* mtl_lib,
|
|
const std::string& hash_name /* = "" */,
|
|
const MTLFCList& func_consts /* = {} */,
|
|
const std::vector<MTL::Function*>& linked_functions /* = {} */) {
|
|
const auto& kname = hash_name.empty() ? base_name : hash_name;
|
|
{
|
|
// Multiple readers allowed
|
|
std::shared_lock lock(kernel_mtx_);
|
|
|
|
// Look for cached kernel
|
|
auto& kernel_map_ = library_kernels_[mtl_lib];
|
|
if (auto it = kernel_map_.find(kname); it != kernel_map_.end()) {
|
|
return it->second;
|
|
}
|
|
}
|
|
return get_kernel_(base_name, mtl_lib, kname, func_consts, linked_functions);
|
|
}
|
|
|
|
MTL::ComputePipelineState* Device::get_kernel(
|
|
const std::string& base_name,
|
|
const std::string& hash_name /* = "" */,
|
|
const MTLFCList& func_consts /* = {} */,
|
|
const std::vector<MTL::Function*>& linked_functions /* = {} */) {
|
|
return get_kernel(
|
|
base_name, default_library_, hash_name, func_consts, linked_functions);
|
|
}
|
|
|
|
void Device::set_residency_set(const MTL::ResidencySet* residency_set) {
|
|
if (residency_set_ != nullptr) {
|
|
throw std::runtime_error(
|
|
"[Device::set_residency_set] Can only be set once.");
|
|
}
|
|
if (residency_set == nullptr) {
|
|
return;
|
|
}
|
|
residency_set_ = residency_set;
|
|
// Attach residency set to existing command queues
|
|
for (auto& [_, stream] : stream_map_) {
|
|
stream.queue->addResidencySet(residency_set_);
|
|
}
|
|
}
|
|
|
|
Device& device(mlx::core::Device) {
|
|
static Device metal_device;
|
|
return metal_device;
|
|
}
|
|
|
|
std::unique_ptr<void, std::function<void(void*)>> new_scoped_memory_pool() {
|
|
auto dtor = [](void* ptr) {
|
|
static_cast<NS::AutoreleasePool*>(ptr)->release();
|
|
};
|
|
return std::unique_ptr<void, std::function<void(void*)>>(
|
|
NS::AutoreleasePool::alloc()->init(), dtor);
|
|
}
|
|
|
|
} // namespace mlx::core::metal
|