1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98
|
#include <ATen/core/dispatch/Dispatcher.h>
#include <ATen/core/type_factory.h>
#include <caffe2/serialize/inline_container.h>
#include <torch/csrc/jit/mobile/compatibility/runtime_compatibility.h>
#include <torch/csrc/jit/mobile/type_parser.h>
#include <torch/csrc/jit/runtime/operator.h>
#include <torch/custom_class.h>
#include <unordered_map>
namespace c10 {
TypePtr parseType(const std::string& pythonStr);
} // namespace c10
namespace torch {
namespace jit {
uint64_t _get_runtime_bytecode_version() {
return caffe2::serialize::kMaxSupportedBytecodeVersion;
}
std::pair<uint64_t, uint64_t> _get_runtime_bytecode_min_max_versions() {
return std::pair<uint64_t, uint64_t>(
caffe2::serialize::kMinSupportedBytecodeVersion,
caffe2::serialize::kMaxSupportedBytecodeVersion);
}
std::pair<uint64_t, uint64_t> _get_runtime_operators_min_max_versions() {
return std::pair<uint64_t, uint64_t>(
caffe2::serialize::kMinSupportedFileFormatVersion,
caffe2::serialize::kMaxSupportedFileFormatVersion);
}
/*
* Returns all registered PyTorch ops and their versioning
*/
std::unordered_map<std::string, OperatorInfo> _get_runtime_ops_and_info() {
std::unordered_map<std::string, OperatorInfo> result;
// Grab the jit operators
auto nonDispatcherOperators = torch::jit::getAllOperators();
for (const auto& full_op : nonDispatcherOperators) {
auto op = full_op->schema();
int num_schema_args = op.arguments().size();
auto op_name = op.name();
if (!op.overload_name().empty()) {
op_name += ("." + op.overload_name());
}
result.emplace(op_name, OperatorInfo{num_schema_args});
}
// Grab the dispatcher operators
auto dispatcherOperators = c10::Dispatcher::singleton().getAllOpNames();
for (auto& op : dispatcherOperators) {
// grab schema
const auto op_handle = c10::Dispatcher::singleton().findOp(op);
c10::optional<int> num_schema_args;
if (op_handle->hasSchema()) {
num_schema_args = op_handle->schema().arguments().size();
}
auto op_name = op.name;
if (!op.overload_name.empty()) {
op_name += ("." + op.overload_name);
}
result.emplace(op_name, OperatorInfo{num_schema_args});
}
return result;
}
RuntimeCompatibilityInfo RuntimeCompatibilityInfo::get() {
return RuntimeCompatibilityInfo{
_get_runtime_bytecode_min_max_versions(),
_get_runtime_ops_and_info(),
_get_mobile_supported_types(),
_get_runtime_operators_min_max_versions()};
}
std::unordered_set<std::string> _get_mobile_supported_types() {
std::unordered_set<std::string> supported_types;
for (const auto& it : c10::DynamicTypeFactory::basePythonTypes()) {
supported_types.insert(it.first);
}
supported_types.insert(
at::TypeParser::getNonSimpleType().begin(),
at::TypeParser::getNonSimpleType().end());
supported_types.insert(
at::TypeParser::getCustomType().begin(),
at::TypeParser::getCustomType().end());
return supported_types;
}
TORCH_API std::unordered_set<std::string> _get_loaded_custom_classes() {
return torch::getAllCustomClassesNames();
}
} // namespace jit
} // namespace torch
|