1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108
|
// Copyright 2021 The Chromium Authors
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
#ifndef COMPONENTS_OPTIMIZATION_GUIDE_CORE_INFERENCE_MODEL_EXECUTOR_H_
#define COMPONENTS_OPTIMIZATION_GUIDE_CORE_INFERENCE_MODEL_EXECUTOR_H_
#include <optional>
#include "base/files/file_path.h"
#include "base/functional/bind.h"
#include "base/functional/callback.h"
#include "base/memory/weak_ptr.h"
#include "base/task/sequenced_task_runner.h"
#include "base/time/time.h"
#include "base/types/optional_ref.h"
#include "components/optimization_guide/proto/models.pb.h"
namespace optimization_guide {
// This class handles the execution, loading, unloading, and associated metrics
// of machine learning models in Optimization Guide on a specified thread. This
// class is meant to be used and owned by an instance of |ModelHandler|. A
// ModelExecutor must be passed to a ModelHandler's constructor, this design
// allows the implementer of a ModelExecutor to define how the model is built
// and executed. See also tflite_model_executor.h, base_model_executor.h, and
// base_model_executor_helpers.h in this directory for helpful derived classes.
//
// Lifetime: This class can be constructed on any thread but cannot do anything
// useful until |InitializeAndMoveToExecutionThread| is called. After that
// method is called, all subsequent calls to this class must be made through the
// |execution_task_runner| that was passed to initialize. Furthermore, all
// WeakPointers of this class must only be dereferenced on the
// |execution_task_runner| thread as well. This in turn means that this class
// must be destroyed on the |execution_task_runner| thread as well.
template <class OutputType, class InputType>
class ModelExecutor {
public:
ModelExecutor() = default;
virtual ~ModelExecutor() = default;
// If |model_inference_timeout| is nullopt a default value will be used,
// controlled by the optimization guide.
virtual void InitializeAndMoveToExecutionThread(
std::optional<base::TimeDelta> model_inference_timeout,
proto::OptimizationTarget optimization_target,
scoped_refptr<base::SequencedTaskRunner> execution_task_runner,
scoped_refptr<base::SequencedTaskRunner> reply_task_runner) = 0;
// Updates model file. If `SetShouldUnloadModelOnComplete` is false,
// immedidately loads model into memory. `file_path` will be nullopt if no
// valid model is found, and the previous model should be unloaded in that
// case.
virtual void UpdateModelFile(
base::optional_ref<const base::FilePath> file_path) = 0;
virtual void UnloadModel() = 0;
// Sets whether the model file should be unloaded from memory after each
// execution. If set to false, use |UnloadModel| to unload the model when
// needed.
virtual void SetShouldUnloadModelOnComplete(bool should_auto_unload) = 0;
// Sets whether the model should be loaded as soon as its file path is
// available.
virtual void SetShouldPreloadModel(bool should_preload_model) = 0;
using ExecutionCallback =
base::OnceCallback<void(const std::optional<OutputType>&)>;
virtual void SendForExecution(ExecutionCallback callback_on_complete,
base::TimeTicks start_time,
InputType input) = 0;
// Some callers define their InputType as a const ref type, but you can't make
// vectors of references. Strip those qualifiers off and add them back to the
// vector instead.
using ConstRefInputVector = const std::vector<typename std::remove_const<
typename std::remove_reference<InputType>::type>::type>&;
// It is guaranteed that the output passed to |BatchExecutionCallback| will
// always be in the same order as the input vector.
using BatchExecutionCallback =
base::OnceCallback<void(const std::vector<std::optional<OutputType>>&)>;
virtual void SendForBatchExecution(
BatchExecutionCallback callback_on_complete,
base::TimeTicks start_time,
ConstRefInputVector inputs) = 0;
// Synchronous batch execution.
virtual std::vector<std::optional<OutputType>> SendForBatchExecutionSync(
ConstRefInputVector inputs) = 0;
// IMPORTANT: These WeakPointers must only be dereferenced on the
// |execution_task_runner| thread.
base::WeakPtr<ModelExecutor> GetWeakPtrForExecutionThread() {
return weak_ptr_factory_.GetWeakPtr();
}
ModelExecutor(const ModelExecutor&) = delete;
ModelExecutor& operator=(const ModelExecutor&) = delete;
private:
base::WeakPtrFactory<ModelExecutor> weak_ptr_factory_{this};
};
} // namespace optimization_guide
#endif // COMPONENTS_OPTIMIZATION_GUIDE_CORE_INFERENCE_MODEL_EXECUTOR_H_
|