Details | Last modification | View Log | RSS feed
Rev | Author | Line No. | Line |
---|---|---|---|
14 | pmbaty | 1 | //===- TrainingLogger.h - mlgo feature/reward logging ----------*- C++ -*-===// |
2 | // |
||
3 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
||
4 | // See https://llvm.org/LICENSE.txt for license information. |
||
5 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
||
6 | // |
||
7 | //===----------------------------------------------------------------------===// |
||
8 | // |
||
9 | // The design goals of the logger are: |
||
10 | // - no dependencies that llvm doesn't already have. |
||
11 | // - support streaming, so that we don't need to buffer data during compilation |
||
12 | // - 0-decoding tensor values. Tensor values are potentially very large buffers |
||
13 | // of scalars. Because of their potentially large size, avoiding |
||
14 | // serialization/deserialization overhead is preferred. |
||
15 | // |
||
16 | // The simple logger produces an output of the form (each line item on its line) |
||
17 | // - header: a json object describing the data that will follow. |
||
18 | // - context: e.g. function name, for regalloc, or "default" for module-wide |
||
19 | // optimizations like the inliner. This is the context to which the subsequent |
||
20 | // data corresponds. |
||
21 | // - observation number. |
||
22 | // - tensor values - raw bytes of the tensors, in the order given in the header. |
||
23 | // The values are in succession, i.e. no separator is found between successive |
||
24 | // tensor values. At the end, there is a new line character. |
||
25 | // - [score] - this is optional, and is present if it was present in the header. |
||
26 | // Currently, for final rewards, we output "0" scores after each observation, |
||
27 | // except for the last one. |
||
28 | // <repeat> |
||
29 | // The file should be read as binary, but the reason we use newlines is mostly |
||
30 | // ease of debugging: the log can be opened in a text editor and, while tensor |
||
31 | // values are inscrutable, at least the sequence of data can be easily observed. |
||
32 | // Of course, the buffer of tensor values could contain '\n' bytes. A reader |
||
33 | // should use the header information to know how much data to read for the |
||
34 | // tensor values, and not use line information for that. |
||
35 | // |
||
36 | // An example reader, used for test, is available at |
||
37 | // Analysis/models/log_reader.py |
||
38 | // |
||
39 | // Example: |
||
40 | // {"features":[list of TensorSpecs], "score":<a tensor spec>} |
||
41 | // {"context": "aFunction"} |
||
42 | // {"observation": 0} |
||
43 | // <bytes> |
||
44 | // {"outcome": 0} |
||
45 | // <bytes for the tensor corresponding to the "score" spec in the header> |
||
46 | // {"observation": 1} |
||
47 | // ... |
||
48 | // {"context": "anotherFunction"} |
||
49 | // {"observation": 0} |
||
50 | // ... |
||
51 | // |
||
52 | |||
53 | #ifndef LLVM_ANALYSIS_UTILS_TRAININGLOGGER_H |
||
54 | #define LLVM_ANALYSIS_UTILS_TRAININGLOGGER_H |
||
55 | |||
56 | #include "llvm/Config/llvm-config.h" |
||
57 | |||
58 | #include "llvm/ADT/StringMap.h" |
||
59 | #include "llvm/Analysis/TensorSpec.h" |
||
60 | #include "llvm/IR/LLVMContext.h" |
||
61 | #include "llvm/Support/JSON.h" |
||
62 | |||
63 | #include <memory> |
||
64 | #include <optional> |
||
65 | #include <vector> |
||
66 | |||
67 | namespace llvm { |
||
68 | |||
69 | /// Logging utility - given an ordered specification of features, and assuming |
||
70 | /// a scalar reward, allow logging feature values and rewards. |
||
71 | /// The assumption is that, for an event to be logged (i.e. a set of feature |
||
72 | /// values and a reward), the user calls the log* API for each feature exactly |
||
73 | /// once, providing the index matching the position in the feature spec list |
||
74 | /// provided at construction. The example assumes the first feature's element |
||
75 | /// type is float, the second is int64, and the reward is float: |
||
76 | /// |
||
77 | /// event 0: |
||
78 | /// logFloatValue(0, ...) |
||
79 | /// logInt64Value(1, ...) |
||
80 | /// ... |
||
81 | /// logFloatReward(...) |
||
82 | /// event 1: |
||
83 | /// logFloatValue(0, ...) |
||
84 | /// logInt64Value(1, ...) |
||
85 | /// ... |
||
86 | /// logFloatReward(...) |
||
87 | /// |
||
88 | /// At the end, call print to generate the log. |
||
89 | /// Alternatively, don't call logReward at the end of each event, just |
||
90 | /// log{Float|Int32|Int64}FinalReward at the end. |
||
91 | class Logger final { |
||
92 | std::unique_ptr<raw_ostream> OS; |
||
93 | const std::vector<TensorSpec> FeatureSpecs; |
||
94 | const TensorSpec RewardSpec; |
||
95 | const bool IncludeReward; |
||
96 | StringMap<size_t> ObservationIDs; |
||
97 | std::string CurrentContext; |
||
98 | |||
99 | void writeHeader(); |
||
100 | void writeTensor(const TensorSpec &Spec, const char *RawData) { |
||
101 | OS->write(RawData, Spec.getTotalTensorBufferSize()); |
||
102 | } |
||
103 | void logRewardImpl(const char *RawData); |
||
104 | |||
105 | public: |
||
106 | /// Construct a Logger. If IncludeReward is false, then logReward or |
||
107 | /// logFinalReward shouldn't be called, and the reward feature won't be |
||
108 | /// printed out. |
||
109 | /// NOTE: the FeatureSpecs are expected to be in the same order (i.e. have |
||
110 | /// corresponding indices) with any MLModelRunner implementations |
||
111 | /// corresponding to the model being trained/logged. |
||
112 | Logger(std::unique_ptr<raw_ostream> OS, |
||
113 | const std::vector<TensorSpec> &FeatureSpecs, |
||
114 | const TensorSpec &RewardSpec, bool IncludeReward); |
||
115 | |||
116 | void switchContext(StringRef Name); |
||
117 | void startObservation(); |
||
118 | void endObservation(); |
||
119 | |||
120 | const std::string ¤tContext() const { return CurrentContext; } |
||
121 | |||
122 | bool hasObservationInProgress() const { |
||
123 | return ObservationIDs.find(CurrentContext) != ObservationIDs.end(); |
||
124 | } |
||
125 | |||
126 | template <typename T> void logReward(T Value) { |
||
127 | logRewardImpl(reinterpret_cast<const char *>(&Value)); |
||
128 | } |
||
129 | |||
130 | void logTensorValue(size_t FeatureID, const char *RawData) { |
||
131 | writeTensor(FeatureSpecs[FeatureID], RawData); |
||
132 | } |
||
133 | }; |
||
134 | |||
135 | } // namespace llvm |
||
136 | #endif // LLVM_ANALYSIS_UTILS_TRAININGLOGGER_H |