File: exceptions.patch

package info (click to toggle)
firefox 146.0.1-1
  • links: PTS, VCS
  • area: main
  • in suites: sid
  • size: 4,653,260 kB
  • sloc: cpp: 7,587,892; javascript: 6,509,455; ansic: 3,755,295; python: 1,410,813; xml: 629,201; asm: 438,677; java: 186,096; sh: 62,697; makefile: 18,086; objc: 13,087; perl: 12,811; yacc: 4,583; cs: 3,846; pascal: 3,448; lex: 1,720; ruby: 1,003; php: 436; lisp: 258; awk: 247; sql: 66; sed: 54; csh: 10; exp: 6
file content (203 lines) | stat: -rw-r--r-- 5,766 bytes parent folder | download | duplicates (3)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
diff --git a/ggml/src/gguf.cpp b/ggml/src/gguf.cpp
index 5ffd12b8b2..c2ae0db107 100644
--- a/ggml/src/gguf.cpp
+++ b/ggml/src/gguf.cpp
@@ -15,6 +15,8 @@
 #include <string>
 #include <vector>
 
+#include "moz-overrides.h"
+
 template <typename T>
 struct type_to_gguf_type;
 
diff --git a/src/llama-adapter.cpp b/src/llama-adapter.cpp
index 8d94034aed..995c5fa41d 100644
--- a/src/llama-adapter.cpp
+++ b/src/llama-adapter.cpp
@@ -8,6 +8,8 @@
 #include <cassert>
 #include <stdexcept>
 
+#include "moz-overrides.h"
+
 // vec
 
 ggml_tensor * llama_adapter_cvec::tensor_for(int il) const {
diff --git a/src/llama-chat.cpp b/src/llama-chat.cpp
index 5d317f4ee6..6a7bed5305 100644
--- a/src/llama-chat.cpp
+++ b/src/llama-chat.cpp
@@ -6,6 +6,8 @@
 #include <sstream>
 #include <algorithm>
 
+#include "moz-overrides.h"
+
 #if __cplusplus >= 202000L
     #define LU8(x) (const char*)(u8##x)
 #else
diff --git a/src/llama-context.cpp b/src/llama-context.cpp
index 06e93b19cb..e932b11ac1 100644
--- a/src/llama-context.cpp
+++ b/src/llama-context.cpp
@@ -12,6 +12,8 @@
 #include <limits>
 #include <stdexcept>
 
+#include "moz-overrides.h"
+
 //
 // llama_context
 //
diff --git a/src/llama-grammar.cpp b/src/llama-grammar.cpp
index bed706bb24..04f4b484e5 100644
--- a/src/llama-grammar.cpp
+++ b/src/llama-grammar.cpp
@@ -8,6 +8,8 @@
 #include <algorithm>
 #include <stdexcept>
 
+#include "moz-overrides.h"
+
 //
 // helpers
 //
diff --git a/src/llama-kv-cache-unified.cpp b/src/llama-kv-cache-unified.cpp
index 8517b722a9..c0e6d05c35 100644
--- a/src/llama-kv-cache-unified.cpp
+++ b/src/llama-kv-cache-unified.cpp
@@ -12,6 +12,8 @@
 #include <map>
 #include <stdexcept>
 
+#include "moz-overrides.h"
+
 //
 // llama_kv_cache_unified
 //
diff --git a/src/llama-memory-recurrent.cpp b/src/llama-memory-recurrent.cpp
index 1b1e95d567..b6624a3864 100644
--- a/src/llama-memory-recurrent.cpp
+++ b/src/llama-memory-recurrent.cpp
@@ -11,6 +11,8 @@
 #include <map>
 #include <stdexcept>
 
+#include "moz-overrides.h"
+
 //
 // llama_memory_recurrent
 //
diff --git a/src/llama-mmap.cpp b/src/llama-mmap.cpp
index 47497cf953..c5f8a28bb3 100644
--- a/src/llama-mmap.cpp
+++ b/src/llama-mmap.cpp
@@ -39,6 +39,8 @@
 #include <TargetConditionals.h>
 #endif
 
+#include "moz-overrides.h"
+
 // TODO: consider moving to llama-impl.h if needed in more places
 #if defined(_WIN32)
 static std::string llama_format_win_err(DWORD err) {
diff --git a/src/llama-model-loader.cpp b/src/llama-model-loader.cpp
index bd9e6da883..1a90a1eb88 100644
--- a/src/llama-model-loader.cpp
+++ b/src/llama-model-loader.cpp
@@ -7,6 +7,8 @@
 #include <cstring>
 #include <future>
 
+#include "moz-overrides.h"
+
 static const size_t kiB = 1024;
 static const size_t MiB = 1024*kiB;
 static const size_t GiB = 1024*MiB;
diff --git a/src/llama-model-loader.h b/src/llama-model-loader.h
index 0f52b011b6..ded54f544a 100644
--- a/src/llama-model-loader.h
+++ b/src/llama-model-loader.h
@@ -34,12 +34,14 @@
         llama_tensor_weight(const llama_file * file, uint16_t idx, const struct gguf_context * gguf_ctx, ggml_tensor * tensor) : idx(idx), tensor(tensor) {
             const int tensor_idx = gguf_find_tensor(gguf_ctx,  ggml_get_name(tensor));
             if (tensor_idx < 0) {
-                throw std::runtime_error(format("tensor '%s' not found in the model", ggml_get_name(tensor)));
+                // throw std::runtime_error(format("tensor '%s' not found in the model", ggml_get_name(tensor)));
+              std::abort();
             }
 
             offs = gguf_get_data_offset(gguf_ctx) + gguf_get_tensor_offset(gguf_ctx, tensor_idx);
             if (offs + ggml_nbytes(tensor) < offs || offs + ggml_nbytes(tensor) > file->size()) {
-                throw std::runtime_error(format("tensor '%s' data is not within the file bounds, model is corrupted or incomplete", ggml_get_name(tensor)));
+              //throw std::runtime_error(format("tensor '%s' data is not within the file bounds, model is corrupted or incomplete", ggml_get_name(tensor)));
+              std::abort();
             }
         }
     };
diff --git a/src/llama-model.cpp b/src/llama-model.cpp
index fc39195ed5..e041f7e1a6 100644
--- a/src/llama-model.cpp
+++ b/src/llama-model.cpp
@@ -25,6 +25,8 @@
 #include <sstream>
 #include <stdexcept>
 
+#include "moz-overrides.h"
+
 const char * llm_type_name(llm_type type) {
     switch (type) {
         case LLM_TYPE_14M:           return "14M";
diff --git a/src/llama-sampling.cpp b/src/llama-sampling.cpp
index bfbf5fa230..dd1d9eee52 100644
--- a/src/llama-sampling.cpp
+++ b/src/llama-sampling.cpp
@@ -17,6 +17,8 @@
 #include <unordered_map>
 #include <stdexcept>
 
+#include "moz-overrides.h"
+
 // the ring buffer works similarly to std::deque, but with a fixed capacity
 template<typename T>
 struct ring_buffer {
diff --git a/src/llama-vocab.cpp b/src/llama-vocab.cpp
index 5c9eb87566..334f862100 100644
--- a/src/llama-vocab.cpp
+++ b/src/llama-vocab.cpp
@@ -20,6 +20,8 @@
 #include <set>
 #include <unordered_map>
 
+#include "moz-overrides.h"
+
 //
 // helpers
 //
diff --git a/src/llama.cpp b/src/llama.cpp
index 34906cdb62..0adb16598e 100644
--- a/src/llama.cpp
+++ b/src/llama.cpp
@@ -17,6 +17,8 @@
 #include <cstring>
 #include <ctime>
 
+#include "moz-overrides.h"
+
 #if defined(_MSC_VER)
 #pragma warning(disable: 4244 4267) // possible loss of data
 #endif
diff --git a/src/unicode.cpp b/src/unicode.cpp
index 43a4581b96..6b3b2dbe7d 100644
--- a/src/unicode.cpp
+++ b/src/unicode.cpp
@@ -19,6 +19,8 @@
 #include <utility>
 #include <vector>
 
+#include "moz-overrides.h"
+
 size_t unicode_len_utf8(char src) {
     const size_t lookup[] = { 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 2, 2, 3, 4 };
     uint8_t highbits = static_cast<uint8_t>(src) >> 4;