File: | build/gcc/wide-int.h |
Warning: | line 953, column 5 3rd function call argument is an uninitialized value |
Press '?' to see keyboard shortcuts
Keyboard shortcuts:
1 | /* Support routines for vrange storage. | |||
2 | Copyright (C) 2022-2023 Free Software Foundation, Inc. | |||
3 | Contributed by Aldy Hernandez <aldyh@redhat.com>. | |||
4 | ||||
5 | This file is part of GCC. | |||
6 | ||||
7 | GCC is free software; you can redistribute it and/or modify | |||
8 | it under the terms of the GNU General Public License as published by | |||
9 | the Free Software Foundation; either version 3, or (at your option) | |||
10 | any later version. | |||
11 | ||||
12 | GCC is distributed in the hope that it will be useful, | |||
13 | but WITHOUT ANY WARRANTY; without even the implied warranty of | |||
14 | MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |||
15 | GNU General Public License for more details. | |||
16 | ||||
17 | You should have received a copy of the GNU General Public License | |||
18 | along with GCC; see the file COPYING3. If not see | |||
19 | <http://www.gnu.org/licenses/>. */ | |||
20 | ||||
21 | #include "config.h" | |||
22 | #include "system.h" | |||
23 | #include "coretypes.h" | |||
24 | #include "backend.h" | |||
25 | #include "tree.h" | |||
26 | #include "gimple.h" | |||
27 | #include "ssa.h" | |||
28 | #include "tree-pretty-print.h" | |||
29 | #include "fold-const.h" | |||
30 | #include "gimple-range.h" | |||
31 | #include "value-range-storage.h" | |||
32 | ||||
33 | // Return a newly allocated slot holding R, or NULL if storing a range | |||
34 | // of R's type is not supported. | |||
35 | ||||
36 | void * | |||
37 | vrange_storage::alloc_slot (const vrange &r) | |||
38 | { | |||
39 | gcc_checking_assert (m_alloc)((void)(!(m_alloc) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 39, __FUNCTION__), 0 : 0)); | |||
40 | ||||
41 | if (is_a <irange> (r)) | |||
42 | return irange_storage_slot::alloc_slot (*m_alloc, as_a <irange> (r)); | |||
43 | if (is_a <frange> (r)) | |||
44 | return frange_storage_slot::alloc_slot (*m_alloc, as_a <frange> (r)); | |||
45 | return NULLnullptr; | |||
46 | } | |||
47 | ||||
48 | // Set SLOT to R. | |||
49 | ||||
50 | void | |||
51 | vrange_storage::set_vrange (void *slot, const vrange &r) | |||
52 | { | |||
53 | if (is_a <irange> (r)) | |||
54 | { | |||
55 | irange_storage_slot *s = static_cast <irange_storage_slot *> (slot); | |||
56 | gcc_checking_assert (s->fits_p (as_a <irange> (r)))((void)(!(s->fits_p (as_a <irange> (r))) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 56, __FUNCTION__), 0 : 0)); | |||
57 | s->set_irange (as_a <irange> (r)); | |||
58 | } | |||
59 | else if (is_a <frange> (r)) | |||
60 | { | |||
61 | frange_storage_slot *s = static_cast <frange_storage_slot *> (slot); | |||
62 | gcc_checking_assert (s->fits_p (as_a <frange> (r)))((void)(!(s->fits_p (as_a <frange> (r))) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 62, __FUNCTION__), 0 : 0)); | |||
63 | s->set_frange (as_a <frange> (r)); | |||
64 | } | |||
65 | else | |||
66 | gcc_unreachable ()(fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 66, __FUNCTION__)); | |||
67 | } | |||
68 | ||||
69 | // Restore R from SLOT. TYPE is the type of R. | |||
70 | ||||
71 | void | |||
72 | vrange_storage::get_vrange (const void *slot, vrange &r, tree type) | |||
73 | { | |||
74 | if (is_a <irange> (r)) | |||
75 | { | |||
76 | const irange_storage_slot *s | |||
77 | = static_cast <const irange_storage_slot *> (slot); | |||
78 | s->get_irange (as_a <irange> (r), type); | |||
79 | } | |||
80 | else if (is_a <frange> (r)) | |||
81 | { | |||
82 | const frange_storage_slot *s | |||
83 | = static_cast <const frange_storage_slot *> (slot); | |||
84 | s->get_frange (as_a <frange> (r), type); | |||
85 | } | |||
86 | else | |||
87 | gcc_unreachable ()(fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 87, __FUNCTION__)); | |||
88 | } | |||
89 | ||||
90 | // Return TRUE if SLOT can fit R. | |||
91 | ||||
92 | bool | |||
93 | vrange_storage::fits_p (const void *slot, const vrange &r) | |||
94 | { | |||
95 | if (is_a <irange> (r)) | |||
96 | { | |||
97 | const irange_storage_slot *s | |||
98 | = static_cast <const irange_storage_slot *> (slot); | |||
99 | return s->fits_p (as_a <irange> (r)); | |||
100 | } | |||
101 | if (is_a <frange> (r)) | |||
102 | { | |||
103 | const frange_storage_slot *s | |||
104 | = static_cast <const frange_storage_slot *> (slot); | |||
105 | return s->fits_p (as_a <frange> (r)); | |||
106 | } | |||
107 | gcc_unreachable ()(fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 107, __FUNCTION__)); | |||
108 | return false; | |||
109 | } | |||
110 | ||||
111 | // Factory that creates a new irange_storage_slot object containing R. | |||
112 | // This is the only way to construct an irange slot as stack creation | |||
113 | // is disallowed. | |||
114 | ||||
115 | irange_storage_slot * | |||
116 | irange_storage_slot::alloc_slot (vrange_allocator &allocator, const irange &r) | |||
117 | { | |||
118 | size_t size = irange_storage_slot::size (r); | |||
119 | irange_storage_slot *p | |||
120 | = static_cast <irange_storage_slot *> (allocator.alloc (size)); | |||
121 | new (p) irange_storage_slot (r); | |||
122 | return p; | |||
123 | } | |||
124 | ||||
125 | // Initialize the current slot with R. | |||
126 | ||||
127 | irange_storage_slot::irange_storage_slot (const irange &r) | |||
128 | { | |||
129 | gcc_checking_assert (!r.undefined_p ())((void)(!(!r.undefined_p ()) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 129, __FUNCTION__), 0 : 0)); | |||
130 | ||||
131 | unsigned prec = TYPE_PRECISION (r.type ())((tree_class_check ((r.type ()), (tcc_type), "/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 131, __FUNCTION__))->type_common.precision); | |||
132 | unsigned n = num_wide_ints_needed (r); | |||
133 | if (n > MAX_INTS) | |||
134 | { | |||
135 | int_range<MAX_PAIRS> squash (r); | |||
136 | m_ints.set_precision (prec, num_wide_ints_needed (squash)); | |||
137 | set_irange (squash); | |||
138 | } | |||
139 | else | |||
140 | { | |||
141 | m_ints.set_precision (prec, n); | |||
142 | set_irange (r); | |||
143 | } | |||
144 | } | |||
145 | ||||
146 | // Store R into the current slot. | |||
147 | ||||
148 | void | |||
149 | irange_storage_slot::set_irange (const irange &r) | |||
150 | { | |||
151 | gcc_checking_assert (fits_p (r))((void)(!(fits_p (r)) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 151, __FUNCTION__), 0 : 0)); | |||
152 | ||||
153 | m_ints[0] = r.get_nonzero_bits (); | |||
154 | ||||
155 | unsigned pairs = r.num_pairs (); | |||
156 | for (unsigned i = 0; i < pairs; ++i) | |||
157 | { | |||
158 | m_ints[i*2 + 1] = r.lower_bound (i); | |||
159 | m_ints[i*2 + 2] = r.upper_bound (i); | |||
160 | } | |||
161 | } | |||
162 | ||||
163 | // Restore a range of TYPE from the current slot into R. | |||
164 | ||||
165 | void | |||
166 | irange_storage_slot::get_irange (irange &r, tree type) const | |||
167 | { | |||
168 | gcc_checking_assert (TYPE_PRECISION (type) == m_ints.get_precision ())((void)(!(((tree_class_check ((type), (tcc_type), "/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 168, __FUNCTION__))->type_common.precision) == m_ints.get_precision ()) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 168, __FUNCTION__), 0 : 0)); | |||
169 | ||||
170 | r.set_undefined (); | |||
171 | unsigned nelements = m_ints.num_elements (); | |||
172 | for (unsigned i = 1; i < nelements; i += 2) | |||
173 | { | |||
174 | int_range<2> tmp (type, m_ints[i], m_ints[i + 1]); | |||
175 | r.union_ (tmp); | |||
176 | } | |||
177 | r.set_nonzero_bits (get_nonzero_bits ()); | |||
178 | } | |||
179 | ||||
180 | // Return the size in bytes to allocate a slot that can hold R. | |||
181 | ||||
182 | size_t | |||
183 | irange_storage_slot::size (const irange &r) | |||
184 | { | |||
185 | gcc_checking_assert (!r.undefined_p ())((void)(!(!r.undefined_p ()) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 185, __FUNCTION__), 0 : 0)); | |||
186 | ||||
187 | unsigned prec = TYPE_PRECISION (r.type ())((tree_class_check ((r.type ()), (tcc_type), "/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 187, __FUNCTION__))->type_common.precision); | |||
188 | unsigned n = num_wide_ints_needed (r); | |||
189 | if (n > MAX_INTS) | |||
190 | n = MAX_INTS; | |||
191 | return (sizeof (irange_storage_slot) | |||
192 | + trailing_wide_ints<MAX_INTS>::extra_size (prec, n)); | |||
193 | } | |||
194 | ||||
195 | // Return the number of wide ints needed to represent R. | |||
196 | ||||
197 | unsigned int | |||
198 | irange_storage_slot::num_wide_ints_needed (const irange &r) | |||
199 | { | |||
200 | return r.num_pairs () * 2 + 1; | |||
201 | } | |||
202 | ||||
203 | // Return TRUE if R fits in the current slot. | |||
204 | ||||
205 | bool | |||
206 | irange_storage_slot::fits_p (const irange &r) const | |||
207 | { | |||
208 | return m_ints.num_elements () >= num_wide_ints_needed (r); | |||
209 | } | |||
210 | ||||
211 | // Dump the current slot. | |||
212 | ||||
213 | void | |||
214 | irange_storage_slot::dump () const | |||
215 | { | |||
216 | fprintf (stderrstderr, "raw irange_storage_slot:\n"); | |||
217 | for (unsigned i = 1; i < m_ints.num_elements (); i += 2) | |||
218 | { | |||
219 | m_ints[i].dump (); | |||
220 | m_ints[i + 1].dump (); | |||
221 | } | |||
222 | fprintf (stderrstderr, "NONZERO "); | |||
223 | wide_int nz = get_nonzero_bits (); | |||
224 | nz.dump (); | |||
225 | } | |||
226 | ||||
227 | DEBUG_FUNCTION__attribute__ ((__used__)) void | |||
228 | debug (const irange_storage_slot &storage) | |||
229 | { | |||
230 | storage.dump (); | |||
| ||||
231 | fprintf (stderrstderr, "\n"); | |||
232 | } | |||
233 | ||||
234 | // Implementation of frange_storage_slot. | |||
235 | ||||
236 | frange_storage_slot * | |||
237 | frange_storage_slot::alloc_slot (vrange_allocator &allocator, const frange &r) | |||
238 | { | |||
239 | size_t size = sizeof (frange_storage_slot); | |||
240 | frange_storage_slot *p | |||
241 | = static_cast <frange_storage_slot *> (allocator.alloc (size)); | |||
242 | new (p) frange_storage_slot (r); | |||
243 | return p; | |||
244 | } | |||
245 | ||||
246 | void | |||
247 | frange_storage_slot::set_frange (const frange &r) | |||
248 | { | |||
249 | gcc_checking_assert (fits_p (r))((void)(!(fits_p (r)) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 249, __FUNCTION__), 0 : 0)); | |||
250 | gcc_checking_assert (!r.undefined_p ())((void)(!(!r.undefined_p ()) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 250, __FUNCTION__), 0 : 0)); | |||
251 | ||||
252 | m_kind = r.m_kind; | |||
253 | m_min = r.m_min; | |||
254 | m_max = r.m_max; | |||
255 | m_pos_nan = r.m_pos_nan; | |||
256 | m_neg_nan = r.m_neg_nan; | |||
257 | } | |||
258 | ||||
259 | void | |||
260 | frange_storage_slot::get_frange (frange &r, tree type) const | |||
261 | { | |||
262 | gcc_checking_assert (r.supports_type_p (type))((void)(!(r.supports_type_p (type)) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.cc" , 262, __FUNCTION__), 0 : 0)); | |||
263 | ||||
264 | // Handle explicit NANs. | |||
265 | if (m_kind == VR_NAN) | |||
266 | { | |||
267 | if (HONOR_NANS (type)) | |||
268 | { | |||
269 | if (m_pos_nan && m_neg_nan) | |||
270 | r.set_nan (type); | |||
271 | else | |||
272 | r.set_nan (type, m_neg_nan); | |||
273 | } | |||
274 | else | |||
275 | r.set_undefined (); | |||
276 | return; | |||
277 | } | |||
278 | ||||
279 | // We use the constructor to create the new range instead of writing | |||
280 | // out the bits into the frange directly, because the global range | |||
281 | // being read may be being inlined into a function with different | |||
282 | // restrictions as when it was originally written. We want to make | |||
283 | // sure the resulting range is canonicalized correctly for the new | |||
284 | // consumer. | |||
285 | r = frange (type, m_min, m_max, m_kind); | |||
286 | ||||
287 | // The constructor will set the NAN bits for HONOR_NANS, but we must | |||
288 | // make sure to set the NAN sign if known. | |||
289 | if (HONOR_NANS (type) && (m_pos_nan ^ m_neg_nan) == 1) | |||
290 | r.update_nan (m_neg_nan); | |||
291 | else if (!m_pos_nan && !m_neg_nan) | |||
292 | r.clear_nan (); | |||
293 | } | |||
294 | ||||
295 | bool | |||
296 | frange_storage_slot::fits_p (const frange &) const | |||
297 | { | |||
298 | return true; | |||
299 | } |
1 | /* Support routines for vrange storage. |
2 | Copyright (C) 2022-2023 Free Software Foundation, Inc. |
3 | Contributed by Aldy Hernandez <aldyh@redhat.com>. |
4 | |
5 | This file is part of GCC. |
6 | |
7 | GCC is free software; you can redistribute it and/or modify |
8 | it under the terms of the GNU General Public License as published by |
9 | the Free Software Foundation; either version 3, or (at your option) |
10 | any later version. |
11 | |
12 | GCC is distributed in the hope that it will be useful, |
13 | but WITHOUT ANY WARRANTY; without even the implied warranty of |
14 | MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
15 | GNU General Public License for more details. |
16 | |
17 | You should have received a copy of the GNU General Public License |
18 | along with GCC; see the file COPYING3. If not see |
19 | <http://www.gnu.org/licenses/>. */ |
20 | |
21 | #ifndef GCC_VALUE_RANGE_STORAGE_H |
22 | #define GCC_VALUE_RANGE_STORAGE_H |
23 | |
24 | // This class is used to allocate the minimum amount of storage needed |
25 | // for a given range. Storage is automatically freed at destruction |
26 | // of the class. |
27 | |
28 | class vrange_allocator |
29 | { |
30 | public: |
31 | vrange_allocator () { } |
32 | virtual ~vrange_allocator () { } |
33 | // Allocate a range of TYPE. |
34 | vrange *alloc_vrange (tree type); |
35 | // Allocate a memory block of BYTES. |
36 | virtual void *alloc (unsigned bytes) = 0; |
37 | virtual void free (void *p) = 0; |
38 | // Return a clone of SRC. |
39 | template <typename T> T *clone (const T &src); |
40 | private: |
41 | irange *alloc_irange (unsigned pairs); |
42 | frange *alloc_frange (); |
43 | void operator= (const vrange_allocator &) = delete; |
44 | }; |
45 | |
46 | // This class is used to allocate chunks of memory that can store |
47 | // ranges as memory efficiently as possible. It is meant to be used |
48 | // when long term storage of a range is needed. The class can be used |
49 | // with any vrange_allocator (i.e. alloca or GC). |
50 | |
51 | class vrange_storage |
52 | { |
53 | public: |
54 | vrange_storage (vrange_allocator *alloc) : m_alloc (alloc) { } |
55 | void *alloc_slot (const vrange &r); |
56 | void free (void *slot) { m_alloc->free (slot); } |
57 | void get_vrange (const void *slot, vrange &r, tree type); |
58 | void set_vrange (void *slot, const vrange &r); |
59 | static bool fits_p (const void *slot, const vrange &r); |
60 | private: |
61 | DISABLE_COPY_AND_ASSIGN (vrange_storage)vrange_storage (const vrange_storage&) = delete; void operator = (const vrange_storage &) = delete; |
62 | vrange_allocator *m_alloc; |
63 | }; |
64 | |
65 | // A chunk of memory pointing to an irange storage. |
66 | |
67 | class GTY ((variable_size)) irange_storage_slot |
68 | { |
69 | public: |
70 | static irange_storage_slot *alloc_slot (vrange_allocator &, const irange &r); |
71 | void set_irange (const irange &r); |
72 | void get_irange (irange &r, tree type) const; |
73 | wide_int get_nonzero_bits () const { return m_ints[0]; } |
74 | bool fits_p (const irange &r) const; |
75 | static size_t size (const irange &r); |
76 | void dump () const; |
77 | private: |
78 | DISABLE_COPY_AND_ASSIGN (irange_storage_slot)irange_storage_slot (const irange_storage_slot&) = delete ; void operator= (const irange_storage_slot &) = delete; |
79 | friend void gt_ggc_mx_irange_storage_slot (void *); |
80 | friend void gt_pch_p_19irange_storage_slot (void *, void *, |
81 | gt_pointer_operator, void *); |
82 | friend void gt_pch_nx_irange_storage_slot (void *); |
83 | |
84 | // This is the maximum number of wide_int's allowed in the trailing |
85 | // ints structure, without going over 16 bytes (128 bits) in the |
86 | // control word that precedes the HOST_WIDE_INTs in |
87 | // trailing_wide_ints::m_val[]. |
88 | static const unsigned MAX_INTS = 12; |
89 | |
90 | // Maximum number of range pairs we can handle, considering the |
91 | // nonzero bits take one wide_int. |
92 | static const unsigned MAX_PAIRS = (MAX_INTS - 1) / 2; |
93 | |
94 | // Constructor is private to disallow stack initialization. Use |
95 | // alloc_slot() to create objects. |
96 | irange_storage_slot (const irange &r); |
97 | |
98 | static unsigned num_wide_ints_needed (const irange &r); |
99 | |
100 | trailing_wide_ints<MAX_INTS> m_ints; |
101 | }; |
102 | |
103 | // A chunk of memory to store an frange to long term memory. |
104 | |
105 | class GTY (()) frange_storage_slot |
106 | { |
107 | public: |
108 | static frange_storage_slot *alloc_slot (vrange_allocator &, const frange &r); |
109 | void set_frange (const frange &r); |
110 | void get_frange (frange &r, tree type) const; |
111 | bool fits_p (const frange &) const; |
112 | private: |
113 | frange_storage_slot (const frange &r) { set_frange (r); } |
114 | DISABLE_COPY_AND_ASSIGN (frange_storage_slot)frange_storage_slot (const frange_storage_slot&) = delete ; void operator= (const frange_storage_slot &) = delete; |
115 | |
116 | enum value_range_kind m_kind; |
117 | REAL_VALUE_TYPEstruct real_value m_min; |
118 | REAL_VALUE_TYPEstruct real_value m_max; |
119 | bool m_pos_nan; |
120 | bool m_neg_nan; |
121 | }; |
122 | |
123 | class obstack_vrange_allocator final: public vrange_allocator |
124 | { |
125 | public: |
126 | obstack_vrange_allocator () |
127 | { |
128 | obstack_init (&m_obstack)_obstack_begin ((&m_obstack), 0, 0, (mempool_obstack_chunk_alloc ), (mempool_obstack_chunk_free)); |
129 | } |
130 | virtual ~obstack_vrange_allocator () final override |
131 | { |
132 | obstack_free (&m_obstack, NULL)__extension__ ({ struct obstack *__o = (&m_obstack); void *__obj = (void *) (nullptr); if (__obj > (void *) __o-> chunk && __obj < (void *) __o->chunk_limit) __o ->next_free = __o->object_base = (char *) __obj; else _obstack_free (__o, __obj); }); |
133 | } |
134 | virtual void *alloc (unsigned bytes) final override |
135 | { |
136 | return obstack_alloc (&m_obstack, bytes)__extension__ ({ struct obstack *__h = (&m_obstack); __extension__ ({ struct obstack *__o = (__h); size_t __len = ((bytes)); if (__extension__ ({ struct obstack const *__o1 = (__o); (size_t ) (__o1->chunk_limit - __o1->next_free); }) < __len) _obstack_newchunk (__o, __len); ((void) ((__o)->next_free += (__len))); }); __extension__ ({ struct obstack *__o1 = (__h ); void *__value = (void *) __o1->object_base; if (__o1-> next_free == __value) __o1->maybe_empty_object = 1; __o1-> next_free = (sizeof (ptrdiff_t) < sizeof (void *) ? ((__o1 ->object_base) + (((__o1->next_free) - (__o1->object_base ) + (__o1->alignment_mask)) & ~(__o1->alignment_mask ))) : (char *) (((ptrdiff_t) (__o1->next_free) + (__o1-> alignment_mask)) & ~(__o1->alignment_mask))); if ((size_t ) (__o1->next_free - (char *) __o1->chunk) > (size_t ) (__o1->chunk_limit - (char *) __o1->chunk)) __o1-> next_free = __o1->chunk_limit; __o1->object_base = __o1 ->next_free; __value; }); }); |
137 | } |
138 | virtual void free (void *) final override { } |
139 | private: |
140 | obstack m_obstack; |
141 | }; |
142 | |
143 | class ggc_vrange_allocator final: public vrange_allocator |
144 | { |
145 | public: |
146 | ggc_vrange_allocator () { } |
147 | virtual ~ggc_vrange_allocator () final override { } |
148 | virtual void *alloc (unsigned bytes) final override |
149 | { |
150 | return ggc_internal_alloc (bytes); |
151 | } |
152 | virtual void free (void *p) final override |
153 | { |
154 | return ggc_free (p); |
155 | } |
156 | }; |
157 | |
158 | // Return a new range to hold ranges of TYPE. The newly allocated |
159 | // range is initialized to VR_UNDEFINED. |
160 | |
161 | inline vrange * |
162 | vrange_allocator::alloc_vrange (tree type) |
163 | { |
164 | if (irange::supports_p (type)) |
165 | return alloc_irange (2); |
166 | if (frange::supports_p (type)) |
167 | return alloc_frange (); |
168 | return NULLnullptr; |
169 | gcc_unreachable ()(fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.h" , 169, __FUNCTION__)); |
170 | } |
171 | |
172 | // Return a new range with NUM_PAIRS. |
173 | |
174 | inline irange * |
175 | vrange_allocator::alloc_irange (unsigned num_pairs) |
176 | { |
177 | // Never allocate 0 pairs. |
178 | // Don't allocate 1 either, or we get legacy value_range's. |
179 | if (num_pairs < 2) |
180 | num_pairs = 2; |
181 | |
182 | size_t nbytes = sizeof (tree) * 2 * num_pairs; |
183 | |
184 | // Allocate the irange and required memory for the vector. |
185 | void *r = alloc (sizeof (irange)); |
186 | tree *mem = static_cast <tree *> (alloc (nbytes)); |
187 | return new (r) irange (mem, num_pairs); |
188 | } |
189 | |
190 | inline frange * |
191 | vrange_allocator::alloc_frange () |
192 | { |
193 | void *r = alloc (sizeof (frange)); |
194 | return new (r) frange (); |
195 | } |
196 | |
197 | // Return a clone of an irange. |
198 | |
199 | template <> |
200 | inline irange * |
201 | vrange_allocator::clone <irange> (const irange &src) |
202 | { |
203 | irange *r = alloc_irange (src.num_pairs ()); |
204 | *r = src; |
205 | return r; |
206 | } |
207 | |
208 | // Return a clone of an frange. |
209 | |
210 | template <> |
211 | inline frange * |
212 | vrange_allocator::clone <frange> (const frange &src) |
213 | { |
214 | frange *r = alloc_frange (); |
215 | *r = src; |
216 | return r; |
217 | } |
218 | |
219 | // Return a clone of a vrange. |
220 | |
221 | template <> |
222 | inline vrange * |
223 | vrange_allocator::clone <vrange> (const vrange &src) |
224 | { |
225 | if (is_a <irange> (src)) |
226 | return clone <irange> (as_a <irange> (src)); |
227 | if (is_a <frange> (src)) |
228 | return clone <frange> (as_a <frange> (src)); |
229 | return NULLnullptr; |
230 | gcc_unreachable ()(fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/value-range-storage.h" , 230, __FUNCTION__)); |
231 | } |
232 | |
233 | #endif // GCC_VALUE_RANGE_STORAGE_H |
1 | /* Operations with very long integers. -*- C++ -*- | |||
2 | Copyright (C) 2012-2023 Free Software Foundation, Inc. | |||
3 | ||||
4 | This file is part of GCC. | |||
5 | ||||
6 | GCC is free software; you can redistribute it and/or modify it | |||
7 | under the terms of the GNU General Public License as published by the | |||
8 | Free Software Foundation; either version 3, or (at your option) any | |||
9 | later version. | |||
10 | ||||
11 | GCC is distributed in the hope that it will be useful, but WITHOUT | |||
12 | ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | |||
13 | FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License | |||
14 | for more details. | |||
15 | ||||
16 | You should have received a copy of the GNU General Public License | |||
17 | along with GCC; see the file COPYING3. If not see | |||
18 | <http://www.gnu.org/licenses/>. */ | |||
19 | ||||
20 | #ifndef WIDE_INT_H | |||
21 | #define WIDE_INT_H | |||
22 | ||||
23 | /* wide-int.[cc|h] implements a class that efficiently performs | |||
24 | mathematical operations on finite precision integers. wide_ints | |||
25 | are designed to be transient - they are not for long term storage | |||
26 | of values. There is tight integration between wide_ints and the | |||
27 | other longer storage GCC representations (rtl and tree). | |||
28 | ||||
29 | The actual precision of a wide_int depends on the flavor. There | |||
30 | are three predefined flavors: | |||
31 | ||||
32 | 1) wide_int (the default). This flavor does the math in the | |||
33 | precision of its input arguments. It is assumed (and checked) | |||
34 | that the precisions of the operands and results are consistent. | |||
35 | This is the most efficient flavor. It is not possible to examine | |||
36 | bits above the precision that has been specified. Because of | |||
37 | this, the default flavor has semantics that are simple to | |||
38 | understand and in general model the underlying hardware that the | |||
39 | compiler is targetted for. | |||
40 | ||||
41 | This flavor must be used at the RTL level of gcc because there | |||
42 | is, in general, not enough information in the RTL representation | |||
43 | to extend a value beyond the precision specified in the mode. | |||
44 | ||||
45 | This flavor should also be used at the TREE and GIMPLE levels of | |||
46 | the compiler except for the circumstances described in the | |||
47 | descriptions of the other two flavors. | |||
48 | ||||
49 | The default wide_int representation does not contain any | |||
50 | information inherent about signedness of the represented value, | |||
51 | so it can be used to represent both signed and unsigned numbers. | |||
52 | For operations where the results depend on signedness (full width | |||
53 | multiply, division, shifts, comparisons, and operations that need | |||
54 | overflow detected), the signedness must be specified separately. | |||
55 | ||||
56 | 2) offset_int. This is a fixed-precision integer that can hold | |||
57 | any address offset, measured in either bits or bytes, with at | |||
58 | least one extra sign bit. At the moment the maximum address | |||
59 | size GCC supports is 64 bits. With 8-bit bytes and an extra | |||
60 | sign bit, offset_int therefore needs to have at least 68 bits | |||
61 | of precision. We round this up to 128 bits for efficiency. | |||
62 | Values of type T are converted to this precision by sign- or | |||
63 | zero-extending them based on the signedness of T. | |||
64 | ||||
65 | The extra sign bit means that offset_int is effectively a signed | |||
66 | 128-bit integer, i.e. it behaves like int128_t. | |||
67 | ||||
68 | Since the values are logically signed, there is no need to | |||
69 | distinguish between signed and unsigned operations. Sign-sensitive | |||
70 | comparison operators <, <=, > and >= are therefore supported. | |||
71 | Shift operators << and >> are also supported, with >> being | |||
72 | an _arithmetic_ right shift. | |||
73 | ||||
74 | [ Note that, even though offset_int is effectively int128_t, | |||
75 | it can still be useful to use unsigned comparisons like | |||
76 | wi::leu_p (a, b) as a more efficient short-hand for | |||
77 | "a >= 0 && a <= b". ] | |||
78 | ||||
79 | 3) widest_int. This representation is an approximation of | |||
80 | infinite precision math. However, it is not really infinite | |||
81 | precision math as in the GMP library. It is really finite | |||
82 | precision math where the precision is 4 times the size of the | |||
83 | largest integer that the target port can represent. | |||
84 | ||||
85 | Like offset_int, widest_int is wider than all the values that | |||
86 | it needs to represent, so the integers are logically signed. | |||
87 | Sign-sensitive comparison operators <, <=, > and >= are supported, | |||
88 | as are << and >>. | |||
89 | ||||
90 | There are several places in the GCC where this should/must be used: | |||
91 | ||||
92 | * Code that does induction variable optimizations. This code | |||
93 | works with induction variables of many different types at the | |||
94 | same time. Because of this, it ends up doing many different | |||
95 | calculations where the operands are not compatible types. The | |||
96 | widest_int makes this easy, because it provides a field where | |||
97 | nothing is lost when converting from any variable, | |||
98 | ||||
99 | * There are a small number of passes that currently use the | |||
100 | widest_int that should use the default. These should be | |||
101 | changed. | |||
102 | ||||
103 | There are surprising features of offset_int and widest_int | |||
104 | that the users should be careful about: | |||
105 | ||||
106 | 1) Shifts and rotations are just weird. You have to specify a | |||
107 | precision in which the shift or rotate is to happen in. The bits | |||
108 | above this precision are zeroed. While this is what you | |||
109 | want, it is clearly non obvious. | |||
110 | ||||
111 | 2) Larger precision math sometimes does not produce the same | |||
112 | answer as would be expected for doing the math at the proper | |||
113 | precision. In particular, a multiply followed by a divide will | |||
114 | produce a different answer if the first product is larger than | |||
115 | what can be represented in the input precision. | |||
116 | ||||
117 | The offset_int and the widest_int flavors are more expensive | |||
118 | than the default wide int, so in addition to the caveats with these | |||
119 | two, the default is the prefered representation. | |||
120 | ||||
121 | All three flavors of wide_int are represented as a vector of | |||
122 | HOST_WIDE_INTs. The default and widest_int vectors contain enough elements | |||
123 | to hold a value of MAX_BITSIZE_MODE_ANY_INT bits. offset_int contains only | |||
124 | enough elements to hold ADDR_MAX_PRECISION bits. The values are stored | |||
125 | in the vector with the least significant HOST_BITS_PER_WIDE_INT bits | |||
126 | in element 0. | |||
127 | ||||
128 | The default wide_int contains three fields: the vector (VAL), | |||
129 | the precision and a length (LEN). The length is the number of HWIs | |||
130 | needed to represent the value. widest_int and offset_int have a | |||
131 | constant precision that cannot be changed, so they only store the | |||
132 | VAL and LEN fields. | |||
133 | ||||
134 | Since most integers used in a compiler are small values, it is | |||
135 | generally profitable to use a representation of the value that is | |||
136 | as small as possible. LEN is used to indicate the number of | |||
137 | elements of the vector that are in use. The numbers are stored as | |||
138 | sign extended numbers as a means of compression. Leading | |||
139 | HOST_WIDE_INTs that contain strings of either -1 or 0 are removed | |||
140 | as long as they can be reconstructed from the top bit that is being | |||
141 | represented. | |||
142 | ||||
143 | The precision and length of a wide_int are always greater than 0. | |||
144 | Any bits in a wide_int above the precision are sign-extended from the | |||
145 | most significant bit. For example, a 4-bit value 0x8 is represented as | |||
146 | VAL = { 0xf...fff8 }. However, as an optimization, we allow other integer | |||
147 | constants to be represented with undefined bits above the precision. | |||
148 | This allows INTEGER_CSTs to be pre-extended according to TYPE_SIGN, | |||
149 | so that the INTEGER_CST representation can be used both in TYPE_PRECISION | |||
150 | and in wider precisions. | |||
151 | ||||
152 | There are constructors to create the various forms of wide_int from | |||
153 | trees, rtl and constants. For trees the options are: | |||
154 | ||||
155 | tree t = ...; | |||
156 | wi::to_wide (t) // Treat T as a wide_int | |||
157 | wi::to_offset (t) // Treat T as an offset_int | |||
158 | wi::to_widest (t) // Treat T as a widest_int | |||
159 | ||||
160 | All three are light-weight accessors that should have no overhead | |||
161 | in release builds. If it is useful for readability reasons to | |||
162 | store the result in a temporary variable, the preferred method is: | |||
163 | ||||
164 | wi::tree_to_wide_ref twide = wi::to_wide (t); | |||
165 | wi::tree_to_offset_ref toffset = wi::to_offset (t); | |||
166 | wi::tree_to_widest_ref twidest = wi::to_widest (t); | |||
167 | ||||
168 | To make an rtx into a wide_int, you have to pair it with a mode. | |||
169 | The canonical way to do this is with rtx_mode_t as in: | |||
170 | ||||
171 | rtx r = ... | |||
172 | wide_int x = rtx_mode_t (r, mode); | |||
173 | ||||
174 | Similarly, a wide_int can only be constructed from a host value if | |||
175 | the target precision is given explicitly, such as in: | |||
176 | ||||
177 | wide_int x = wi::shwi (c, prec); // sign-extend C if necessary | |||
178 | wide_int y = wi::uhwi (c, prec); // zero-extend C if necessary | |||
179 | ||||
180 | However, offset_int and widest_int have an inherent precision and so | |||
181 | can be initialized directly from a host value: | |||
182 | ||||
183 | offset_int x = (int) c; // sign-extend C | |||
184 | widest_int x = (unsigned int) c; // zero-extend C | |||
185 | ||||
186 | It is also possible to do arithmetic directly on rtx_mode_ts and | |||
187 | constants. For example: | |||
188 | ||||
189 | wi::add (r1, r2); // add equal-sized rtx_mode_ts r1 and r2 | |||
190 | wi::add (r1, 1); // add 1 to rtx_mode_t r1 | |||
191 | wi::lshift (1, 100); // 1 << 100 as a widest_int | |||
192 | ||||
193 | Many binary operations place restrictions on the combinations of inputs, | |||
194 | using the following rules: | |||
195 | ||||
196 | - {rtx, wide_int} op {rtx, wide_int} -> wide_int | |||
197 | The inputs must be the same precision. The result is a wide_int | |||
198 | of the same precision | |||
199 | ||||
200 | - {rtx, wide_int} op (un)signed HOST_WIDE_INT -> wide_int | |||
201 | (un)signed HOST_WIDE_INT op {rtx, wide_int} -> wide_int | |||
202 | The HOST_WIDE_INT is extended or truncated to the precision of | |||
203 | the other input. The result is a wide_int of the same precision | |||
204 | as that input. | |||
205 | ||||
206 | - (un)signed HOST_WIDE_INT op (un)signed HOST_WIDE_INT -> widest_int | |||
207 | The inputs are extended to widest_int precision and produce a | |||
208 | widest_int result. | |||
209 | ||||
210 | - offset_int op offset_int -> offset_int | |||
211 | offset_int op (un)signed HOST_WIDE_INT -> offset_int | |||
212 | (un)signed HOST_WIDE_INT op offset_int -> offset_int | |||
213 | ||||
214 | - widest_int op widest_int -> widest_int | |||
215 | widest_int op (un)signed HOST_WIDE_INT -> widest_int | |||
216 | (un)signed HOST_WIDE_INT op widest_int -> widest_int | |||
217 | ||||
218 | Other combinations like: | |||
219 | ||||
220 | - widest_int op offset_int and | |||
221 | - wide_int op offset_int | |||
222 | ||||
223 | are not allowed. The inputs should instead be extended or truncated | |||
224 | so that they match. | |||
225 | ||||
226 | The inputs to comparison functions like wi::eq_p and wi::lts_p | |||
227 | follow the same compatibility rules, although their return types | |||
228 | are different. Unary functions on X produce the same result as | |||
229 | a binary operation X + X. Shift functions X op Y also produce | |||
230 | the same result as X + X; the precision of the shift amount Y | |||
231 | can be arbitrarily different from X. */ | |||
232 | ||||
233 | /* The MAX_BITSIZE_MODE_ANY_INT is automatically generated by a very | |||
234 | early examination of the target's mode file. The WIDE_INT_MAX_ELTS | |||
235 | can accomodate at least 1 more bit so that unsigned numbers of that | |||
236 | mode can be represented as a signed value. Note that it is still | |||
237 | possible to create fixed_wide_ints that have precisions greater than | |||
238 | MAX_BITSIZE_MODE_ANY_INT. This can be useful when representing a | |||
239 | double-width multiplication result, for example. */ | |||
240 | #define WIDE_INT_MAX_ELTS(((64*(8)) + 64) / 64) \ | |||
241 | ((MAX_BITSIZE_MODE_ANY_INT(64*(8)) + HOST_BITS_PER_WIDE_INT64) / HOST_BITS_PER_WIDE_INT64) | |||
242 | ||||
243 | #define WIDE_INT_MAX_PRECISION((((64*(8)) + 64) / 64) * 64) (WIDE_INT_MAX_ELTS(((64*(8)) + 64) / 64) * HOST_BITS_PER_WIDE_INT64) | |||
244 | ||||
245 | /* This is the max size of any pointer on any machine. It does not | |||
246 | seem to be as easy to sniff this out of the machine description as | |||
247 | it is for MAX_BITSIZE_MODE_ANY_INT since targets may support | |||
248 | multiple address sizes and may have different address sizes for | |||
249 | different address spaces. However, currently the largest pointer | |||
250 | on any platform is 64 bits. When that changes, then it is likely | |||
251 | that a target hook should be defined so that targets can make this | |||
252 | value larger for those targets. */ | |||
253 | #define ADDR_MAX_BITSIZE64 64 | |||
254 | ||||
255 | /* This is the internal precision used when doing any address | |||
256 | arithmetic. The '4' is really 3 + 1. Three of the bits are for | |||
257 | the number of extra bits needed to do bit addresses and the other bit | |||
258 | is to allow everything to be signed without loosing any precision. | |||
259 | Then everything is rounded up to the next HWI for efficiency. */ | |||
260 | #define ADDR_MAX_PRECISION((64 + 4 + 64 - 1) & ~(64 - 1)) \ | |||
261 | ((ADDR_MAX_BITSIZE64 + 4 + HOST_BITS_PER_WIDE_INT64 - 1) \ | |||
262 | & ~(HOST_BITS_PER_WIDE_INT64 - 1)) | |||
263 | ||||
264 | /* The number of HWIs needed to store an offset_int. */ | |||
265 | #define OFFSET_INT_ELTS(((64 + 4 + 64 - 1) & ~(64 - 1)) / 64) (ADDR_MAX_PRECISION((64 + 4 + 64 - 1) & ~(64 - 1)) / HOST_BITS_PER_WIDE_INT64) | |||
266 | ||||
267 | /* The type of result produced by a binary operation on types T1 and T2. | |||
268 | Defined purely for brevity. */ | |||
269 | #define WI_BINARY_RESULT(T1, T2)typename wi::binary_traits <T1, T2>::result_type \ | |||
270 | typename wi::binary_traits <T1, T2>::result_type | |||
271 | ||||
272 | /* Likewise for binary operators, which excludes the case in which neither | |||
273 | T1 nor T2 is a wide-int-based type. */ | |||
274 | #define WI_BINARY_OPERATOR_RESULT(T1, T2)typename wi::binary_traits <T1, T2>::operator_result \ | |||
275 | typename wi::binary_traits <T1, T2>::operator_result | |||
276 | ||||
277 | /* The type of result produced by T1 << T2. Leads to substitution failure | |||
278 | if the operation isn't supported. Defined purely for brevity. */ | |||
279 | #define WI_SIGNED_SHIFT_RESULT(T1, T2)typename wi::binary_traits <T1, T2>::signed_shift_result_type \ | |||
280 | typename wi::binary_traits <T1, T2>::signed_shift_result_type | |||
281 | ||||
282 | /* The type of result produced by a sign-agnostic binary predicate on | |||
283 | types T1 and T2. This is bool if wide-int operations make sense for | |||
284 | T1 and T2 and leads to substitution failure otherwise. */ | |||
285 | #define WI_BINARY_PREDICATE_RESULT(T1, T2)typename wi::binary_traits <T1, T2>::predicate_result \ | |||
286 | typename wi::binary_traits <T1, T2>::predicate_result | |||
287 | ||||
288 | /* The type of result produced by a signed binary predicate on types T1 and T2. | |||
289 | This is bool if signed comparisons make sense for T1 and T2 and leads to | |||
290 | substitution failure otherwise. */ | |||
291 | #define WI_SIGNED_BINARY_PREDICATE_RESULT(T1, T2)typename wi::binary_traits <T1, T2>::signed_predicate_result \ | |||
292 | typename wi::binary_traits <T1, T2>::signed_predicate_result | |||
293 | ||||
294 | /* The type of result produced by a unary operation on type T. */ | |||
295 | #define WI_UNARY_RESULT(T)typename wi::binary_traits <T, T>::result_type \ | |||
296 | typename wi::binary_traits <T, T>::result_type | |||
297 | ||||
298 | /* Define a variable RESULT to hold the result of a binary operation on | |||
299 | X and Y, which have types T1 and T2 respectively. Define VAL to | |||
300 | point to the blocks of RESULT. Once the user of the macro has | |||
301 | filled in VAL, it should call RESULT.set_len to set the number | |||
302 | of initialized blocks. */ | |||
303 | #define WI_BINARY_RESULT_VAR(RESULT, VAL, T1, X, T2, Y)typename wi::binary_traits <T1, T2>::result_type RESULT = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (X, Y); long *VAL = RESULT .write_val () \ | |||
304 | WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type RESULT = \ | |||
305 | wi::int_traits <WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type>::get_binary_result (X, Y); \ | |||
306 | HOST_WIDE_INTlong *VAL = RESULT.write_val () | |||
307 | ||||
308 | /* Similar for the result of a unary operation on X, which has type T. */ | |||
309 | #define WI_UNARY_RESULT_VAR(RESULT, VAL, T, X)typename wi::binary_traits <T, T>::result_type RESULT = wi::int_traits <typename wi::binary_traits <T, T>:: result_type>::get_binary_result (X, X); long *VAL = RESULT .write_val () \ | |||
310 | WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type RESULT = \ | |||
311 | wi::int_traits <WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type>::get_binary_result (X, X); \ | |||
312 | HOST_WIDE_INTlong *VAL = RESULT.write_val () | |||
313 | ||||
314 | template <typename T> class generic_wide_int; | |||
315 | template <int N> class fixed_wide_int_storage; | |||
316 | class wide_int_storage; | |||
317 | ||||
318 | /* An N-bit integer. Until we can use typedef templates, use this instead. */ | |||
319 | #define FIXED_WIDE_INT(N)generic_wide_int < fixed_wide_int_storage <N> > \ | |||
320 | generic_wide_int < fixed_wide_int_storage <N> > | |||
321 | ||||
322 | typedef generic_wide_int <wide_int_storage> wide_int; | |||
323 | typedef FIXED_WIDE_INT (ADDR_MAX_PRECISION)generic_wide_int < fixed_wide_int_storage <((64 + 4 + 64 - 1) & ~(64 - 1))> > offset_int; | |||
324 | typedef FIXED_WIDE_INT (WIDE_INT_MAX_PRECISION)generic_wide_int < fixed_wide_int_storage <((((64*(8)) + 64) / 64) * 64)> > widest_int; | |||
325 | /* Spelled out explicitly (rather than through FIXED_WIDE_INT) | |||
326 | so as not to confuse gengtype. */ | |||
327 | typedef generic_wide_int < fixed_wide_int_storage <WIDE_INT_MAX_PRECISION((((64*(8)) + 64) / 64) * 64) * 2> > widest2_int; | |||
328 | ||||
329 | /* wi::storage_ref can be a reference to a primitive type, | |||
330 | so this is the conservatively-correct setting. */ | |||
331 | template <bool SE, bool HDP = true> | |||
332 | class wide_int_ref_storage; | |||
333 | ||||
334 | typedef generic_wide_int <wide_int_ref_storage <false> > wide_int_ref; | |||
335 | ||||
336 | /* This can be used instead of wide_int_ref if the referenced value is | |||
337 | known to have type T. It carries across properties of T's representation, | |||
338 | such as whether excess upper bits in a HWI are defined, and can therefore | |||
339 | help avoid redundant work. | |||
340 | ||||
341 | The macro could be replaced with a template typedef, once we're able | |||
342 | to use those. */ | |||
343 | #define WIDE_INT_REF_FOR(T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > \ | |||
344 | generic_wide_int \ | |||
345 | <wide_int_ref_storage <wi::int_traits <T>::is_sign_extended, \ | |||
346 | wi::int_traits <T>::host_dependent_precision> > | |||
347 | ||||
348 | namespace wi | |||
349 | { | |||
350 | /* Operations that calculate overflow do so even for | |||
351 | TYPE_OVERFLOW_WRAPS types. For example, adding 1 to +MAX_INT in | |||
352 | an unsigned int is 0 and does not overflow in C/C++, but wi::add | |||
353 | will set the overflow argument in case it's needed for further | |||
354 | analysis. | |||
355 | ||||
356 | For operations that require overflow, these are the different | |||
357 | types of overflow. */ | |||
358 | enum overflow_type { | |||
359 | OVF_NONE = 0, | |||
360 | OVF_UNDERFLOW = -1, | |||
361 | OVF_OVERFLOW = 1, | |||
362 | /* There was an overflow, but we are unsure whether it was an | |||
363 | overflow or an underflow. */ | |||
364 | OVF_UNKNOWN = 2 | |||
365 | }; | |||
366 | ||||
367 | /* Classifies an integer based on its precision. */ | |||
368 | enum precision_type { | |||
369 | /* The integer has both a precision and defined signedness. This allows | |||
370 | the integer to be converted to any width, since we know whether to fill | |||
371 | any extra bits with zeros or signs. */ | |||
372 | FLEXIBLE_PRECISION, | |||
373 | ||||
374 | /* The integer has a variable precision but no defined signedness. */ | |||
375 | VAR_PRECISION, | |||
376 | ||||
377 | /* The integer has a constant precision (known at GCC compile time) | |||
378 | and is signed. */ | |||
379 | CONST_PRECISION | |||
380 | }; | |||
381 | ||||
382 | /* This class, which has no default implementation, is expected to | |||
383 | provide the following members: | |||
384 | ||||
385 | static const enum precision_type precision_type; | |||
386 | Classifies the type of T. | |||
387 | ||||
388 | static const unsigned int precision; | |||
389 | Only defined if precision_type == CONST_PRECISION. Specifies the | |||
390 | precision of all integers of type T. | |||
391 | ||||
392 | static const bool host_dependent_precision; | |||
393 | True if the precision of T depends (or can depend) on the host. | |||
394 | ||||
395 | static unsigned int get_precision (const T &x) | |||
396 | Return the number of bits in X. | |||
397 | ||||
398 | static wi::storage_ref *decompose (HOST_WIDE_INT *scratch, | |||
399 | unsigned int precision, const T &x) | |||
400 | Decompose X as a PRECISION-bit integer, returning the associated | |||
401 | wi::storage_ref. SCRATCH is available as scratch space if needed. | |||
402 | The routine should assert that PRECISION is acceptable. */ | |||
403 | template <typename T> struct int_traits; | |||
404 | ||||
405 | /* This class provides a single type, result_type, which specifies the | |||
406 | type of integer produced by a binary operation whose inputs have | |||
407 | types T1 and T2. The definition should be symmetric. */ | |||
408 | template <typename T1, typename T2, | |||
409 | enum precision_type P1 = int_traits <T1>::precision_type, | |||
410 | enum precision_type P2 = int_traits <T2>::precision_type> | |||
411 | struct binary_traits; | |||
412 | ||||
413 | /* Specify the result type for each supported combination of binary | |||
414 | inputs. Note that CONST_PRECISION and VAR_PRECISION cannot be | |||
415 | mixed, in order to give stronger type checking. When both inputs | |||
416 | are CONST_PRECISION, they must have the same precision. */ | |||
417 | template <typename T1, typename T2> | |||
418 | struct binary_traits <T1, T2, FLEXIBLE_PRECISION, FLEXIBLE_PRECISION> | |||
419 | { | |||
420 | typedef widest_int result_type; | |||
421 | /* Don't define operators for this combination. */ | |||
422 | }; | |||
423 | ||||
424 | template <typename T1, typename T2> | |||
425 | struct binary_traits <T1, T2, FLEXIBLE_PRECISION, VAR_PRECISION> | |||
426 | { | |||
427 | typedef wide_int result_type; | |||
428 | typedef result_type operator_result; | |||
429 | typedef bool predicate_result; | |||
430 | }; | |||
431 | ||||
432 | template <typename T1, typename T2> | |||
433 | struct binary_traits <T1, T2, FLEXIBLE_PRECISION, CONST_PRECISION> | |||
434 | { | |||
435 | /* Spelled out explicitly (rather than through FIXED_WIDE_INT) | |||
436 | so as not to confuse gengtype. */ | |||
437 | typedef generic_wide_int < fixed_wide_int_storage | |||
438 | <int_traits <T2>::precision> > result_type; | |||
439 | typedef result_type operator_result; | |||
440 | typedef bool predicate_result; | |||
441 | typedef result_type signed_shift_result_type; | |||
442 | typedef bool signed_predicate_result; | |||
443 | }; | |||
444 | ||||
445 | template <typename T1, typename T2> | |||
446 | struct binary_traits <T1, T2, VAR_PRECISION, FLEXIBLE_PRECISION> | |||
447 | { | |||
448 | typedef wide_int result_type; | |||
449 | typedef result_type operator_result; | |||
450 | typedef bool predicate_result; | |||
451 | }; | |||
452 | ||||
453 | template <typename T1, typename T2> | |||
454 | struct binary_traits <T1, T2, CONST_PRECISION, FLEXIBLE_PRECISION> | |||
455 | { | |||
456 | /* Spelled out explicitly (rather than through FIXED_WIDE_INT) | |||
457 | so as not to confuse gengtype. */ | |||
458 | typedef generic_wide_int < fixed_wide_int_storage | |||
459 | <int_traits <T1>::precision> > result_type; | |||
460 | typedef result_type operator_result; | |||
461 | typedef bool predicate_result; | |||
462 | typedef result_type signed_shift_result_type; | |||
463 | typedef bool signed_predicate_result; | |||
464 | }; | |||
465 | ||||
466 | template <typename T1, typename T2> | |||
467 | struct binary_traits <T1, T2, CONST_PRECISION, CONST_PRECISION> | |||
468 | { | |||
469 | STATIC_ASSERT (int_traits <T1>::precision == int_traits <T2>::precision)static_assert ((int_traits <T1>::precision == int_traits <T2>::precision), "int_traits <T1>::precision == int_traits <T2>::precision" ); | |||
470 | /* Spelled out explicitly (rather than through FIXED_WIDE_INT) | |||
471 | so as not to confuse gengtype. */ | |||
472 | typedef generic_wide_int < fixed_wide_int_storage | |||
473 | <int_traits <T1>::precision> > result_type; | |||
474 | typedef result_type operator_result; | |||
475 | typedef bool predicate_result; | |||
476 | typedef result_type signed_shift_result_type; | |||
477 | typedef bool signed_predicate_result; | |||
478 | }; | |||
479 | ||||
480 | template <typename T1, typename T2> | |||
481 | struct binary_traits <T1, T2, VAR_PRECISION, VAR_PRECISION> | |||
482 | { | |||
483 | typedef wide_int result_type; | |||
484 | typedef result_type operator_result; | |||
485 | typedef bool predicate_result; | |||
486 | }; | |||
487 | } | |||
488 | ||||
489 | /* Public functions for querying and operating on integers. */ | |||
490 | namespace wi | |||
491 | { | |||
492 | template <typename T> | |||
493 | unsigned int get_precision (const T &); | |||
494 | ||||
495 | template <typename T1, typename T2> | |||
496 | unsigned int get_binary_precision (const T1 &, const T2 &); | |||
497 | ||||
498 | template <typename T1, typename T2> | |||
499 | void copy (T1 &, const T2 &); | |||
500 | ||||
501 | #define UNARY_PREDICATE \ | |||
502 | template <typename T> bool | |||
503 | #define UNARY_FUNCTION \ | |||
504 | template <typename T> WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type | |||
505 | #define BINARY_PREDICATE \ | |||
506 | template <typename T1, typename T2> bool | |||
507 | #define BINARY_FUNCTION \ | |||
508 | template <typename T1, typename T2> WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
509 | #define SHIFT_FUNCTION \ | |||
510 | template <typename T1, typename T2> WI_UNARY_RESULT (T1)typename wi::binary_traits <T1, T1>::result_type | |||
511 | ||||
512 | UNARY_PREDICATE fits_shwi_p (const T &); | |||
513 | UNARY_PREDICATE fits_uhwi_p (const T &); | |||
514 | UNARY_PREDICATE neg_p (const T &, signop = SIGNED); | |||
515 | ||||
516 | template <typename T> | |||
517 | HOST_WIDE_INTlong sign_mask (const T &); | |||
518 | ||||
519 | BINARY_PREDICATE eq_p (const T1 &, const T2 &); | |||
520 | BINARY_PREDICATE ne_p (const T1 &, const T2 &); | |||
521 | BINARY_PREDICATE lt_p (const T1 &, const T2 &, signop); | |||
522 | BINARY_PREDICATE lts_p (const T1 &, const T2 &); | |||
523 | BINARY_PREDICATE ltu_p (const T1 &, const T2 &); | |||
524 | BINARY_PREDICATE le_p (const T1 &, const T2 &, signop); | |||
525 | BINARY_PREDICATE les_p (const T1 &, const T2 &); | |||
526 | BINARY_PREDICATE leu_p (const T1 &, const T2 &); | |||
527 | BINARY_PREDICATE gt_p (const T1 &, const T2 &, signop); | |||
528 | BINARY_PREDICATE gts_p (const T1 &, const T2 &); | |||
529 | BINARY_PREDICATE gtu_p (const T1 &, const T2 &); | |||
530 | BINARY_PREDICATE ge_p (const T1 &, const T2 &, signop); | |||
531 | BINARY_PREDICATE ges_p (const T1 &, const T2 &); | |||
532 | BINARY_PREDICATE geu_p (const T1 &, const T2 &); | |||
533 | ||||
534 | template <typename T1, typename T2> | |||
535 | int cmp (const T1 &, const T2 &, signop); | |||
536 | ||||
537 | template <typename T1, typename T2> | |||
538 | int cmps (const T1 &, const T2 &); | |||
539 | ||||
540 | template <typename T1, typename T2> | |||
541 | int cmpu (const T1 &, const T2 &); | |||
542 | ||||
543 | UNARY_FUNCTION bit_not (const T &); | |||
544 | UNARY_FUNCTION neg (const T &); | |||
545 | UNARY_FUNCTION neg (const T &, overflow_type *); | |||
546 | UNARY_FUNCTION abs (const T &); | |||
547 | UNARY_FUNCTION ext (const T &, unsigned int, signop); | |||
548 | UNARY_FUNCTION sext (const T &, unsigned int); | |||
549 | UNARY_FUNCTION zext (const T &, unsigned int); | |||
550 | UNARY_FUNCTION set_bit (const T &, unsigned int); | |||
551 | ||||
552 | BINARY_FUNCTION min (const T1 &, const T2 &, signop); | |||
553 | BINARY_FUNCTION smin (const T1 &, const T2 &); | |||
554 | BINARY_FUNCTION umin (const T1 &, const T2 &); | |||
555 | BINARY_FUNCTION max (const T1 &, const T2 &, signop); | |||
556 | BINARY_FUNCTION smax (const T1 &, const T2 &); | |||
557 | BINARY_FUNCTION umax (const T1 &, const T2 &); | |||
558 | ||||
559 | BINARY_FUNCTION bit_and (const T1 &, const T2 &); | |||
560 | BINARY_FUNCTION bit_and_not (const T1 &, const T2 &); | |||
561 | BINARY_FUNCTION bit_or (const T1 &, const T2 &); | |||
562 | BINARY_FUNCTION bit_or_not (const T1 &, const T2 &); | |||
563 | BINARY_FUNCTION bit_xor (const T1 &, const T2 &); | |||
564 | BINARY_FUNCTION add (const T1 &, const T2 &); | |||
565 | BINARY_FUNCTION add (const T1 &, const T2 &, signop, overflow_type *); | |||
566 | BINARY_FUNCTION sub (const T1 &, const T2 &); | |||
567 | BINARY_FUNCTION sub (const T1 &, const T2 &, signop, overflow_type *); | |||
568 | BINARY_FUNCTION mul (const T1 &, const T2 &); | |||
569 | BINARY_FUNCTION mul (const T1 &, const T2 &, signop, overflow_type *); | |||
570 | BINARY_FUNCTION smul (const T1 &, const T2 &, overflow_type *); | |||
571 | BINARY_FUNCTION umul (const T1 &, const T2 &, overflow_type *); | |||
572 | BINARY_FUNCTION mul_high (const T1 &, const T2 &, signop); | |||
573 | BINARY_FUNCTION div_trunc (const T1 &, const T2 &, signop, | |||
574 | overflow_type * = 0); | |||
575 | BINARY_FUNCTION sdiv_trunc (const T1 &, const T2 &); | |||
576 | BINARY_FUNCTION udiv_trunc (const T1 &, const T2 &); | |||
577 | BINARY_FUNCTION div_floor (const T1 &, const T2 &, signop, | |||
578 | overflow_type * = 0); | |||
579 | BINARY_FUNCTION udiv_floor (const T1 &, const T2 &); | |||
580 | BINARY_FUNCTION sdiv_floor (const T1 &, const T2 &); | |||
581 | BINARY_FUNCTION div_ceil (const T1 &, const T2 &, signop, | |||
582 | overflow_type * = 0); | |||
583 | BINARY_FUNCTION udiv_ceil (const T1 &, const T2 &); | |||
584 | BINARY_FUNCTION div_round (const T1 &, const T2 &, signop, | |||
585 | overflow_type * = 0); | |||
586 | BINARY_FUNCTION divmod_trunc (const T1 &, const T2 &, signop, | |||
587 | WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type *); | |||
588 | BINARY_FUNCTION gcd (const T1 &, const T2 &, signop = UNSIGNED); | |||
589 | BINARY_FUNCTION mod_trunc (const T1 &, const T2 &, signop, | |||
590 | overflow_type * = 0); | |||
591 | BINARY_FUNCTION smod_trunc (const T1 &, const T2 &); | |||
592 | BINARY_FUNCTION umod_trunc (const T1 &, const T2 &); | |||
593 | BINARY_FUNCTION mod_floor (const T1 &, const T2 &, signop, | |||
594 | overflow_type * = 0); | |||
595 | BINARY_FUNCTION umod_floor (const T1 &, const T2 &); | |||
596 | BINARY_FUNCTION mod_ceil (const T1 &, const T2 &, signop, | |||
597 | overflow_type * = 0); | |||
598 | BINARY_FUNCTION mod_round (const T1 &, const T2 &, signop, | |||
599 | overflow_type * = 0); | |||
600 | ||||
601 | template <typename T1, typename T2> | |||
602 | bool multiple_of_p (const T1 &, const T2 &, signop); | |||
603 | ||||
604 | template <typename T1, typename T2> | |||
605 | bool multiple_of_p (const T1 &, const T2 &, signop, | |||
606 | WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type *); | |||
607 | ||||
608 | SHIFT_FUNCTION lshift (const T1 &, const T2 &); | |||
609 | SHIFT_FUNCTION lrshift (const T1 &, const T2 &); | |||
610 | SHIFT_FUNCTION arshift (const T1 &, const T2 &); | |||
611 | SHIFT_FUNCTION rshift (const T1 &, const T2 &, signop sgn); | |||
612 | SHIFT_FUNCTION lrotate (const T1 &, const T2 &, unsigned int = 0); | |||
613 | SHIFT_FUNCTION rrotate (const T1 &, const T2 &, unsigned int = 0); | |||
614 | ||||
615 | #undef SHIFT_FUNCTION | |||
616 | #undef BINARY_PREDICATE | |||
617 | #undef BINARY_FUNCTION | |||
618 | #undef UNARY_PREDICATE | |||
619 | #undef UNARY_FUNCTION | |||
620 | ||||
621 | bool only_sign_bit_p (const wide_int_ref &, unsigned int); | |||
622 | bool only_sign_bit_p (const wide_int_ref &); | |||
623 | int clz (const wide_int_ref &); | |||
624 | int clrsb (const wide_int_ref &); | |||
625 | int ctz (const wide_int_ref &); | |||
626 | int exact_log2 (const wide_int_ref &); | |||
627 | int floor_log2 (const wide_int_ref &); | |||
628 | int ffs (const wide_int_ref &); | |||
629 | int popcount (const wide_int_ref &); | |||
630 | int parity (const wide_int_ref &); | |||
631 | ||||
632 | template <typename T> | |||
633 | unsigned HOST_WIDE_INTlong extract_uhwi (const T &, unsigned int, unsigned int); | |||
634 | ||||
635 | template <typename T> | |||
636 | unsigned int min_precision (const T &, signop); | |||
637 | ||||
638 | static inline void accumulate_overflow (overflow_type &, overflow_type); | |||
639 | } | |||
640 | ||||
641 | namespace wi | |||
642 | { | |||
643 | /* Contains the components of a decomposed integer for easy, direct | |||
644 | access. */ | |||
645 | class storage_ref | |||
646 | { | |||
647 | public: | |||
648 | storage_ref () {} | |||
649 | storage_ref (const HOST_WIDE_INTlong *, unsigned int, unsigned int); | |||
650 | ||||
651 | const HOST_WIDE_INTlong *val; | |||
652 | unsigned int len; | |||
653 | unsigned int precision; | |||
654 | ||||
655 | /* Provide enough trappings for this class to act as storage for | |||
656 | generic_wide_int. */ | |||
657 | unsigned int get_len () const; | |||
658 | unsigned int get_precision () const; | |||
659 | const HOST_WIDE_INTlong *get_val () const; | |||
660 | }; | |||
661 | } | |||
662 | ||||
663 | inline::wi::storage_ref::storage_ref (const HOST_WIDE_INTlong *val_in, | |||
664 | unsigned int len_in, | |||
665 | unsigned int precision_in) | |||
666 | : val (val_in), len (len_in), precision (precision_in) | |||
667 | { | |||
668 | } | |||
669 | ||||
670 | inline unsigned int | |||
671 | wi::storage_ref::get_len () const | |||
672 | { | |||
673 | return len; | |||
674 | } | |||
675 | ||||
676 | inline unsigned int | |||
677 | wi::storage_ref::get_precision () const | |||
678 | { | |||
679 | return precision; | |||
680 | } | |||
681 | ||||
682 | inline const HOST_WIDE_INTlong * | |||
683 | wi::storage_ref::get_val () const | |||
684 | { | |||
685 | return val; | |||
686 | } | |||
687 | ||||
688 | /* This class defines an integer type using the storage provided by the | |||
689 | template argument. The storage class must provide the following | |||
690 | functions: | |||
691 | ||||
692 | unsigned int get_precision () const | |||
693 | Return the number of bits in the integer. | |||
694 | ||||
695 | HOST_WIDE_INT *get_val () const | |||
696 | Return a pointer to the array of blocks that encodes the integer. | |||
697 | ||||
698 | unsigned int get_len () const | |||
699 | Return the number of blocks in get_val (). If this is smaller | |||
700 | than the number of blocks implied by get_precision (), the | |||
701 | remaining blocks are sign extensions of block get_len () - 1. | |||
702 | ||||
703 | Although not required by generic_wide_int itself, writable storage | |||
704 | classes can also provide the following functions: | |||
705 | ||||
706 | HOST_WIDE_INT *write_val () | |||
707 | Get a modifiable version of get_val () | |||
708 | ||||
709 | unsigned int set_len (unsigned int len) | |||
710 | Set the value returned by get_len () to LEN. */ | |||
711 | template <typename storage> | |||
712 | class GTY(()) generic_wide_int : public storage | |||
713 | { | |||
714 | public: | |||
715 | generic_wide_int (); | |||
716 | ||||
717 | template <typename T> | |||
718 | generic_wide_int (const T &); | |||
719 | ||||
720 | template <typename T> | |||
721 | generic_wide_int (const T &, unsigned int); | |||
722 | ||||
723 | /* Conversions. */ | |||
724 | HOST_WIDE_INTlong to_shwi (unsigned int) const; | |||
725 | HOST_WIDE_INTlong to_shwi () const; | |||
726 | unsigned HOST_WIDE_INTlong to_uhwi (unsigned int) const; | |||
727 | unsigned HOST_WIDE_INTlong to_uhwi () const; | |||
728 | HOST_WIDE_INTlong to_short_addr () const; | |||
729 | ||||
730 | /* Public accessors for the interior of a wide int. */ | |||
731 | HOST_WIDE_INTlong sign_mask () const; | |||
732 | HOST_WIDE_INTlong elt (unsigned int) const; | |||
733 | HOST_WIDE_INTlong sext_elt (unsigned int) const; | |||
734 | unsigned HOST_WIDE_INTlong ulow () const; | |||
735 | unsigned HOST_WIDE_INTlong uhigh () const; | |||
736 | HOST_WIDE_INTlong slow () const; | |||
737 | HOST_WIDE_INTlong shigh () const; | |||
738 | ||||
739 | template <typename T> | |||
740 | generic_wide_int &operator = (const T &); | |||
741 | ||||
742 | #define ASSIGNMENT_OPERATOR(OP, F) \ | |||
743 | template <typename T> \ | |||
744 | generic_wide_int &OP (const T &c) { return (*this = wi::F (*this, c)); } | |||
745 | ||||
746 | /* Restrict these to cases where the shift operator is defined. */ | |||
747 | #define SHIFT_ASSIGNMENT_OPERATOR(OP, OP2) \ | |||
748 | template <typename T> \ | |||
749 | generic_wide_int &OP (const T &c) { return (*this = *this OP2 c); } | |||
750 | ||||
751 | #define INCDEC_OPERATOR(OP, DELTA) \ | |||
752 | generic_wide_int &OP () { *this += DELTA; return *this; } | |||
753 | ||||
754 | ASSIGNMENT_OPERATOR (operator &=, bit_and) | |||
755 | ASSIGNMENT_OPERATOR (operator |=, bit_or) | |||
756 | ASSIGNMENT_OPERATOR (operator ^=, bit_xor) | |||
757 | ASSIGNMENT_OPERATOR (operator +=, add) | |||
758 | ASSIGNMENT_OPERATOR (operator -=, sub) | |||
759 | ASSIGNMENT_OPERATOR (operator *=, mul) | |||
760 | ASSIGNMENT_OPERATOR (operator <<=, lshift) | |||
761 | SHIFT_ASSIGNMENT_OPERATOR (operator >>=, >>) | |||
762 | INCDEC_OPERATOR (operator ++, 1) | |||
763 | INCDEC_OPERATOR (operator --, -1) | |||
764 | ||||
765 | #undef SHIFT_ASSIGNMENT_OPERATOR | |||
766 | #undef ASSIGNMENT_OPERATOR | |||
767 | #undef INCDEC_OPERATOR | |||
768 | ||||
769 | /* Debugging functions. */ | |||
770 | void dump () const; | |||
771 | ||||
772 | static const bool is_sign_extended | |||
773 | = wi::int_traits <generic_wide_int <storage> >::is_sign_extended; | |||
774 | }; | |||
775 | ||||
776 | template <typename storage> | |||
777 | inline generic_wide_int <storage>::generic_wide_int () {} | |||
778 | ||||
779 | template <typename storage> | |||
780 | template <typename T> | |||
781 | inline generic_wide_int <storage>::generic_wide_int (const T &x) | |||
782 | : storage (x) | |||
783 | { | |||
784 | } | |||
785 | ||||
786 | template <typename storage> | |||
787 | template <typename T> | |||
788 | inline generic_wide_int <storage>::generic_wide_int (const T &x, | |||
789 | unsigned int precision) | |||
790 | : storage (x, precision) | |||
791 | { | |||
792 | } | |||
793 | ||||
794 | /* Return THIS as a signed HOST_WIDE_INT, sign-extending from PRECISION. | |||
795 | If THIS does not fit in PRECISION, the information is lost. */ | |||
796 | template <typename storage> | |||
797 | inline HOST_WIDE_INTlong | |||
798 | generic_wide_int <storage>::to_shwi (unsigned int precision) const | |||
799 | { | |||
800 | if (precision < HOST_BITS_PER_WIDE_INT64) | |||
801 | return sext_hwi (this->get_val ()[0], precision); | |||
802 | else | |||
803 | return this->get_val ()[0]; | |||
804 | } | |||
805 | ||||
806 | /* Return THIS as a signed HOST_WIDE_INT, in its natural precision. */ | |||
807 | template <typename storage> | |||
808 | inline HOST_WIDE_INTlong | |||
809 | generic_wide_int <storage>::to_shwi () const | |||
810 | { | |||
811 | if (is_sign_extended) | |||
812 | return this->get_val ()[0]; | |||
813 | else | |||
814 | return to_shwi (this->get_precision ()); | |||
815 | } | |||
816 | ||||
817 | /* Return THIS as an unsigned HOST_WIDE_INT, zero-extending from | |||
818 | PRECISION. If THIS does not fit in PRECISION, the information | |||
819 | is lost. */ | |||
820 | template <typename storage> | |||
821 | inline unsigned HOST_WIDE_INTlong | |||
822 | generic_wide_int <storage>::to_uhwi (unsigned int precision) const | |||
823 | { | |||
824 | if (precision < HOST_BITS_PER_WIDE_INT64) | |||
825 | return zext_hwi (this->get_val ()[0], precision); | |||
826 | else | |||
827 | return this->get_val ()[0]; | |||
828 | } | |||
829 | ||||
830 | /* Return THIS as an signed HOST_WIDE_INT, in its natural precision. */ | |||
831 | template <typename storage> | |||
832 | inline unsigned HOST_WIDE_INTlong | |||
833 | generic_wide_int <storage>::to_uhwi () const | |||
834 | { | |||
835 | return to_uhwi (this->get_precision ()); | |||
836 | } | |||
837 | ||||
838 | /* TODO: The compiler is half converted from using HOST_WIDE_INT to | |||
839 | represent addresses to using offset_int to represent addresses. | |||
840 | We use to_short_addr at the interface from new code to old, | |||
841 | unconverted code. */ | |||
842 | template <typename storage> | |||
843 | inline HOST_WIDE_INTlong | |||
844 | generic_wide_int <storage>::to_short_addr () const | |||
845 | { | |||
846 | return this->get_val ()[0]; | |||
847 | } | |||
848 | ||||
849 | /* Return the implicit value of blocks above get_len (). */ | |||
850 | template <typename storage> | |||
851 | inline HOST_WIDE_INTlong | |||
852 | generic_wide_int <storage>::sign_mask () const | |||
853 | { | |||
854 | unsigned int len = this->get_len (); | |||
855 | gcc_assert (len > 0)((void)(!(len > 0) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/wide-int.h" , 855, __FUNCTION__), 0 : 0)); | |||
856 | ||||
857 | unsigned HOST_WIDE_INTlong high = this->get_val ()[len - 1]; | |||
858 | if (!is_sign_extended) | |||
859 | { | |||
860 | unsigned int precision = this->get_precision (); | |||
861 | int excess = len * HOST_BITS_PER_WIDE_INT64 - precision; | |||
862 | if (excess > 0) | |||
863 | high <<= excess; | |||
864 | } | |||
865 | return (HOST_WIDE_INTlong) (high) < 0 ? -1 : 0; | |||
866 | } | |||
867 | ||||
868 | /* Return the signed value of the least-significant explicitly-encoded | |||
869 | block. */ | |||
870 | template <typename storage> | |||
871 | inline HOST_WIDE_INTlong | |||
872 | generic_wide_int <storage>::slow () const | |||
873 | { | |||
874 | return this->get_val ()[0]; | |||
875 | } | |||
876 | ||||
877 | /* Return the signed value of the most-significant explicitly-encoded | |||
878 | block. */ | |||
879 | template <typename storage> | |||
880 | inline HOST_WIDE_INTlong | |||
881 | generic_wide_int <storage>::shigh () const | |||
882 | { | |||
883 | return this->get_val ()[this->get_len () - 1]; | |||
884 | } | |||
885 | ||||
886 | /* Return the unsigned value of the least-significant | |||
887 | explicitly-encoded block. */ | |||
888 | template <typename storage> | |||
889 | inline unsigned HOST_WIDE_INTlong | |||
890 | generic_wide_int <storage>::ulow () const | |||
891 | { | |||
892 | return this->get_val ()[0]; | |||
893 | } | |||
894 | ||||
895 | /* Return the unsigned value of the most-significant | |||
896 | explicitly-encoded block. */ | |||
897 | template <typename storage> | |||
898 | inline unsigned HOST_WIDE_INTlong | |||
899 | generic_wide_int <storage>::uhigh () const | |||
900 | { | |||
901 | return this->get_val ()[this->get_len () - 1]; | |||
902 | } | |||
903 | ||||
904 | /* Return block I, which might be implicitly or explicit encoded. */ | |||
905 | template <typename storage> | |||
906 | inline HOST_WIDE_INTlong | |||
907 | generic_wide_int <storage>::elt (unsigned int i) const | |||
908 | { | |||
909 | if (i >= this->get_len ()) | |||
910 | return sign_mask (); | |||
911 | else | |||
912 | return this->get_val ()[i]; | |||
913 | } | |||
914 | ||||
915 | /* Like elt, but sign-extend beyond the upper bit, instead of returning | |||
916 | the raw encoding. */ | |||
917 | template <typename storage> | |||
918 | inline HOST_WIDE_INTlong | |||
919 | generic_wide_int <storage>::sext_elt (unsigned int i) const | |||
920 | { | |||
921 | HOST_WIDE_INTlong elt_i = elt (i); | |||
922 | if (!is_sign_extended) | |||
923 | { | |||
924 | unsigned int precision = this->get_precision (); | |||
925 | unsigned int lsb = i * HOST_BITS_PER_WIDE_INT64; | |||
926 | if (precision - lsb < HOST_BITS_PER_WIDE_INT64) | |||
927 | elt_i = sext_hwi (elt_i, precision - lsb); | |||
928 | } | |||
929 | return elt_i; | |||
930 | } | |||
931 | ||||
932 | template <typename storage> | |||
933 | template <typename T> | |||
934 | inline generic_wide_int <storage> & | |||
935 | generic_wide_int <storage>::operator = (const T &x) | |||
936 | { | |||
937 | storage::operator = (x); | |||
938 | return *this; | |||
939 | } | |||
940 | ||||
941 | /* Dump the contents of the integer to stderr, for debugging. */ | |||
942 | template <typename storage> | |||
943 | void | |||
944 | generic_wide_int <storage>::dump () const | |||
945 | { | |||
946 | unsigned int len = this->get_len (); | |||
947 | const HOST_WIDE_INTlong *val = this->get_val (); | |||
948 | unsigned int precision = this->get_precision (); | |||
949 | fprintf (stderrstderr, "["); | |||
950 | if (len * HOST_BITS_PER_WIDE_INT64 < precision) | |||
951 | fprintf (stderrstderr, "...,"); | |||
952 | for (unsigned int i = 0; i < len - 1; ++i) | |||
953 | fprintf (stderrstderr, HOST_WIDE_INT_PRINT_HEX"%#" "l" "x" ",", val[len - 1 - i]); | |||
| ||||
954 | fprintf (stderrstderr, HOST_WIDE_INT_PRINT_HEX"%#" "l" "x" "], precision = %d\n", | |||
955 | val[0], precision); | |||
956 | } | |||
957 | ||||
958 | namespace wi | |||
959 | { | |||
960 | template <typename storage> | |||
961 | struct int_traits < generic_wide_int <storage> > | |||
962 | : public wi::int_traits <storage> | |||
963 | { | |||
964 | static unsigned int get_precision (const generic_wide_int <storage> &); | |||
965 | static wi::storage_ref decompose (HOST_WIDE_INTlong *, unsigned int, | |||
966 | const generic_wide_int <storage> &); | |||
967 | }; | |||
968 | } | |||
969 | ||||
970 | template <typename storage> | |||
971 | inline unsigned int | |||
972 | wi::int_traits < generic_wide_int <storage> >:: | |||
973 | get_precision (const generic_wide_int <storage> &x) | |||
974 | { | |||
975 | return x.get_precision (); | |||
976 | } | |||
977 | ||||
978 | template <typename storage> | |||
979 | inline wi::storage_ref | |||
980 | wi::int_traits < generic_wide_int <storage> >:: | |||
981 | decompose (HOST_WIDE_INTlong *, unsigned int precision, | |||
982 | const generic_wide_int <storage> &x) | |||
983 | { | |||
984 | gcc_checking_assert (precision == x.get_precision ())((void)(!(precision == x.get_precision ()) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/wide-int.h" , 984, __FUNCTION__), 0 : 0)); | |||
985 | return wi::storage_ref (x.get_val (), x.get_len (), precision); | |||
986 | } | |||
987 | ||||
988 | /* Provide the storage for a wide_int_ref. This acts like a read-only | |||
989 | wide_int, with the optimization that VAL is normally a pointer to | |||
990 | another integer's storage, so that no array copy is needed. */ | |||
991 | template <bool SE, bool HDP> | |||
992 | class wide_int_ref_storage : public wi::storage_ref | |||
993 | { | |||
994 | private: | |||
995 | /* Scratch space that can be used when decomposing the original integer. | |||
996 | It must live as long as this object. */ | |||
997 | HOST_WIDE_INTlong scratch[2]; | |||
998 | ||||
999 | public: | |||
1000 | wide_int_ref_storage () {} | |||
1001 | ||||
1002 | wide_int_ref_storage (const wi::storage_ref &); | |||
1003 | ||||
1004 | template <typename T> | |||
1005 | wide_int_ref_storage (const T &); | |||
1006 | ||||
1007 | template <typename T> | |||
1008 | wide_int_ref_storage (const T &, unsigned int); | |||
1009 | }; | |||
1010 | ||||
1011 | /* Create a reference from an existing reference. */ | |||
1012 | template <bool SE, bool HDP> | |||
1013 | inline wide_int_ref_storage <SE, HDP>:: | |||
1014 | wide_int_ref_storage (const wi::storage_ref &x) | |||
1015 | : storage_ref (x) | |||
1016 | {} | |||
1017 | ||||
1018 | /* Create a reference to integer X in its natural precision. Note | |||
1019 | that the natural precision is host-dependent for primitive | |||
1020 | types. */ | |||
1021 | template <bool SE, bool HDP> | |||
1022 | template <typename T> | |||
1023 | inline wide_int_ref_storage <SE, HDP>::wide_int_ref_storage (const T &x) | |||
1024 | : storage_ref (wi::int_traits <T>::decompose (scratch, | |||
1025 | wi::get_precision (x), x)) | |||
1026 | { | |||
1027 | } | |||
1028 | ||||
1029 | /* Create a reference to integer X in precision PRECISION. */ | |||
1030 | template <bool SE, bool HDP> | |||
1031 | template <typename T> | |||
1032 | inline wide_int_ref_storage <SE, HDP>:: | |||
1033 | wide_int_ref_storage (const T &x, unsigned int precision) | |||
1034 | : storage_ref (wi::int_traits <T>::decompose (scratch, precision, x)) | |||
1035 | { | |||
1036 | } | |||
1037 | ||||
1038 | namespace wi | |||
1039 | { | |||
1040 | template <bool SE, bool HDP> | |||
1041 | struct int_traits <wide_int_ref_storage <SE, HDP> > | |||
1042 | { | |||
1043 | static const enum precision_type precision_type = VAR_PRECISION; | |||
1044 | static const bool host_dependent_precision = HDP; | |||
1045 | static const bool is_sign_extended = SE; | |||
1046 | }; | |||
1047 | } | |||
1048 | ||||
1049 | namespace wi | |||
1050 | { | |||
1051 | unsigned int force_to_size (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1052 | unsigned int, unsigned int, unsigned int, | |||
1053 | signop sgn); | |||
1054 | unsigned int from_array (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1055 | unsigned int, unsigned int, bool = true); | |||
1056 | } | |||
1057 | ||||
1058 | /* The storage used by wide_int. */ | |||
1059 | class GTY(()) wide_int_storage | |||
1060 | { | |||
1061 | private: | |||
1062 | HOST_WIDE_INTlong val[WIDE_INT_MAX_ELTS(((64*(8)) + 64) / 64)]; | |||
1063 | unsigned int len; | |||
1064 | unsigned int precision; | |||
1065 | ||||
1066 | public: | |||
1067 | wide_int_storage (); | |||
1068 | template <typename T> | |||
1069 | wide_int_storage (const T &); | |||
1070 | ||||
1071 | /* The standard generic_wide_int storage methods. */ | |||
1072 | unsigned int get_precision () const; | |||
1073 | const HOST_WIDE_INTlong *get_val () const; | |||
1074 | unsigned int get_len () const; | |||
1075 | HOST_WIDE_INTlong *write_val (); | |||
1076 | void set_len (unsigned int, bool = false); | |||
1077 | ||||
1078 | template <typename T> | |||
1079 | wide_int_storage &operator = (const T &); | |||
1080 | ||||
1081 | static wide_int from (const wide_int_ref &, unsigned int, signop); | |||
1082 | static wide_int from_array (const HOST_WIDE_INTlong *, unsigned int, | |||
1083 | unsigned int, bool = true); | |||
1084 | static wide_int create (unsigned int); | |||
1085 | ||||
1086 | /* FIXME: target-dependent, so should disappear. */ | |||
1087 | wide_int bswap () const; | |||
1088 | }; | |||
1089 | ||||
1090 | namespace wi | |||
1091 | { | |||
1092 | template <> | |||
1093 | struct int_traits <wide_int_storage> | |||
1094 | { | |||
1095 | static const enum precision_type precision_type = VAR_PRECISION; | |||
1096 | /* Guaranteed by a static assert in the wide_int_storage constructor. */ | |||
1097 | static const bool host_dependent_precision = false; | |||
1098 | static const bool is_sign_extended = true; | |||
1099 | template <typename T1, typename T2> | |||
1100 | static wide_int get_binary_result (const T1 &, const T2 &); | |||
1101 | }; | |||
1102 | } | |||
1103 | ||||
1104 | inline wide_int_storage::wide_int_storage () {} | |||
1105 | ||||
1106 | /* Initialize the storage from integer X, in its natural precision. | |||
1107 | Note that we do not allow integers with host-dependent precision | |||
1108 | to become wide_ints; wide_ints must always be logically independent | |||
1109 | of the host. */ | |||
1110 | template <typename T> | |||
1111 | inline wide_int_storage::wide_int_storage (const T &x) | |||
1112 | { | |||
1113 | { STATIC_ASSERT (!wi::int_traits<T>::host_dependent_precision)static_assert ((!wi::int_traits<T>::host_dependent_precision ), "!wi::int_traits<T>::host_dependent_precision"); } | |||
1114 | { STATIC_ASSERT (wi::int_traits<T>::precision_type != wi::CONST_PRECISION)static_assert ((wi::int_traits<T>::precision_type != wi ::CONST_PRECISION), "wi::int_traits<T>::precision_type != wi::CONST_PRECISION" ); } | |||
1115 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x); | |||
1116 | precision = xi.precision; | |||
1117 | wi::copy (*this, xi); | |||
1118 | } | |||
1119 | ||||
1120 | template <typename T> | |||
1121 | inline wide_int_storage& | |||
1122 | wide_int_storage::operator = (const T &x) | |||
1123 | { | |||
1124 | { STATIC_ASSERT (!wi::int_traits<T>::host_dependent_precision)static_assert ((!wi::int_traits<T>::host_dependent_precision ), "!wi::int_traits<T>::host_dependent_precision"); } | |||
1125 | { STATIC_ASSERT (wi::int_traits<T>::precision_type != wi::CONST_PRECISION)static_assert ((wi::int_traits<T>::precision_type != wi ::CONST_PRECISION), "wi::int_traits<T>::precision_type != wi::CONST_PRECISION" ); } | |||
1126 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x); | |||
1127 | precision = xi.precision; | |||
1128 | wi::copy (*this, xi); | |||
1129 | return *this; | |||
1130 | } | |||
1131 | ||||
1132 | inline unsigned int | |||
1133 | wide_int_storage::get_precision () const | |||
1134 | { | |||
1135 | return precision; | |||
1136 | } | |||
1137 | ||||
1138 | inline const HOST_WIDE_INTlong * | |||
1139 | wide_int_storage::get_val () const | |||
1140 | { | |||
1141 | return val; | |||
1142 | } | |||
1143 | ||||
1144 | inline unsigned int | |||
1145 | wide_int_storage::get_len () const | |||
1146 | { | |||
1147 | return len; | |||
1148 | } | |||
1149 | ||||
1150 | inline HOST_WIDE_INTlong * | |||
1151 | wide_int_storage::write_val () | |||
1152 | { | |||
1153 | return val; | |||
1154 | } | |||
1155 | ||||
1156 | inline void | |||
1157 | wide_int_storage::set_len (unsigned int l, bool is_sign_extended) | |||
1158 | { | |||
1159 | len = l; | |||
1160 | if (!is_sign_extended && len * HOST_BITS_PER_WIDE_INT64 > precision) | |||
1161 | val[len - 1] = sext_hwi (val[len - 1], | |||
1162 | precision % HOST_BITS_PER_WIDE_INT64); | |||
1163 | } | |||
1164 | ||||
1165 | /* Treat X as having signedness SGN and convert it to a PRECISION-bit | |||
1166 | number. */ | |||
1167 | inline wide_int | |||
1168 | wide_int_storage::from (const wide_int_ref &x, unsigned int precision, | |||
1169 | signop sgn) | |||
1170 | { | |||
1171 | wide_int result = wide_int::create (precision); | |||
1172 | result.set_len (wi::force_to_size (result.write_val (), x.val, x.len, | |||
1173 | x.precision, precision, sgn)); | |||
1174 | return result; | |||
1175 | } | |||
1176 | ||||
1177 | /* Create a wide_int from the explicit block encoding given by VAL and | |||
1178 | LEN. PRECISION is the precision of the integer. NEED_CANON_P is | |||
1179 | true if the encoding may have redundant trailing blocks. */ | |||
1180 | inline wide_int | |||
1181 | wide_int_storage::from_array (const HOST_WIDE_INTlong *val, unsigned int len, | |||
1182 | unsigned int precision, bool need_canon_p) | |||
1183 | { | |||
1184 | wide_int result = wide_int::create (precision); | |||
1185 | result.set_len (wi::from_array (result.write_val (), val, len, precision, | |||
1186 | need_canon_p)); | |||
1187 | return result; | |||
1188 | } | |||
1189 | ||||
1190 | /* Return an uninitialized wide_int with precision PRECISION. */ | |||
1191 | inline wide_int | |||
1192 | wide_int_storage::create (unsigned int precision) | |||
1193 | { | |||
1194 | wide_int x; | |||
1195 | x.precision = precision; | |||
1196 | return x; | |||
1197 | } | |||
1198 | ||||
1199 | template <typename T1, typename T2> | |||
1200 | inline wide_int | |||
1201 | wi::int_traits <wide_int_storage>::get_binary_result (const T1 &x, const T2 &y) | |||
1202 | { | |||
1203 | /* This shouldn't be used for two flexible-precision inputs. */ | |||
1204 | STATIC_ASSERT (wi::int_traits <T1>::precision_type != FLEXIBLE_PRECISIONstatic_assert ((wi::int_traits <T1>::precision_type != FLEXIBLE_PRECISION || wi::int_traits <T2>::precision_type != FLEXIBLE_PRECISION ), "wi::int_traits <T1>::precision_type != FLEXIBLE_PRECISION || wi::int_traits <T2>::precision_type != FLEXIBLE_PRECISION" ) | |||
1205 | || wi::int_traits <T2>::precision_type != FLEXIBLE_PRECISION)static_assert ((wi::int_traits <T1>::precision_type != FLEXIBLE_PRECISION || wi::int_traits <T2>::precision_type != FLEXIBLE_PRECISION ), "wi::int_traits <T1>::precision_type != FLEXIBLE_PRECISION || wi::int_traits <T2>::precision_type != FLEXIBLE_PRECISION" ); | |||
1206 | if (wi::int_traits <T1>::precision_type == FLEXIBLE_PRECISION) | |||
1207 | return wide_int::create (wi::get_precision (y)); | |||
1208 | else | |||
1209 | return wide_int::create (wi::get_precision (x)); | |||
1210 | } | |||
1211 | ||||
1212 | /* The storage used by FIXED_WIDE_INT (N). */ | |||
1213 | template <int N> | |||
1214 | class GTY(()) fixed_wide_int_storage | |||
1215 | { | |||
1216 | private: | |||
1217 | HOST_WIDE_INTlong val[(N + HOST_BITS_PER_WIDE_INT64 + 1) / HOST_BITS_PER_WIDE_INT64]; | |||
1218 | unsigned int len; | |||
1219 | ||||
1220 | public: | |||
1221 | fixed_wide_int_storage (); | |||
1222 | template <typename T> | |||
1223 | fixed_wide_int_storage (const T &); | |||
1224 | ||||
1225 | /* The standard generic_wide_int storage methods. */ | |||
1226 | unsigned int get_precision () const; | |||
1227 | const HOST_WIDE_INTlong *get_val () const; | |||
1228 | unsigned int get_len () const; | |||
1229 | HOST_WIDE_INTlong *write_val (); | |||
1230 | void set_len (unsigned int, bool = false); | |||
1231 | ||||
1232 | static FIXED_WIDE_INT (N)generic_wide_int < fixed_wide_int_storage <N> > from (const wide_int_ref &, signop); | |||
1233 | static FIXED_WIDE_INT (N)generic_wide_int < fixed_wide_int_storage <N> > from_array (const HOST_WIDE_INTlong *, unsigned int, | |||
1234 | bool = true); | |||
1235 | }; | |||
1236 | ||||
1237 | namespace wi | |||
1238 | { | |||
1239 | template <int N> | |||
1240 | struct int_traits < fixed_wide_int_storage <N> > | |||
1241 | { | |||
1242 | static const enum precision_type precision_type = CONST_PRECISION; | |||
1243 | static const bool host_dependent_precision = false; | |||
1244 | static const bool is_sign_extended = true; | |||
1245 | static const unsigned int precision = N; | |||
1246 | template <typename T1, typename T2> | |||
1247 | static FIXED_WIDE_INT (N)generic_wide_int < fixed_wide_int_storage <N> > get_binary_result (const T1 &, const T2 &); | |||
1248 | }; | |||
1249 | } | |||
1250 | ||||
1251 | template <int N> | |||
1252 | inline fixed_wide_int_storage <N>::fixed_wide_int_storage () {} | |||
1253 | ||||
1254 | /* Initialize the storage from integer X, in precision N. */ | |||
1255 | template <int N> | |||
1256 | template <typename T> | |||
1257 | inline fixed_wide_int_storage <N>::fixed_wide_int_storage (const T &x) | |||
1258 | { | |||
1259 | /* Check for type compatibility. We don't want to initialize a | |||
1260 | fixed-width integer from something like a wide_int. */ | |||
1261 | WI_BINARY_RESULT (T, FIXED_WIDE_INT (N))typename wi::binary_traits <T, generic_wide_int < fixed_wide_int_storage <N> > >::result_type *assertion ATTRIBUTE_UNUSED__attribute__ ((__unused__)); | |||
1262 | wi::copy (*this, WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > (x, N)); | |||
1263 | } | |||
1264 | ||||
1265 | template <int N> | |||
1266 | inline unsigned int | |||
1267 | fixed_wide_int_storage <N>::get_precision () const | |||
1268 | { | |||
1269 | return N; | |||
1270 | } | |||
1271 | ||||
1272 | template <int N> | |||
1273 | inline const HOST_WIDE_INTlong * | |||
1274 | fixed_wide_int_storage <N>::get_val () const | |||
1275 | { | |||
1276 | return val; | |||
1277 | } | |||
1278 | ||||
1279 | template <int N> | |||
1280 | inline unsigned int | |||
1281 | fixed_wide_int_storage <N>::get_len () const | |||
1282 | { | |||
1283 | return len; | |||
1284 | } | |||
1285 | ||||
1286 | template <int N> | |||
1287 | inline HOST_WIDE_INTlong * | |||
1288 | fixed_wide_int_storage <N>::write_val () | |||
1289 | { | |||
1290 | return val; | |||
1291 | } | |||
1292 | ||||
1293 | template <int N> | |||
1294 | inline void | |||
1295 | fixed_wide_int_storage <N>::set_len (unsigned int l, bool) | |||
1296 | { | |||
1297 | len = l; | |||
1298 | /* There are no excess bits in val[len - 1]. */ | |||
1299 | STATIC_ASSERT (N % HOST_BITS_PER_WIDE_INT == 0)static_assert ((N % 64 == 0), "N % HOST_BITS_PER_WIDE_INT == 0" ); | |||
1300 | } | |||
1301 | ||||
1302 | /* Treat X as having signedness SGN and convert it to an N-bit number. */ | |||
1303 | template <int N> | |||
1304 | inline FIXED_WIDE_INT (N)generic_wide_int < fixed_wide_int_storage <N> > | |||
1305 | fixed_wide_int_storage <N>::from (const wide_int_ref &x, signop sgn) | |||
1306 | { | |||
1307 | FIXED_WIDE_INT (N)generic_wide_int < fixed_wide_int_storage <N> > result; | |||
1308 | result.set_len (wi::force_to_size (result.write_val (), x.val, x.len, | |||
1309 | x.precision, N, sgn)); | |||
1310 | return result; | |||
1311 | } | |||
1312 | ||||
1313 | /* Create a FIXED_WIDE_INT (N) from the explicit block encoding given by | |||
1314 | VAL and LEN. NEED_CANON_P is true if the encoding may have redundant | |||
1315 | trailing blocks. */ | |||
1316 | template <int N> | |||
1317 | inline FIXED_WIDE_INT (N)generic_wide_int < fixed_wide_int_storage <N> > | |||
1318 | fixed_wide_int_storage <N>::from_array (const HOST_WIDE_INTlong *val, | |||
1319 | unsigned int len, | |||
1320 | bool need_canon_p) | |||
1321 | { | |||
1322 | FIXED_WIDE_INT (N)generic_wide_int < fixed_wide_int_storage <N> > result; | |||
1323 | result.set_len (wi::from_array (result.write_val (), val, len, | |||
1324 | N, need_canon_p)); | |||
1325 | return result; | |||
1326 | } | |||
1327 | ||||
1328 | template <int N> | |||
1329 | template <typename T1, typename T2> | |||
1330 | inline FIXED_WIDE_INT (N)generic_wide_int < fixed_wide_int_storage <N> > | |||
1331 | wi::int_traits < fixed_wide_int_storage <N> >:: | |||
1332 | get_binary_result (const T1 &, const T2 &) | |||
1333 | { | |||
1334 | return FIXED_WIDE_INT (N)generic_wide_int < fixed_wide_int_storage <N> > (); | |||
1335 | } | |||
1336 | ||||
1337 | /* A reference to one element of a trailing_wide_ints structure. */ | |||
1338 | class trailing_wide_int_storage | |||
1339 | { | |||
1340 | private: | |||
1341 | /* The precision of the integer, which is a fixed property of the | |||
1342 | parent trailing_wide_ints. */ | |||
1343 | unsigned int m_precision; | |||
1344 | ||||
1345 | /* A pointer to the length field. */ | |||
1346 | unsigned char *m_len; | |||
1347 | ||||
1348 | /* A pointer to the HWI array. There are enough elements to hold all | |||
1349 | values of precision M_PRECISION. */ | |||
1350 | HOST_WIDE_INTlong *m_val; | |||
1351 | ||||
1352 | public: | |||
1353 | trailing_wide_int_storage (unsigned int, unsigned char *, HOST_WIDE_INTlong *); | |||
1354 | ||||
1355 | /* The standard generic_wide_int storage methods. */ | |||
1356 | unsigned int get_len () const; | |||
1357 | unsigned int get_precision () const; | |||
1358 | const HOST_WIDE_INTlong *get_val () const; | |||
1359 | HOST_WIDE_INTlong *write_val (); | |||
1360 | void set_len (unsigned int, bool = false); | |||
1361 | ||||
1362 | template <typename T> | |||
1363 | trailing_wide_int_storage &operator = (const T &); | |||
1364 | }; | |||
1365 | ||||
1366 | typedef generic_wide_int <trailing_wide_int_storage> trailing_wide_int; | |||
1367 | ||||
1368 | /* trailing_wide_int behaves like a wide_int. */ | |||
1369 | namespace wi | |||
1370 | { | |||
1371 | template <> | |||
1372 | struct int_traits <trailing_wide_int_storage> | |||
1373 | : public int_traits <wide_int_storage> {}; | |||
1374 | } | |||
1375 | ||||
1376 | /* A variable-length array of wide_int-like objects that can be put | |||
1377 | at the end of a variable-sized structure. The number of objects is | |||
1378 | at most N and can be set at runtime by using set_precision(). | |||
1379 | ||||
1380 | Use extra_size to calculate how many bytes beyond the | |||
1381 | sizeof need to be allocated. Use set_precision to initialize the | |||
1382 | structure. */ | |||
1383 | template <int N> | |||
1384 | struct GTY((user)) trailing_wide_ints | |||
1385 | { | |||
1386 | private: | |||
1387 | /* The shared precision of each number. */ | |||
1388 | unsigned short m_precision; | |||
1389 | ||||
1390 | /* The shared maximum length of each number. */ | |||
1391 | unsigned char m_max_len; | |||
1392 | ||||
1393 | /* The number of elements. */ | |||
1394 | unsigned char m_num_elements; | |||
1395 | ||||
1396 | /* The current length of each number. | |||
1397 | Avoid char array so the whole structure is not a typeless storage | |||
1398 | that will, in turn, turn off TBAA on gimple, trees and RTL. */ | |||
1399 | struct {unsigned char len;} m_len[N]; | |||
1400 | ||||
1401 | /* The variable-length part of the structure, which always contains | |||
1402 | at least one HWI. Element I starts at index I * M_MAX_LEN. */ | |||
1403 | HOST_WIDE_INTlong m_val[1]; | |||
1404 | ||||
1405 | public: | |||
1406 | typedef WIDE_INT_REF_FOR (trailing_wide_int_storage)generic_wide_int <wide_int_ref_storage <wi::int_traits < trailing_wide_int_storage>::is_sign_extended, wi::int_traits <trailing_wide_int_storage>::host_dependent_precision> > const_reference; | |||
1407 | ||||
1408 | void set_precision (unsigned int precision, unsigned int num_elements = N); | |||
1409 | unsigned int get_precision () const { return m_precision; } | |||
1410 | unsigned int num_elements () const { return m_num_elements; } | |||
1411 | trailing_wide_int operator [] (unsigned int); | |||
1412 | const_reference operator [] (unsigned int) const; | |||
1413 | static size_t extra_size (unsigned int precision, | |||
1414 | unsigned int num_elements = N); | |||
1415 | size_t extra_size () const { return extra_size (m_precision, | |||
1416 | m_num_elements); } | |||
1417 | }; | |||
1418 | ||||
1419 | inline trailing_wide_int_storage:: | |||
1420 | trailing_wide_int_storage (unsigned int precision, unsigned char *len, | |||
1421 | HOST_WIDE_INTlong *val) | |||
1422 | : m_precision (precision), m_len (len), m_val (val) | |||
1423 | { | |||
1424 | } | |||
1425 | ||||
1426 | inline unsigned int | |||
1427 | trailing_wide_int_storage::get_len () const | |||
1428 | { | |||
1429 | return *m_len; | |||
1430 | } | |||
1431 | ||||
1432 | inline unsigned int | |||
1433 | trailing_wide_int_storage::get_precision () const | |||
1434 | { | |||
1435 | return m_precision; | |||
1436 | } | |||
1437 | ||||
1438 | inline const HOST_WIDE_INTlong * | |||
1439 | trailing_wide_int_storage::get_val () const | |||
1440 | { | |||
1441 | return m_val; | |||
1442 | } | |||
1443 | ||||
1444 | inline HOST_WIDE_INTlong * | |||
1445 | trailing_wide_int_storage::write_val () | |||
1446 | { | |||
1447 | return m_val; | |||
1448 | } | |||
1449 | ||||
1450 | inline void | |||
1451 | trailing_wide_int_storage::set_len (unsigned int len, bool is_sign_extended) | |||
1452 | { | |||
1453 | *m_len = len; | |||
1454 | if (!is_sign_extended && len * HOST_BITS_PER_WIDE_INT64 > m_precision) | |||
1455 | m_val[len - 1] = sext_hwi (m_val[len - 1], | |||
1456 | m_precision % HOST_BITS_PER_WIDE_INT64); | |||
1457 | } | |||
1458 | ||||
1459 | template <typename T> | |||
1460 | inline trailing_wide_int_storage & | |||
1461 | trailing_wide_int_storage::operator = (const T &x) | |||
1462 | { | |||
1463 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x, m_precision); | |||
1464 | wi::copy (*this, xi); | |||
1465 | return *this; | |||
1466 | } | |||
1467 | ||||
1468 | /* Initialize the structure and record that all elements have precision | |||
1469 | PRECISION. NUM_ELEMENTS can be no more than N. */ | |||
1470 | template <int N> | |||
1471 | inline void | |||
1472 | trailing_wide_ints <N>::set_precision (unsigned int precision, | |||
1473 | unsigned int num_elements) | |||
1474 | { | |||
1475 | gcc_checking_assert (num_elements <= N)((void)(!(num_elements <= N) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/wide-int.h" , 1475, __FUNCTION__), 0 : 0)); | |||
1476 | m_num_elements = num_elements; | |||
1477 | m_precision = precision; | |||
1478 | m_max_len = ((precision + HOST_BITS_PER_WIDE_INT64 - 1) | |||
1479 | / HOST_BITS_PER_WIDE_INT64); | |||
1480 | } | |||
1481 | ||||
1482 | /* Return a reference to element INDEX. */ | |||
1483 | template <int N> | |||
1484 | inline trailing_wide_int | |||
1485 | trailing_wide_ints <N>::operator [] (unsigned int index) | |||
1486 | { | |||
1487 | return trailing_wide_int_storage (m_precision, &m_len[index].len, | |||
1488 | &m_val[index * m_max_len]); | |||
1489 | } | |||
1490 | ||||
1491 | template <int N> | |||
1492 | inline typename trailing_wide_ints <N>::const_reference | |||
1493 | trailing_wide_ints <N>::operator [] (unsigned int index) const | |||
1494 | { | |||
1495 | return wi::storage_ref (&m_val[index * m_max_len], | |||
1496 | m_len[index].len, m_precision); | |||
1497 | } | |||
1498 | ||||
1499 | /* Return how many extra bytes need to be added to the end of the | |||
1500 | structure in order to handle NUM_ELEMENTS wide_ints of precision | |||
1501 | PRECISION. NUM_ELEMENTS is the number of elements, and defaults | |||
1502 | to N. */ | |||
1503 | template <int N> | |||
1504 | inline size_t | |||
1505 | trailing_wide_ints <N>::extra_size (unsigned int precision, | |||
1506 | unsigned int num_elements) | |||
1507 | { | |||
1508 | unsigned int max_len = ((precision + HOST_BITS_PER_WIDE_INT64 - 1) | |||
1509 | / HOST_BITS_PER_WIDE_INT64); | |||
1510 | gcc_checking_assert (num_elements <= N)((void)(!(num_elements <= N) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/wide-int.h" , 1510, __FUNCTION__), 0 : 0)); | |||
1511 | return (num_elements * max_len - 1) * sizeof (HOST_WIDE_INTlong); | |||
1512 | } | |||
1513 | ||||
1514 | /* This macro is used in structures that end with a trailing_wide_ints field | |||
1515 | called FIELD. It declares get_NAME() and set_NAME() methods to access | |||
1516 | element I of FIELD. */ | |||
1517 | #define TRAILING_WIDE_INT_ACCESSOR(NAME, FIELD, I)trailing_wide_int get_NAME () { return FIELD[I]; } template < typename T> void set_NAME (const T &x) { FIELD[I] = x; } \ | |||
1518 | trailing_wide_int get_##NAME () { return FIELD[I]; } \ | |||
1519 | template <typename T> void set_##NAME (const T &x) { FIELD[I] = x; } | |||
1520 | ||||
1521 | namespace wi | |||
1522 | { | |||
1523 | /* Implementation of int_traits for primitive integer types like "int". */ | |||
1524 | template <typename T, bool signed_p> | |||
1525 | struct primitive_int_traits | |||
1526 | { | |||
1527 | static const enum precision_type precision_type = FLEXIBLE_PRECISION; | |||
1528 | static const bool host_dependent_precision = true; | |||
1529 | static const bool is_sign_extended = true; | |||
1530 | static unsigned int get_precision (T); | |||
1531 | static wi::storage_ref decompose (HOST_WIDE_INTlong *, unsigned int, T); | |||
1532 | }; | |||
1533 | } | |||
1534 | ||||
1535 | template <typename T, bool signed_p> | |||
1536 | inline unsigned int | |||
1537 | wi::primitive_int_traits <T, signed_p>::get_precision (T) | |||
1538 | { | |||
1539 | return sizeof (T) * CHAR_BIT8; | |||
1540 | } | |||
1541 | ||||
1542 | template <typename T, bool signed_p> | |||
1543 | inline wi::storage_ref | |||
1544 | wi::primitive_int_traits <T, signed_p>::decompose (HOST_WIDE_INTlong *scratch, | |||
1545 | unsigned int precision, T x) | |||
1546 | { | |||
1547 | scratch[0] = x; | |||
1548 | if (signed_p || scratch[0] >= 0 || precision <= HOST_BITS_PER_WIDE_INT64) | |||
1549 | return wi::storage_ref (scratch, 1, precision); | |||
1550 | scratch[1] = 0; | |||
1551 | return wi::storage_ref (scratch, 2, precision); | |||
1552 | } | |||
1553 | ||||
1554 | /* Allow primitive C types to be used in wi:: routines. */ | |||
1555 | namespace wi | |||
1556 | { | |||
1557 | template <> | |||
1558 | struct int_traits <unsigned char> | |||
1559 | : public primitive_int_traits <unsigned char, false> {}; | |||
1560 | ||||
1561 | template <> | |||
1562 | struct int_traits <unsigned short> | |||
1563 | : public primitive_int_traits <unsigned short, false> {}; | |||
1564 | ||||
1565 | template <> | |||
1566 | struct int_traits <int> | |||
1567 | : public primitive_int_traits <int, true> {}; | |||
1568 | ||||
1569 | template <> | |||
1570 | struct int_traits <unsigned int> | |||
1571 | : public primitive_int_traits <unsigned int, false> {}; | |||
1572 | ||||
1573 | template <> | |||
1574 | struct int_traits <long> | |||
1575 | : public primitive_int_traits <long, true> {}; | |||
1576 | ||||
1577 | template <> | |||
1578 | struct int_traits <unsigned long> | |||
1579 | : public primitive_int_traits <unsigned long, false> {}; | |||
1580 | ||||
1581 | #if defined HAVE_LONG_LONG1 | |||
1582 | template <> | |||
1583 | struct int_traits <long long> | |||
1584 | : public primitive_int_traits <long long, true> {}; | |||
1585 | ||||
1586 | template <> | |||
1587 | struct int_traits <unsigned long long> | |||
1588 | : public primitive_int_traits <unsigned long long, false> {}; | |||
1589 | #endif | |||
1590 | } | |||
1591 | ||||
1592 | namespace wi | |||
1593 | { | |||
1594 | /* Stores HWI-sized integer VAL, treating it as having signedness SGN | |||
1595 | and precision PRECISION. */ | |||
1596 | class hwi_with_prec | |||
1597 | { | |||
1598 | public: | |||
1599 | hwi_with_prec () {} | |||
1600 | hwi_with_prec (HOST_WIDE_INTlong, unsigned int, signop); | |||
1601 | HOST_WIDE_INTlong val; | |||
1602 | unsigned int precision; | |||
1603 | signop sgn; | |||
1604 | }; | |||
1605 | ||||
1606 | hwi_with_prec shwi (HOST_WIDE_INTlong, unsigned int); | |||
1607 | hwi_with_prec uhwi (unsigned HOST_WIDE_INTlong, unsigned int); | |||
1608 | ||||
1609 | hwi_with_prec minus_one (unsigned int); | |||
1610 | hwi_with_prec zero (unsigned int); | |||
1611 | hwi_with_prec one (unsigned int); | |||
1612 | hwi_with_prec two (unsigned int); | |||
1613 | } | |||
1614 | ||||
1615 | inline wi::hwi_with_prec::hwi_with_prec (HOST_WIDE_INTlong v, unsigned int p, | |||
1616 | signop s) | |||
1617 | : precision (p), sgn (s) | |||
1618 | { | |||
1619 | if (precision < HOST_BITS_PER_WIDE_INT64) | |||
1620 | val = sext_hwi (v, precision); | |||
1621 | else | |||
1622 | val = v; | |||
1623 | } | |||
1624 | ||||
1625 | /* Return a signed integer that has value VAL and precision PRECISION. */ | |||
1626 | inline wi::hwi_with_prec | |||
1627 | wi::shwi (HOST_WIDE_INTlong val, unsigned int precision) | |||
1628 | { | |||
1629 | return hwi_with_prec (val, precision, SIGNED); | |||
1630 | } | |||
1631 | ||||
1632 | /* Return an unsigned integer that has value VAL and precision PRECISION. */ | |||
1633 | inline wi::hwi_with_prec | |||
1634 | wi::uhwi (unsigned HOST_WIDE_INTlong val, unsigned int precision) | |||
1635 | { | |||
1636 | return hwi_with_prec (val, precision, UNSIGNED); | |||
1637 | } | |||
1638 | ||||
1639 | /* Return a wide int of -1 with precision PRECISION. */ | |||
1640 | inline wi::hwi_with_prec | |||
1641 | wi::minus_one (unsigned int precision) | |||
1642 | { | |||
1643 | return wi::shwi (-1, precision); | |||
1644 | } | |||
1645 | ||||
1646 | /* Return a wide int of 0 with precision PRECISION. */ | |||
1647 | inline wi::hwi_with_prec | |||
1648 | wi::zero (unsigned int precision) | |||
1649 | { | |||
1650 | return wi::shwi (0, precision); | |||
1651 | } | |||
1652 | ||||
1653 | /* Return a wide int of 1 with precision PRECISION. */ | |||
1654 | inline wi::hwi_with_prec | |||
1655 | wi::one (unsigned int precision) | |||
1656 | { | |||
1657 | return wi::shwi (1, precision); | |||
1658 | } | |||
1659 | ||||
1660 | /* Return a wide int of 2 with precision PRECISION. */ | |||
1661 | inline wi::hwi_with_prec | |||
1662 | wi::two (unsigned int precision) | |||
1663 | { | |||
1664 | return wi::shwi (2, precision); | |||
1665 | } | |||
1666 | ||||
1667 | namespace wi | |||
1668 | { | |||
1669 | /* ints_for<T>::zero (X) returns a zero that, when asssigned to a T, | |||
1670 | gives that T the same precision as X. */ | |||
1671 | template<typename T, precision_type = int_traits<T>::precision_type> | |||
1672 | struct ints_for | |||
1673 | { | |||
1674 | static int zero (const T &) { return 0; } | |||
1675 | }; | |||
1676 | ||||
1677 | template<typename T> | |||
1678 | struct ints_for<T, VAR_PRECISION> | |||
1679 | { | |||
1680 | static hwi_with_prec zero (const T &); | |||
1681 | }; | |||
1682 | } | |||
1683 | ||||
1684 | template<typename T> | |||
1685 | inline wi::hwi_with_prec | |||
1686 | wi::ints_for<T, wi::VAR_PRECISION>::zero (const T &x) | |||
1687 | { | |||
1688 | return wi::zero (wi::get_precision (x)); | |||
1689 | } | |||
1690 | ||||
1691 | namespace wi | |||
1692 | { | |||
1693 | template <> | |||
1694 | struct int_traits <wi::hwi_with_prec> | |||
1695 | { | |||
1696 | static const enum precision_type precision_type = VAR_PRECISION; | |||
1697 | /* hwi_with_prec has an explicitly-given precision, rather than the | |||
1698 | precision of HOST_WIDE_INT. */ | |||
1699 | static const bool host_dependent_precision = false; | |||
1700 | static const bool is_sign_extended = true; | |||
1701 | static unsigned int get_precision (const wi::hwi_with_prec &); | |||
1702 | static wi::storage_ref decompose (HOST_WIDE_INTlong *, unsigned int, | |||
1703 | const wi::hwi_with_prec &); | |||
1704 | }; | |||
1705 | } | |||
1706 | ||||
1707 | inline unsigned int | |||
1708 | wi::int_traits <wi::hwi_with_prec>::get_precision (const wi::hwi_with_prec &x) | |||
1709 | { | |||
1710 | return x.precision; | |||
1711 | } | |||
1712 | ||||
1713 | inline wi::storage_ref | |||
1714 | wi::int_traits <wi::hwi_with_prec>:: | |||
1715 | decompose (HOST_WIDE_INTlong *scratch, unsigned int precision, | |||
1716 | const wi::hwi_with_prec &x) | |||
1717 | { | |||
1718 | gcc_checking_assert (precision == x.precision)((void)(!(precision == x.precision) ? fancy_abort ("/buildworker/marxinbox-gcc-clang-static-analyzer/build/gcc/wide-int.h" , 1718, __FUNCTION__), 0 : 0)); | |||
1719 | scratch[0] = x.val; | |||
1720 | if (x.sgn == SIGNED || x.val >= 0 || precision <= HOST_BITS_PER_WIDE_INT64) | |||
1721 | return wi::storage_ref (scratch, 1, precision); | |||
1722 | scratch[1] = 0; | |||
1723 | return wi::storage_ref (scratch, 2, precision); | |||
1724 | } | |||
1725 | ||||
1726 | /* Private functions for handling large cases out of line. They take | |||
1727 | individual length and array parameters because that is cheaper for | |||
1728 | the inline caller than constructing an object on the stack and | |||
1729 | passing a reference to it. (Although many callers use wide_int_refs, | |||
1730 | we generally want those to be removed by SRA.) */ | |||
1731 | namespace wi | |||
1732 | { | |||
1733 | bool eq_p_large (const HOST_WIDE_INTlong *, unsigned int, | |||
1734 | const HOST_WIDE_INTlong *, unsigned int, unsigned int); | |||
1735 | bool lts_p_large (const HOST_WIDE_INTlong *, unsigned int, unsigned int, | |||
1736 | const HOST_WIDE_INTlong *, unsigned int); | |||
1737 | bool ltu_p_large (const HOST_WIDE_INTlong *, unsigned int, unsigned int, | |||
1738 | const HOST_WIDE_INTlong *, unsigned int); | |||
1739 | int cmps_large (const HOST_WIDE_INTlong *, unsigned int, unsigned int, | |||
1740 | const HOST_WIDE_INTlong *, unsigned int); | |||
1741 | int cmpu_large (const HOST_WIDE_INTlong *, unsigned int, unsigned int, | |||
1742 | const HOST_WIDE_INTlong *, unsigned int); | |||
1743 | unsigned int sext_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1744 | unsigned int, | |||
1745 | unsigned int, unsigned int); | |||
1746 | unsigned int zext_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1747 | unsigned int, | |||
1748 | unsigned int, unsigned int); | |||
1749 | unsigned int set_bit_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1750 | unsigned int, unsigned int, unsigned int); | |||
1751 | unsigned int lshift_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1752 | unsigned int, unsigned int, unsigned int); | |||
1753 | unsigned int lrshift_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1754 | unsigned int, unsigned int, unsigned int, | |||
1755 | unsigned int); | |||
1756 | unsigned int arshift_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1757 | unsigned int, unsigned int, unsigned int, | |||
1758 | unsigned int); | |||
1759 | unsigned int and_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, unsigned int, | |||
1760 | const HOST_WIDE_INTlong *, unsigned int, unsigned int); | |||
1761 | unsigned int and_not_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1762 | unsigned int, const HOST_WIDE_INTlong *, | |||
1763 | unsigned int, unsigned int); | |||
1764 | unsigned int or_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, unsigned int, | |||
1765 | const HOST_WIDE_INTlong *, unsigned int, unsigned int); | |||
1766 | unsigned int or_not_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1767 | unsigned int, const HOST_WIDE_INTlong *, | |||
1768 | unsigned int, unsigned int); | |||
1769 | unsigned int xor_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, unsigned int, | |||
1770 | const HOST_WIDE_INTlong *, unsigned int, unsigned int); | |||
1771 | unsigned int add_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, unsigned int, | |||
1772 | const HOST_WIDE_INTlong *, unsigned int, unsigned int, | |||
1773 | signop, overflow_type *); | |||
1774 | unsigned int sub_large (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, unsigned int, | |||
1775 | const HOST_WIDE_INTlong *, unsigned int, unsigned int, | |||
1776 | signop, overflow_type *); | |||
1777 | unsigned int mul_internal (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1778 | unsigned int, const HOST_WIDE_INTlong *, | |||
1779 | unsigned int, unsigned int, signop, | |||
1780 | overflow_type *, bool); | |||
1781 | unsigned int divmod_internal (HOST_WIDE_INTlong *, unsigned int *, | |||
1782 | HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
1783 | unsigned int, unsigned int, | |||
1784 | const HOST_WIDE_INTlong *, | |||
1785 | unsigned int, unsigned int, | |||
1786 | signop, overflow_type *); | |||
1787 | } | |||
1788 | ||||
1789 | /* Return the number of bits that integer X can hold. */ | |||
1790 | template <typename T> | |||
1791 | inline unsigned int | |||
1792 | wi::get_precision (const T &x) | |||
1793 | { | |||
1794 | return wi::int_traits <T>::get_precision (x); | |||
1795 | } | |||
1796 | ||||
1797 | /* Return the number of bits that the result of a binary operation can | |||
1798 | hold when the input operands are X and Y. */ | |||
1799 | template <typename T1, typename T2> | |||
1800 | inline unsigned int | |||
1801 | wi::get_binary_precision (const T1 &x, const T2 &y) | |||
1802 | { | |||
1803 | return get_precision (wi::int_traits <WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type>:: | |||
1804 | get_binary_result (x, y)); | |||
1805 | } | |||
1806 | ||||
1807 | /* Copy the contents of Y to X, but keeping X's current precision. */ | |||
1808 | template <typename T1, typename T2> | |||
1809 | inline void | |||
1810 | wi::copy (T1 &x, const T2 &y) | |||
1811 | { | |||
1812 | HOST_WIDE_INTlong *xval = x.write_val (); | |||
1813 | const HOST_WIDE_INTlong *yval = y.get_val (); | |||
1814 | unsigned int len = y.get_len (); | |||
1815 | unsigned int i = 0; | |||
1816 | do | |||
1817 | xval[i] = yval[i]; | |||
1818 | while (++i < len); | |||
1819 | x.set_len (len, y.is_sign_extended); | |||
1820 | } | |||
1821 | ||||
1822 | /* Return true if X fits in a HOST_WIDE_INT with no loss of precision. */ | |||
1823 | template <typename T> | |||
1824 | inline bool | |||
1825 | wi::fits_shwi_p (const T &x) | |||
1826 | { | |||
1827 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x); | |||
1828 | return xi.len == 1; | |||
1829 | } | |||
1830 | ||||
1831 | /* Return true if X fits in an unsigned HOST_WIDE_INT with no loss of | |||
1832 | precision. */ | |||
1833 | template <typename T> | |||
1834 | inline bool | |||
1835 | wi::fits_uhwi_p (const T &x) | |||
1836 | { | |||
1837 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x); | |||
1838 | if (xi.precision <= HOST_BITS_PER_WIDE_INT64) | |||
1839 | return true; | |||
1840 | if (xi.len == 1) | |||
1841 | return xi.slow () >= 0; | |||
1842 | return xi.len == 2 && xi.uhigh () == 0; | |||
1843 | } | |||
1844 | ||||
1845 | /* Return true if X is negative based on the interpretation of SGN. | |||
1846 | For UNSIGNED, this is always false. */ | |||
1847 | template <typename T> | |||
1848 | inline bool | |||
1849 | wi::neg_p (const T &x, signop sgn) | |||
1850 | { | |||
1851 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x); | |||
1852 | if (sgn == UNSIGNED) | |||
1853 | return false; | |||
1854 | return xi.sign_mask () < 0; | |||
1855 | } | |||
1856 | ||||
1857 | /* Return -1 if the top bit of X is set and 0 if the top bit is clear. */ | |||
1858 | template <typename T> | |||
1859 | inline HOST_WIDE_INTlong | |||
1860 | wi::sign_mask (const T &x) | |||
1861 | { | |||
1862 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x); | |||
1863 | return xi.sign_mask (); | |||
1864 | } | |||
1865 | ||||
1866 | /* Return true if X == Y. X and Y must be binary-compatible. */ | |||
1867 | template <typename T1, typename T2> | |||
1868 | inline bool | |||
1869 | wi::eq_p (const T1 &x, const T2 &y) | |||
1870 | { | |||
1871 | unsigned int precision = get_binary_precision (x, y); | |||
1872 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
1873 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
1874 | if (xi.is_sign_extended && yi.is_sign_extended) | |||
1875 | { | |||
1876 | /* This case reduces to array equality. */ | |||
1877 | if (xi.len != yi.len) | |||
1878 | return false; | |||
1879 | unsigned int i = 0; | |||
1880 | do | |||
1881 | if (xi.val[i] != yi.val[i]) | |||
1882 | return false; | |||
1883 | while (++i != xi.len); | |||
1884 | return true; | |||
1885 | } | |||
1886 | if (LIKELY (yi.len == 1)(__builtin_expect ((yi.len == 1), 1))) | |||
1887 | { | |||
1888 | /* XI is only equal to YI if it too has a single HWI. */ | |||
1889 | if (xi.len != 1) | |||
1890 | return false; | |||
1891 | /* Excess bits in xi.val[0] will be signs or zeros, so comparisons | |||
1892 | with 0 are simple. */ | |||
1893 | if (STATIC_CONSTANT_P (yi.val[0] == 0)(__builtin_constant_p (yi.val[0] == 0) && (yi.val[0] == 0))) | |||
1894 | return xi.val[0] == 0; | |||
1895 | /* Otherwise flush out any excess bits first. */ | |||
1896 | unsigned HOST_WIDE_INTlong diff = xi.val[0] ^ yi.val[0]; | |||
1897 | int excess = HOST_BITS_PER_WIDE_INT64 - precision; | |||
1898 | if (excess > 0) | |||
1899 | diff <<= excess; | |||
1900 | return diff == 0; | |||
1901 | } | |||
1902 | return eq_p_large (xi.val, xi.len, yi.val, yi.len, precision); | |||
1903 | } | |||
1904 | ||||
1905 | /* Return true if X != Y. X and Y must be binary-compatible. */ | |||
1906 | template <typename T1, typename T2> | |||
1907 | inline bool | |||
1908 | wi::ne_p (const T1 &x, const T2 &y) | |||
1909 | { | |||
1910 | return !eq_p (x, y); | |||
1911 | } | |||
1912 | ||||
1913 | /* Return true if X < Y when both are treated as signed values. */ | |||
1914 | template <typename T1, typename T2> | |||
1915 | inline bool | |||
1916 | wi::lts_p (const T1 &x, const T2 &y) | |||
1917 | { | |||
1918 | unsigned int precision = get_binary_precision (x, y); | |||
1919 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
1920 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
1921 | /* We optimize x < y, where y is 64 or fewer bits. */ | |||
1922 | if (wi::fits_shwi_p (yi)) | |||
1923 | { | |||
1924 | /* Make lts_p (x, 0) as efficient as wi::neg_p (x). */ | |||
1925 | if (STATIC_CONSTANT_P (yi.val[0] == 0)(__builtin_constant_p (yi.val[0] == 0) && (yi.val[0] == 0))) | |||
1926 | return neg_p (xi); | |||
1927 | /* If x fits directly into a shwi, we can compare directly. */ | |||
1928 | if (wi::fits_shwi_p (xi)) | |||
1929 | return xi.to_shwi () < yi.to_shwi (); | |||
1930 | /* If x doesn't fit and is negative, then it must be more | |||
1931 | negative than any value in y, and hence smaller than y. */ | |||
1932 | if (neg_p (xi)) | |||
1933 | return true; | |||
1934 | /* If x is positive, then it must be larger than any value in y, | |||
1935 | and hence greater than y. */ | |||
1936 | return false; | |||
1937 | } | |||
1938 | /* Optimize the opposite case, if it can be detected at compile time. */ | |||
1939 | if (STATIC_CONSTANT_P (xi.len == 1)(__builtin_constant_p (xi.len == 1) && (xi.len == 1))) | |||
1940 | /* If YI is negative it is lower than the least HWI. | |||
1941 | If YI is positive it is greater than the greatest HWI. */ | |||
1942 | return !neg_p (yi); | |||
1943 | return lts_p_large (xi.val, xi.len, precision, yi.val, yi.len); | |||
1944 | } | |||
1945 | ||||
1946 | /* Return true if X < Y when both are treated as unsigned values. */ | |||
1947 | template <typename T1, typename T2> | |||
1948 | inline bool | |||
1949 | wi::ltu_p (const T1 &x, const T2 &y) | |||
1950 | { | |||
1951 | unsigned int precision = get_binary_precision (x, y); | |||
1952 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
1953 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
1954 | /* Optimize comparisons with constants. */ | |||
1955 | if (STATIC_CONSTANT_P (yi.len == 1 && yi.val[0] >= 0)(__builtin_constant_p (yi.len == 1 && yi.val[0] >= 0) && (yi.len == 1 && yi.val[0] >= 0))) | |||
1956 | return xi.len == 1 && xi.to_uhwi () < (unsigned HOST_WIDE_INTlong) yi.val[0]; | |||
1957 | if (STATIC_CONSTANT_P (xi.len == 1 && xi.val[0] >= 0)(__builtin_constant_p (xi.len == 1 && xi.val[0] >= 0) && (xi.len == 1 && xi.val[0] >= 0))) | |||
1958 | return yi.len != 1 || yi.to_uhwi () > (unsigned HOST_WIDE_INTlong) xi.val[0]; | |||
1959 | /* Optimize the case of two HWIs. The HWIs are implicitly sign-extended | |||
1960 | for precisions greater than HOST_BITS_WIDE_INT, but sign-extending both | |||
1961 | values does not change the result. */ | |||
1962 | if (LIKELY (xi.len + yi.len == 2)(__builtin_expect ((xi.len + yi.len == 2), 1))) | |||
1963 | { | |||
1964 | unsigned HOST_WIDE_INTlong xl = xi.to_uhwi (); | |||
1965 | unsigned HOST_WIDE_INTlong yl = yi.to_uhwi (); | |||
1966 | return xl < yl; | |||
1967 | } | |||
1968 | return ltu_p_large (xi.val, xi.len, precision, yi.val, yi.len); | |||
1969 | } | |||
1970 | ||||
1971 | /* Return true if X < Y. Signedness of X and Y is indicated by SGN. */ | |||
1972 | template <typename T1, typename T2> | |||
1973 | inline bool | |||
1974 | wi::lt_p (const T1 &x, const T2 &y, signop sgn) | |||
1975 | { | |||
1976 | if (sgn == SIGNED) | |||
1977 | return lts_p (x, y); | |||
1978 | else | |||
1979 | return ltu_p (x, y); | |||
1980 | } | |||
1981 | ||||
1982 | /* Return true if X <= Y when both are treated as signed values. */ | |||
1983 | template <typename T1, typename T2> | |||
1984 | inline bool | |||
1985 | wi::les_p (const T1 &x, const T2 &y) | |||
1986 | { | |||
1987 | return !lts_p (y, x); | |||
1988 | } | |||
1989 | ||||
1990 | /* Return true if X <= Y when both are treated as unsigned values. */ | |||
1991 | template <typename T1, typename T2> | |||
1992 | inline bool | |||
1993 | wi::leu_p (const T1 &x, const T2 &y) | |||
1994 | { | |||
1995 | return !ltu_p (y, x); | |||
1996 | } | |||
1997 | ||||
1998 | /* Return true if X <= Y. Signedness of X and Y is indicated by SGN. */ | |||
1999 | template <typename T1, typename T2> | |||
2000 | inline bool | |||
2001 | wi::le_p (const T1 &x, const T2 &y, signop sgn) | |||
2002 | { | |||
2003 | if (sgn == SIGNED) | |||
2004 | return les_p (x, y); | |||
2005 | else | |||
2006 | return leu_p (x, y); | |||
2007 | } | |||
2008 | ||||
2009 | /* Return true if X > Y when both are treated as signed values. */ | |||
2010 | template <typename T1, typename T2> | |||
2011 | inline bool | |||
2012 | wi::gts_p (const T1 &x, const T2 &y) | |||
2013 | { | |||
2014 | return lts_p (y, x); | |||
2015 | } | |||
2016 | ||||
2017 | /* Return true if X > Y when both are treated as unsigned values. */ | |||
2018 | template <typename T1, typename T2> | |||
2019 | inline bool | |||
2020 | wi::gtu_p (const T1 &x, const T2 &y) | |||
2021 | { | |||
2022 | return ltu_p (y, x); | |||
2023 | } | |||
2024 | ||||
2025 | /* Return true if X > Y. Signedness of X and Y is indicated by SGN. */ | |||
2026 | template <typename T1, typename T2> | |||
2027 | inline bool | |||
2028 | wi::gt_p (const T1 &x, const T2 &y, signop sgn) | |||
2029 | { | |||
2030 | if (sgn == SIGNED) | |||
2031 | return gts_p (x, y); | |||
2032 | else | |||
2033 | return gtu_p (x, y); | |||
2034 | } | |||
2035 | ||||
2036 | /* Return true if X >= Y when both are treated as signed values. */ | |||
2037 | template <typename T1, typename T2> | |||
2038 | inline bool | |||
2039 | wi::ges_p (const T1 &x, const T2 &y) | |||
2040 | { | |||
2041 | return !lts_p (x, y); | |||
2042 | } | |||
2043 | ||||
2044 | /* Return true if X >= Y when both are treated as unsigned values. */ | |||
2045 | template <typename T1, typename T2> | |||
2046 | inline bool | |||
2047 | wi::geu_p (const T1 &x, const T2 &y) | |||
2048 | { | |||
2049 | return !ltu_p (x, y); | |||
2050 | } | |||
2051 | ||||
2052 | /* Return true if X >= Y. Signedness of X and Y is indicated by SGN. */ | |||
2053 | template <typename T1, typename T2> | |||
2054 | inline bool | |||
2055 | wi::ge_p (const T1 &x, const T2 &y, signop sgn) | |||
2056 | { | |||
2057 | if (sgn == SIGNED) | |||
2058 | return ges_p (x, y); | |||
2059 | else | |||
2060 | return geu_p (x, y); | |||
2061 | } | |||
2062 | ||||
2063 | /* Return -1 if X < Y, 0 if X == Y and 1 if X > Y. Treat both X and Y | |||
2064 | as signed values. */ | |||
2065 | template <typename T1, typename T2> | |||
2066 | inline int | |||
2067 | wi::cmps (const T1 &x, const T2 &y) | |||
2068 | { | |||
2069 | unsigned int precision = get_binary_precision (x, y); | |||
2070 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2071 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2072 | if (wi::fits_shwi_p (yi)) | |||
2073 | { | |||
2074 | /* Special case for comparisons with 0. */ | |||
2075 | if (STATIC_CONSTANT_P (yi.val[0] == 0)(__builtin_constant_p (yi.val[0] == 0) && (yi.val[0] == 0))) | |||
2076 | return neg_p (xi) ? -1 : !(xi.len == 1 && xi.val[0] == 0); | |||
2077 | /* If x fits into a signed HWI, we can compare directly. */ | |||
2078 | if (wi::fits_shwi_p (xi)) | |||
2079 | { | |||
2080 | HOST_WIDE_INTlong xl = xi.to_shwi (); | |||
2081 | HOST_WIDE_INTlong yl = yi.to_shwi (); | |||
2082 | return xl < yl ? -1 : xl > yl; | |||
2083 | } | |||
2084 | /* If x doesn't fit and is negative, then it must be more | |||
2085 | negative than any signed HWI, and hence smaller than y. */ | |||
2086 | if (neg_p (xi)) | |||
2087 | return -1; | |||
2088 | /* If x is positive, then it must be larger than any signed HWI, | |||
2089 | and hence greater than y. */ | |||
2090 | return 1; | |||
2091 | } | |||
2092 | /* Optimize the opposite case, if it can be detected at compile time. */ | |||
2093 | if (STATIC_CONSTANT_P (xi.len == 1)(__builtin_constant_p (xi.len == 1) && (xi.len == 1))) | |||
2094 | /* If YI is negative it is lower than the least HWI. | |||
2095 | If YI is positive it is greater than the greatest HWI. */ | |||
2096 | return neg_p (yi) ? 1 : -1; | |||
2097 | return cmps_large (xi.val, xi.len, precision, yi.val, yi.len); | |||
2098 | } | |||
2099 | ||||
2100 | /* Return -1 if X < Y, 0 if X == Y and 1 if X > Y. Treat both X and Y | |||
2101 | as unsigned values. */ | |||
2102 | template <typename T1, typename T2> | |||
2103 | inline int | |||
2104 | wi::cmpu (const T1 &x, const T2 &y) | |||
2105 | { | |||
2106 | unsigned int precision = get_binary_precision (x, y); | |||
2107 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2108 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2109 | /* Optimize comparisons with constants. */ | |||
2110 | if (STATIC_CONSTANT_P (yi.len == 1 && yi.val[0] >= 0)(__builtin_constant_p (yi.len == 1 && yi.val[0] >= 0) && (yi.len == 1 && yi.val[0] >= 0))) | |||
2111 | { | |||
2112 | /* If XI doesn't fit in a HWI then it must be larger than YI. */ | |||
2113 | if (xi.len != 1) | |||
2114 | return 1; | |||
2115 | /* Otherwise compare directly. */ | |||
2116 | unsigned HOST_WIDE_INTlong xl = xi.to_uhwi (); | |||
2117 | unsigned HOST_WIDE_INTlong yl = yi.val[0]; | |||
2118 | return xl < yl ? -1 : xl > yl; | |||
2119 | } | |||
2120 | if (STATIC_CONSTANT_P (xi.len == 1 && xi.val[0] >= 0)(__builtin_constant_p (xi.len == 1 && xi.val[0] >= 0) && (xi.len == 1 && xi.val[0] >= 0))) | |||
2121 | { | |||
2122 | /* If YI doesn't fit in a HWI then it must be larger than XI. */ | |||
2123 | if (yi.len != 1) | |||
2124 | return -1; | |||
2125 | /* Otherwise compare directly. */ | |||
2126 | unsigned HOST_WIDE_INTlong xl = xi.val[0]; | |||
2127 | unsigned HOST_WIDE_INTlong yl = yi.to_uhwi (); | |||
2128 | return xl < yl ? -1 : xl > yl; | |||
2129 | } | |||
2130 | /* Optimize the case of two HWIs. The HWIs are implicitly sign-extended | |||
2131 | for precisions greater than HOST_BITS_WIDE_INT, but sign-extending both | |||
2132 | values does not change the result. */ | |||
2133 | if (LIKELY (xi.len + yi.len == 2)(__builtin_expect ((xi.len + yi.len == 2), 1))) | |||
2134 | { | |||
2135 | unsigned HOST_WIDE_INTlong xl = xi.to_uhwi (); | |||
2136 | unsigned HOST_WIDE_INTlong yl = yi.to_uhwi (); | |||
2137 | return xl < yl ? -1 : xl > yl; | |||
2138 | } | |||
2139 | return cmpu_large (xi.val, xi.len, precision, yi.val, yi.len); | |||
2140 | } | |||
2141 | ||||
2142 | /* Return -1 if X < Y, 0 if X == Y and 1 if X > Y. Signedness of | |||
2143 | X and Y indicated by SGN. */ | |||
2144 | template <typename T1, typename T2> | |||
2145 | inline int | |||
2146 | wi::cmp (const T1 &x, const T2 &y, signop sgn) | |||
2147 | { | |||
2148 | if (sgn == SIGNED) | |||
2149 | return cmps (x, y); | |||
2150 | else | |||
2151 | return cmpu (x, y); | |||
2152 | } | |||
2153 | ||||
2154 | /* Return ~x. */ | |||
2155 | template <typename T> | |||
2156 | inline WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type | |||
2157 | wi::bit_not (const T &x) | |||
2158 | { | |||
2159 | WI_UNARY_RESULT_VAR (result, val, T, x)typename wi::binary_traits <T, T>::result_type result = wi::int_traits <typename wi::binary_traits <T, T>:: result_type>::get_binary_result (x, x); long *val = result .write_val (); | |||
2160 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x, get_precision (result)); | |||
2161 | for (unsigned int i = 0; i < xi.len; ++i) | |||
2162 | val[i] = ~xi.val[i]; | |||
2163 | result.set_len (xi.len); | |||
2164 | return result; | |||
2165 | } | |||
2166 | ||||
2167 | /* Return -x. */ | |||
2168 | template <typename T> | |||
2169 | inline WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type | |||
2170 | wi::neg (const T &x) | |||
2171 | { | |||
2172 | return sub (0, x); | |||
2173 | } | |||
2174 | ||||
2175 | /* Return -x. Indicate in *OVERFLOW if performing the negation would | |||
2176 | cause an overflow. */ | |||
2177 | template <typename T> | |||
2178 | inline WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type | |||
2179 | wi::neg (const T &x, overflow_type *overflow) | |||
2180 | { | |||
2181 | *overflow = only_sign_bit_p (x) ? OVF_OVERFLOW : OVF_NONE; | |||
2182 | return sub (0, x); | |||
2183 | } | |||
2184 | ||||
2185 | /* Return the absolute value of x. */ | |||
2186 | template <typename T> | |||
2187 | inline WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type | |||
2188 | wi::abs (const T &x) | |||
2189 | { | |||
2190 | return neg_p (x) ? neg (x) : WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type (x); | |||
2191 | } | |||
2192 | ||||
2193 | /* Return the result of sign-extending the low OFFSET bits of X. */ | |||
2194 | template <typename T> | |||
2195 | inline WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type | |||
2196 | wi::sext (const T &x, unsigned int offset) | |||
2197 | { | |||
2198 | WI_UNARY_RESULT_VAR (result, val, T, x)typename wi::binary_traits <T, T>::result_type result = wi::int_traits <typename wi::binary_traits <T, T>:: result_type>::get_binary_result (x, x); long *val = result .write_val (); | |||
2199 | unsigned int precision = get_precision (result); | |||
2200 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x, precision); | |||
2201 | ||||
2202 | if (offset <= HOST_BITS_PER_WIDE_INT64) | |||
2203 | { | |||
2204 | val[0] = sext_hwi (xi.ulow (), offset); | |||
2205 | result.set_len (1, true); | |||
2206 | } | |||
2207 | else | |||
2208 | result.set_len (sext_large (val, xi.val, xi.len, precision, offset)); | |||
2209 | return result; | |||
2210 | } | |||
2211 | ||||
2212 | /* Return the result of zero-extending the low OFFSET bits of X. */ | |||
2213 | template <typename T> | |||
2214 | inline WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type | |||
2215 | wi::zext (const T &x, unsigned int offset) | |||
2216 | { | |||
2217 | WI_UNARY_RESULT_VAR (result, val, T, x)typename wi::binary_traits <T, T>::result_type result = wi::int_traits <typename wi::binary_traits <T, T>:: result_type>::get_binary_result (x, x); long *val = result .write_val (); | |||
2218 | unsigned int precision = get_precision (result); | |||
2219 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x, precision); | |||
2220 | ||||
2221 | /* This is not just an optimization, it is actually required to | |||
2222 | maintain canonization. */ | |||
2223 | if (offset >= precision) | |||
2224 | { | |||
2225 | wi::copy (result, xi); | |||
2226 | return result; | |||
2227 | } | |||
2228 | ||||
2229 | /* In these cases we know that at least the top bit will be clear, | |||
2230 | so no sign extension is necessary. */ | |||
2231 | if (offset < HOST_BITS_PER_WIDE_INT64) | |||
2232 | { | |||
2233 | val[0] = zext_hwi (xi.ulow (), offset); | |||
2234 | result.set_len (1, true); | |||
2235 | } | |||
2236 | else | |||
2237 | result.set_len (zext_large (val, xi.val, xi.len, precision, offset), true); | |||
2238 | return result; | |||
2239 | } | |||
2240 | ||||
2241 | /* Return the result of extending the low OFFSET bits of X according to | |||
2242 | signedness SGN. */ | |||
2243 | template <typename T> | |||
2244 | inline WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type | |||
2245 | wi::ext (const T &x, unsigned int offset, signop sgn) | |||
2246 | { | |||
2247 | return sgn == SIGNED ? sext (x, offset) : zext (x, offset); | |||
2248 | } | |||
2249 | ||||
2250 | /* Return an integer that represents X | (1 << bit). */ | |||
2251 | template <typename T> | |||
2252 | inline WI_UNARY_RESULT (T)typename wi::binary_traits <T, T>::result_type | |||
2253 | wi::set_bit (const T &x, unsigned int bit) | |||
2254 | { | |||
2255 | WI_UNARY_RESULT_VAR (result, val, T, x)typename wi::binary_traits <T, T>::result_type result = wi::int_traits <typename wi::binary_traits <T, T>:: result_type>::get_binary_result (x, x); long *val = result .write_val (); | |||
2256 | unsigned int precision = get_precision (result); | |||
2257 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x, precision); | |||
2258 | if (precision <= HOST_BITS_PER_WIDE_INT64) | |||
2259 | { | |||
2260 | val[0] = xi.ulow () | (HOST_WIDE_INT_1U1UL << bit); | |||
2261 | result.set_len (1); | |||
2262 | } | |||
2263 | else | |||
2264 | result.set_len (set_bit_large (val, xi.val, xi.len, precision, bit)); | |||
2265 | return result; | |||
2266 | } | |||
2267 | ||||
2268 | /* Return the mininum of X and Y, treating them both as having | |||
2269 | signedness SGN. */ | |||
2270 | template <typename T1, typename T2> | |||
2271 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2272 | wi::min (const T1 &x, const T2 &y, signop sgn) | |||
2273 | { | |||
2274 | WI_BINARY_RESULT_VAR (result, val ATTRIBUTE_UNUSED, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val __attribute__ ((__unused__)) = result.write_val (); | |||
2275 | unsigned int precision = get_precision (result); | |||
2276 | if (wi::le_p (x, y, sgn)) | |||
2277 | wi::copy (result, WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > (x, precision)); | |||
2278 | else | |||
2279 | wi::copy (result, WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > (y, precision)); | |||
2280 | return result; | |||
2281 | } | |||
2282 | ||||
2283 | /* Return the minimum of X and Y, treating both as signed values. */ | |||
2284 | template <typename T1, typename T2> | |||
2285 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2286 | wi::smin (const T1 &x, const T2 &y) | |||
2287 | { | |||
2288 | return wi::min (x, y, SIGNED); | |||
2289 | } | |||
2290 | ||||
2291 | /* Return the minimum of X and Y, treating both as unsigned values. */ | |||
2292 | template <typename T1, typename T2> | |||
2293 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2294 | wi::umin (const T1 &x, const T2 &y) | |||
2295 | { | |||
2296 | return wi::min (x, y, UNSIGNED); | |||
2297 | } | |||
2298 | ||||
2299 | /* Return the maxinum of X and Y, treating them both as having | |||
2300 | signedness SGN. */ | |||
2301 | template <typename T1, typename T2> | |||
2302 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2303 | wi::max (const T1 &x, const T2 &y, signop sgn) | |||
2304 | { | |||
2305 | WI_BINARY_RESULT_VAR (result, val ATTRIBUTE_UNUSED, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val __attribute__ ((__unused__)) = result.write_val (); | |||
2306 | unsigned int precision = get_precision (result); | |||
2307 | if (wi::ge_p (x, y, sgn)) | |||
2308 | wi::copy (result, WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > (x, precision)); | |||
2309 | else | |||
2310 | wi::copy (result, WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > (y, precision)); | |||
2311 | return result; | |||
2312 | } | |||
2313 | ||||
2314 | /* Return the maximum of X and Y, treating both as signed values. */ | |||
2315 | template <typename T1, typename T2> | |||
2316 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2317 | wi::smax (const T1 &x, const T2 &y) | |||
2318 | { | |||
2319 | return wi::max (x, y, SIGNED); | |||
2320 | } | |||
2321 | ||||
2322 | /* Return the maximum of X and Y, treating both as unsigned values. */ | |||
2323 | template <typename T1, typename T2> | |||
2324 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2325 | wi::umax (const T1 &x, const T2 &y) | |||
2326 | { | |||
2327 | return wi::max (x, y, UNSIGNED); | |||
2328 | } | |||
2329 | ||||
2330 | /* Return X & Y. */ | |||
2331 | template <typename T1, typename T2> | |||
2332 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2333 | wi::bit_and (const T1 &x, const T2 &y) | |||
2334 | { | |||
2335 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2336 | unsigned int precision = get_precision (result); | |||
2337 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2338 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2339 | bool is_sign_extended = xi.is_sign_extended && yi.is_sign_extended; | |||
2340 | if (LIKELY (xi.len + yi.len == 2)(__builtin_expect ((xi.len + yi.len == 2), 1))) | |||
2341 | { | |||
2342 | val[0] = xi.ulow () & yi.ulow (); | |||
2343 | result.set_len (1, is_sign_extended); | |||
2344 | } | |||
2345 | else | |||
2346 | result.set_len (and_large (val, xi.val, xi.len, yi.val, yi.len, | |||
2347 | precision), is_sign_extended); | |||
2348 | return result; | |||
2349 | } | |||
2350 | ||||
2351 | /* Return X & ~Y. */ | |||
2352 | template <typename T1, typename T2> | |||
2353 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2354 | wi::bit_and_not (const T1 &x, const T2 &y) | |||
2355 | { | |||
2356 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2357 | unsigned int precision = get_precision (result); | |||
2358 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2359 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2360 | bool is_sign_extended = xi.is_sign_extended && yi.is_sign_extended; | |||
2361 | if (LIKELY (xi.len + yi.len == 2)(__builtin_expect ((xi.len + yi.len == 2), 1))) | |||
2362 | { | |||
2363 | val[0] = xi.ulow () & ~yi.ulow (); | |||
2364 | result.set_len (1, is_sign_extended); | |||
2365 | } | |||
2366 | else | |||
2367 | result.set_len (and_not_large (val, xi.val, xi.len, yi.val, yi.len, | |||
2368 | precision), is_sign_extended); | |||
2369 | return result; | |||
2370 | } | |||
2371 | ||||
2372 | /* Return X | Y. */ | |||
2373 | template <typename T1, typename T2> | |||
2374 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2375 | wi::bit_or (const T1 &x, const T2 &y) | |||
2376 | { | |||
2377 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2378 | unsigned int precision = get_precision (result); | |||
2379 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2380 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2381 | bool is_sign_extended = xi.is_sign_extended && yi.is_sign_extended; | |||
2382 | if (LIKELY (xi.len + yi.len == 2)(__builtin_expect ((xi.len + yi.len == 2), 1))) | |||
2383 | { | |||
2384 | val[0] = xi.ulow () | yi.ulow (); | |||
2385 | result.set_len (1, is_sign_extended); | |||
2386 | } | |||
2387 | else | |||
2388 | result.set_len (or_large (val, xi.val, xi.len, | |||
2389 | yi.val, yi.len, precision), is_sign_extended); | |||
2390 | return result; | |||
2391 | } | |||
2392 | ||||
2393 | /* Return X | ~Y. */ | |||
2394 | template <typename T1, typename T2> | |||
2395 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2396 | wi::bit_or_not (const T1 &x, const T2 &y) | |||
2397 | { | |||
2398 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2399 | unsigned int precision = get_precision (result); | |||
2400 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2401 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2402 | bool is_sign_extended = xi.is_sign_extended && yi.is_sign_extended; | |||
2403 | if (LIKELY (xi.len + yi.len == 2)(__builtin_expect ((xi.len + yi.len == 2), 1))) | |||
2404 | { | |||
2405 | val[0] = xi.ulow () | ~yi.ulow (); | |||
2406 | result.set_len (1, is_sign_extended); | |||
2407 | } | |||
2408 | else | |||
2409 | result.set_len (or_not_large (val, xi.val, xi.len, yi.val, yi.len, | |||
2410 | precision), is_sign_extended); | |||
2411 | return result; | |||
2412 | } | |||
2413 | ||||
2414 | /* Return X ^ Y. */ | |||
2415 | template <typename T1, typename T2> | |||
2416 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2417 | wi::bit_xor (const T1 &x, const T2 &y) | |||
2418 | { | |||
2419 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2420 | unsigned int precision = get_precision (result); | |||
2421 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2422 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2423 | bool is_sign_extended = xi.is_sign_extended && yi.is_sign_extended; | |||
2424 | if (LIKELY (xi.len + yi.len == 2)(__builtin_expect ((xi.len + yi.len == 2), 1))) | |||
2425 | { | |||
2426 | val[0] = xi.ulow () ^ yi.ulow (); | |||
2427 | result.set_len (1, is_sign_extended); | |||
2428 | } | |||
2429 | else | |||
2430 | result.set_len (xor_large (val, xi.val, xi.len, | |||
2431 | yi.val, yi.len, precision), is_sign_extended); | |||
2432 | return result; | |||
2433 | } | |||
2434 | ||||
2435 | /* Return X + Y. */ | |||
2436 | template <typename T1, typename T2> | |||
2437 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2438 | wi::add (const T1 &x, const T2 &y) | |||
2439 | { | |||
2440 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2441 | unsigned int precision = get_precision (result); | |||
2442 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2443 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2444 | if (precision <= HOST_BITS_PER_WIDE_INT64) | |||
2445 | { | |||
2446 | val[0] = xi.ulow () + yi.ulow (); | |||
2447 | result.set_len (1); | |||
2448 | } | |||
2449 | /* If the precision is known at compile time to be greater than | |||
2450 | HOST_BITS_PER_WIDE_INT, we can optimize the single-HWI case | |||
2451 | knowing that (a) all bits in those HWIs are significant and | |||
2452 | (b) the result has room for at least two HWIs. This provides | |||
2453 | a fast path for things like offset_int and widest_int. | |||
2454 | ||||
2455 | The STATIC_CONSTANT_P test prevents this path from being | |||
2456 | used for wide_ints. wide_ints with precisions greater than | |||
2457 | HOST_BITS_PER_WIDE_INT are relatively rare and there's not much | |||
2458 | point handling them inline. */ | |||
2459 | else if (STATIC_CONSTANT_P (precision > HOST_BITS_PER_WIDE_INT)(__builtin_constant_p (precision > 64) && (precision > 64)) | |||
2460 | && LIKELY (xi.len + yi.len == 2)(__builtin_expect ((xi.len + yi.len == 2), 1))) | |||
2461 | { | |||
2462 | unsigned HOST_WIDE_INTlong xl = xi.ulow (); | |||
2463 | unsigned HOST_WIDE_INTlong yl = yi.ulow (); | |||
2464 | unsigned HOST_WIDE_INTlong resultl = xl + yl; | |||
2465 | val[0] = resultl; | |||
2466 | val[1] = (HOST_WIDE_INTlong) resultl < 0 ? 0 : -1; | |||
2467 | result.set_len (1 + (((resultl ^ xl) & (resultl ^ yl)) | |||
2468 | >> (HOST_BITS_PER_WIDE_INT64 - 1))); | |||
2469 | } | |||
2470 | else | |||
2471 | result.set_len (add_large (val, xi.val, xi.len, | |||
2472 | yi.val, yi.len, precision, | |||
2473 | UNSIGNED, 0)); | |||
2474 | return result; | |||
2475 | } | |||
2476 | ||||
2477 | /* Return X + Y. Treat X and Y as having the signednes given by SGN | |||
2478 | and indicate in *OVERFLOW whether the operation overflowed. */ | |||
2479 | template <typename T1, typename T2> | |||
2480 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2481 | wi::add (const T1 &x, const T2 &y, signop sgn, overflow_type *overflow) | |||
2482 | { | |||
2483 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2484 | unsigned int precision = get_precision (result); | |||
2485 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2486 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2487 | if (precision <= HOST_BITS_PER_WIDE_INT64) | |||
2488 | { | |||
2489 | unsigned HOST_WIDE_INTlong xl = xi.ulow (); | |||
2490 | unsigned HOST_WIDE_INTlong yl = yi.ulow (); | |||
2491 | unsigned HOST_WIDE_INTlong resultl = xl + yl; | |||
2492 | if (sgn == SIGNED) | |||
2493 | { | |||
2494 | if ((((resultl ^ xl) & (resultl ^ yl)) | |||
2495 | >> (precision - 1)) & 1) | |||
2496 | { | |||
2497 | if (xl > resultl) | |||
2498 | *overflow = OVF_UNDERFLOW; | |||
2499 | else if (xl < resultl) | |||
2500 | *overflow = OVF_OVERFLOW; | |||
2501 | else | |||
2502 | *overflow = OVF_NONE; | |||
2503 | } | |||
2504 | else | |||
2505 | *overflow = OVF_NONE; | |||
2506 | } | |||
2507 | else | |||
2508 | *overflow = ((resultl << (HOST_BITS_PER_WIDE_INT64 - precision)) | |||
2509 | < (xl << (HOST_BITS_PER_WIDE_INT64 - precision))) | |||
2510 | ? OVF_OVERFLOW : OVF_NONE; | |||
2511 | val[0] = resultl; | |||
2512 | result.set_len (1); | |||
2513 | } | |||
2514 | else | |||
2515 | result.set_len (add_large (val, xi.val, xi.len, | |||
2516 | yi.val, yi.len, precision, | |||
2517 | sgn, overflow)); | |||
2518 | return result; | |||
2519 | } | |||
2520 | ||||
2521 | /* Return X - Y. */ | |||
2522 | template <typename T1, typename T2> | |||
2523 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2524 | wi::sub (const T1 &x, const T2 &y) | |||
2525 | { | |||
2526 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2527 | unsigned int precision = get_precision (result); | |||
2528 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2529 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2530 | if (precision <= HOST_BITS_PER_WIDE_INT64) | |||
2531 | { | |||
2532 | val[0] = xi.ulow () - yi.ulow (); | |||
2533 | result.set_len (1); | |||
2534 | } | |||
2535 | /* If the precision is known at compile time to be greater than | |||
2536 | HOST_BITS_PER_WIDE_INT, we can optimize the single-HWI case | |||
2537 | knowing that (a) all bits in those HWIs are significant and | |||
2538 | (b) the result has room for at least two HWIs. This provides | |||
2539 | a fast path for things like offset_int and widest_int. | |||
2540 | ||||
2541 | The STATIC_CONSTANT_P test prevents this path from being | |||
2542 | used for wide_ints. wide_ints with precisions greater than | |||
2543 | HOST_BITS_PER_WIDE_INT are relatively rare and there's not much | |||
2544 | point handling them inline. */ | |||
2545 | else if (STATIC_CONSTANT_P (precision > HOST_BITS_PER_WIDE_INT)(__builtin_constant_p (precision > 64) && (precision > 64)) | |||
2546 | && LIKELY (xi.len + yi.len == 2)(__builtin_expect ((xi.len + yi.len == 2), 1))) | |||
2547 | { | |||
2548 | unsigned HOST_WIDE_INTlong xl = xi.ulow (); | |||
2549 | unsigned HOST_WIDE_INTlong yl = yi.ulow (); | |||
2550 | unsigned HOST_WIDE_INTlong resultl = xl - yl; | |||
2551 | val[0] = resultl; | |||
2552 | val[1] = (HOST_WIDE_INTlong) resultl < 0 ? 0 : -1; | |||
2553 | result.set_len (1 + (((resultl ^ xl) & (xl ^ yl)) | |||
2554 | >> (HOST_BITS_PER_WIDE_INT64 - 1))); | |||
2555 | } | |||
2556 | else | |||
2557 | result.set_len (sub_large (val, xi.val, xi.len, | |||
2558 | yi.val, yi.len, precision, | |||
2559 | UNSIGNED, 0)); | |||
2560 | return result; | |||
2561 | } | |||
2562 | ||||
2563 | /* Return X - Y. Treat X and Y as having the signednes given by SGN | |||
2564 | and indicate in *OVERFLOW whether the operation overflowed. */ | |||
2565 | template <typename T1, typename T2> | |||
2566 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2567 | wi::sub (const T1 &x, const T2 &y, signop sgn, overflow_type *overflow) | |||
2568 | { | |||
2569 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2570 | unsigned int precision = get_precision (result); | |||
2571 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2572 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2573 | if (precision <= HOST_BITS_PER_WIDE_INT64) | |||
2574 | { | |||
2575 | unsigned HOST_WIDE_INTlong xl = xi.ulow (); | |||
2576 | unsigned HOST_WIDE_INTlong yl = yi.ulow (); | |||
2577 | unsigned HOST_WIDE_INTlong resultl = xl - yl; | |||
2578 | if (sgn == SIGNED) | |||
2579 | { | |||
2580 | if ((((xl ^ yl) & (resultl ^ xl)) >> (precision - 1)) & 1) | |||
2581 | { | |||
2582 | if (xl > yl) | |||
2583 | *overflow = OVF_UNDERFLOW; | |||
2584 | else if (xl < yl) | |||
2585 | *overflow = OVF_OVERFLOW; | |||
2586 | else | |||
2587 | *overflow = OVF_NONE; | |||
2588 | } | |||
2589 | else | |||
2590 | *overflow = OVF_NONE; | |||
2591 | } | |||
2592 | else | |||
2593 | *overflow = ((resultl << (HOST_BITS_PER_WIDE_INT64 - precision)) | |||
2594 | > (xl << (HOST_BITS_PER_WIDE_INT64 - precision))) | |||
2595 | ? OVF_UNDERFLOW : OVF_NONE; | |||
2596 | val[0] = resultl; | |||
2597 | result.set_len (1); | |||
2598 | } | |||
2599 | else | |||
2600 | result.set_len (sub_large (val, xi.val, xi.len, | |||
2601 | yi.val, yi.len, precision, | |||
2602 | sgn, overflow)); | |||
2603 | return result; | |||
2604 | } | |||
2605 | ||||
2606 | /* Return X * Y. */ | |||
2607 | template <typename T1, typename T2> | |||
2608 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2609 | wi::mul (const T1 &x, const T2 &y) | |||
2610 | { | |||
2611 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2612 | unsigned int precision = get_precision (result); | |||
2613 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2614 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2615 | if (precision <= HOST_BITS_PER_WIDE_INT64) | |||
2616 | { | |||
2617 | val[0] = xi.ulow () * yi.ulow (); | |||
2618 | result.set_len (1); | |||
2619 | } | |||
2620 | else | |||
2621 | result.set_len (mul_internal (val, xi.val, xi.len, yi.val, yi.len, | |||
2622 | precision, UNSIGNED, 0, false)); | |||
2623 | return result; | |||
2624 | } | |||
2625 | ||||
2626 | /* Return X * Y. Treat X and Y as having the signednes given by SGN | |||
2627 | and indicate in *OVERFLOW whether the operation overflowed. */ | |||
2628 | template <typename T1, typename T2> | |||
2629 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2630 | wi::mul (const T1 &x, const T2 &y, signop sgn, overflow_type *overflow) | |||
2631 | { | |||
2632 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2633 | unsigned int precision = get_precision (result); | |||
2634 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2635 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2636 | result.set_len (mul_internal (val, xi.val, xi.len, | |||
2637 | yi.val, yi.len, precision, | |||
2638 | sgn, overflow, false)); | |||
2639 | return result; | |||
2640 | } | |||
2641 | ||||
2642 | /* Return X * Y, treating both X and Y as signed values. Indicate in | |||
2643 | *OVERFLOW whether the operation overflowed. */ | |||
2644 | template <typename T1, typename T2> | |||
2645 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2646 | wi::smul (const T1 &x, const T2 &y, overflow_type *overflow) | |||
2647 | { | |||
2648 | return mul (x, y, SIGNED, overflow); | |||
2649 | } | |||
2650 | ||||
2651 | /* Return X * Y, treating both X and Y as unsigned values. Indicate in | |||
2652 | *OVERFLOW if the result overflows. */ | |||
2653 | template <typename T1, typename T2> | |||
2654 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2655 | wi::umul (const T1 &x, const T2 &y, overflow_type *overflow) | |||
2656 | { | |||
2657 | return mul (x, y, UNSIGNED, overflow); | |||
2658 | } | |||
2659 | ||||
2660 | /* Perform a widening multiplication of X and Y, extending the values | |||
2661 | according to SGN, and return the high part of the result. */ | |||
2662 | template <typename T1, typename T2> | |||
2663 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2664 | wi::mul_high (const T1 &x, const T2 &y, signop sgn) | |||
2665 | { | |||
2666 | WI_BINARY_RESULT_VAR (result, val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *val = result .write_val (); | |||
2667 | unsigned int precision = get_precision (result); | |||
2668 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2669 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y, precision); | |||
2670 | result.set_len (mul_internal (val, xi.val, xi.len, | |||
2671 | yi.val, yi.len, precision, | |||
2672 | sgn, 0, true)); | |||
2673 | return result; | |||
2674 | } | |||
2675 | ||||
2676 | /* Return X / Y, rouding towards 0. Treat X and Y as having the | |||
2677 | signedness given by SGN. Indicate in *OVERFLOW if the result | |||
2678 | overflows. */ | |||
2679 | template <typename T1, typename T2> | |||
2680 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2681 | wi::div_trunc (const T1 &x, const T2 &y, signop sgn, overflow_type *overflow) | |||
2682 | { | |||
2683 | WI_BINARY_RESULT_VAR (quotient, quotient_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type quotient = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *quotient_val = quotient.write_val (); | |||
2684 | unsigned int precision = get_precision (quotient); | |||
2685 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2686 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
2687 | ||||
2688 | quotient.set_len (divmod_internal (quotient_val, 0, 0, xi.val, xi.len, | |||
2689 | precision, | |||
2690 | yi.val, yi.len, yi.precision, | |||
2691 | sgn, overflow)); | |||
2692 | return quotient; | |||
2693 | } | |||
2694 | ||||
2695 | /* Return X / Y, rouding towards 0. Treat X and Y as signed values. */ | |||
2696 | template <typename T1, typename T2> | |||
2697 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2698 | wi::sdiv_trunc (const T1 &x, const T2 &y) | |||
2699 | { | |||
2700 | return div_trunc (x, y, SIGNED); | |||
2701 | } | |||
2702 | ||||
2703 | /* Return X / Y, rouding towards 0. Treat X and Y as unsigned values. */ | |||
2704 | template <typename T1, typename T2> | |||
2705 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2706 | wi::udiv_trunc (const T1 &x, const T2 &y) | |||
2707 | { | |||
2708 | return div_trunc (x, y, UNSIGNED); | |||
2709 | } | |||
2710 | ||||
2711 | /* Return X / Y, rouding towards -inf. Treat X and Y as having the | |||
2712 | signedness given by SGN. Indicate in *OVERFLOW if the result | |||
2713 | overflows. */ | |||
2714 | template <typename T1, typename T2> | |||
2715 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2716 | wi::div_floor (const T1 &x, const T2 &y, signop sgn, overflow_type *overflow) | |||
2717 | { | |||
2718 | WI_BINARY_RESULT_VAR (quotient, quotient_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type quotient = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *quotient_val = quotient.write_val (); | |||
2719 | WI_BINARY_RESULT_VAR (remainder, remainder_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type remainder = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *remainder_val = remainder.write_val (); | |||
2720 | unsigned int precision = get_precision (quotient); | |||
2721 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2722 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
2723 | ||||
2724 | unsigned int remainder_len; | |||
2725 | quotient.set_len (divmod_internal (quotient_val, | |||
2726 | &remainder_len, remainder_val, | |||
2727 | xi.val, xi.len, precision, | |||
2728 | yi.val, yi.len, yi.precision, sgn, | |||
2729 | overflow)); | |||
2730 | remainder.set_len (remainder_len); | |||
2731 | if (wi::neg_p (x, sgn) != wi::neg_p (y, sgn) && remainder != 0) | |||
2732 | return quotient - 1; | |||
2733 | return quotient; | |||
2734 | } | |||
2735 | ||||
2736 | /* Return X / Y, rouding towards -inf. Treat X and Y as signed values. */ | |||
2737 | template <typename T1, typename T2> | |||
2738 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2739 | wi::sdiv_floor (const T1 &x, const T2 &y) | |||
2740 | { | |||
2741 | return div_floor (x, y, SIGNED); | |||
2742 | } | |||
2743 | ||||
2744 | /* Return X / Y, rouding towards -inf. Treat X and Y as unsigned values. */ | |||
2745 | /* ??? Why do we have both this and udiv_trunc. Aren't they the same? */ | |||
2746 | template <typename T1, typename T2> | |||
2747 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2748 | wi::udiv_floor (const T1 &x, const T2 &y) | |||
2749 | { | |||
2750 | return div_floor (x, y, UNSIGNED); | |||
2751 | } | |||
2752 | ||||
2753 | /* Return X / Y, rouding towards +inf. Treat X and Y as having the | |||
2754 | signedness given by SGN. Indicate in *OVERFLOW if the result | |||
2755 | overflows. */ | |||
2756 | template <typename T1, typename T2> | |||
2757 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2758 | wi::div_ceil (const T1 &x, const T2 &y, signop sgn, overflow_type *overflow) | |||
2759 | { | |||
2760 | WI_BINARY_RESULT_VAR (quotient, quotient_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type quotient = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *quotient_val = quotient.write_val (); | |||
2761 | WI_BINARY_RESULT_VAR (remainder, remainder_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type remainder = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *remainder_val = remainder.write_val (); | |||
2762 | unsigned int precision = get_precision (quotient); | |||
2763 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2764 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
2765 | ||||
2766 | unsigned int remainder_len; | |||
2767 | quotient.set_len (divmod_internal (quotient_val, | |||
2768 | &remainder_len, remainder_val, | |||
2769 | xi.val, xi.len, precision, | |||
2770 | yi.val, yi.len, yi.precision, sgn, | |||
2771 | overflow)); | |||
2772 | remainder.set_len (remainder_len); | |||
2773 | if (wi::neg_p (x, sgn) == wi::neg_p (y, sgn) && remainder != 0) | |||
2774 | return quotient + 1; | |||
2775 | return quotient; | |||
2776 | } | |||
2777 | ||||
2778 | /* Return X / Y, rouding towards +inf. Treat X and Y as unsigned values. */ | |||
2779 | template <typename T1, typename T2> | |||
2780 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2781 | wi::udiv_ceil (const T1 &x, const T2 &y) | |||
2782 | { | |||
2783 | return div_ceil (x, y, UNSIGNED); | |||
2784 | } | |||
2785 | ||||
2786 | /* Return X / Y, rouding towards nearest with ties away from zero. | |||
2787 | Treat X and Y as having the signedness given by SGN. Indicate | |||
2788 | in *OVERFLOW if the result overflows. */ | |||
2789 | template <typename T1, typename T2> | |||
2790 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2791 | wi::div_round (const T1 &x, const T2 &y, signop sgn, overflow_type *overflow) | |||
2792 | { | |||
2793 | WI_BINARY_RESULT_VAR (quotient, quotient_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type quotient = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *quotient_val = quotient.write_val (); | |||
2794 | WI_BINARY_RESULT_VAR (remainder, remainder_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type remainder = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *remainder_val = remainder.write_val (); | |||
2795 | unsigned int precision = get_precision (quotient); | |||
2796 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2797 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
2798 | ||||
2799 | unsigned int remainder_len; | |||
2800 | quotient.set_len (divmod_internal (quotient_val, | |||
2801 | &remainder_len, remainder_val, | |||
2802 | xi.val, xi.len, precision, | |||
2803 | yi.val, yi.len, yi.precision, sgn, | |||
2804 | overflow)); | |||
2805 | remainder.set_len (remainder_len); | |||
2806 | ||||
2807 | if (remainder != 0) | |||
2808 | { | |||
2809 | if (sgn == SIGNED) | |||
2810 | { | |||
2811 | WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type abs_remainder = wi::abs (remainder); | |||
2812 | if (wi::geu_p (abs_remainder, wi::sub (wi::abs (y), abs_remainder))) | |||
2813 | { | |||
2814 | if (wi::neg_p (x, sgn) != wi::neg_p (y, sgn)) | |||
2815 | return quotient - 1; | |||
2816 | else | |||
2817 | return quotient + 1; | |||
2818 | } | |||
2819 | } | |||
2820 | else | |||
2821 | { | |||
2822 | if (wi::geu_p (remainder, wi::sub (y, remainder))) | |||
2823 | return quotient + 1; | |||
2824 | } | |||
2825 | } | |||
2826 | return quotient; | |||
2827 | } | |||
2828 | ||||
2829 | /* Return X / Y, rouding towards 0. Treat X and Y as having the | |||
2830 | signedness given by SGN. Store the remainder in *REMAINDER_PTR. */ | |||
2831 | template <typename T1, typename T2> | |||
2832 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2833 | wi::divmod_trunc (const T1 &x, const T2 &y, signop sgn, | |||
2834 | WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type *remainder_ptr) | |||
2835 | { | |||
2836 | WI_BINARY_RESULT_VAR (quotient, quotient_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type quotient = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *quotient_val = quotient.write_val (); | |||
2837 | WI_BINARY_RESULT_VAR (remainder, remainder_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type remainder = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *remainder_val = remainder.write_val (); | |||
2838 | unsigned int precision = get_precision (quotient); | |||
2839 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2840 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
2841 | ||||
2842 | unsigned int remainder_len; | |||
2843 | quotient.set_len (divmod_internal (quotient_val, | |||
2844 | &remainder_len, remainder_val, | |||
2845 | xi.val, xi.len, precision, | |||
2846 | yi.val, yi.len, yi.precision, sgn, 0)); | |||
2847 | remainder.set_len (remainder_len); | |||
2848 | ||||
2849 | *remainder_ptr = remainder; | |||
2850 | return quotient; | |||
2851 | } | |||
2852 | ||||
2853 | /* Compute the greatest common divisor of two numbers A and B using | |||
2854 | Euclid's algorithm. */ | |||
2855 | template <typename T1, typename T2> | |||
2856 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2857 | wi::gcd (const T1 &a, const T2 &b, signop sgn) | |||
2858 | { | |||
2859 | T1 x, y, z; | |||
2860 | ||||
2861 | x = wi::abs (a); | |||
2862 | y = wi::abs (b); | |||
2863 | ||||
2864 | while (gt_p (x, 0, sgn)) | |||
2865 | { | |||
2866 | z = mod_trunc (y, x, sgn); | |||
2867 | y = x; | |||
2868 | x = z; | |||
2869 | } | |||
2870 | ||||
2871 | return y; | |||
2872 | } | |||
2873 | ||||
2874 | /* Compute X / Y, rouding towards 0, and return the remainder. | |||
2875 | Treat X and Y as having the signedness given by SGN. Indicate | |||
2876 | in *OVERFLOW if the division overflows. */ | |||
2877 | template <typename T1, typename T2> | |||
2878 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2879 | wi::mod_trunc (const T1 &x, const T2 &y, signop sgn, overflow_type *overflow) | |||
2880 | { | |||
2881 | WI_BINARY_RESULT_VAR (remainder, remainder_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type remainder = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *remainder_val = remainder.write_val (); | |||
2882 | unsigned int precision = get_precision (remainder); | |||
2883 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2884 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
2885 | ||||
2886 | unsigned int remainder_len; | |||
2887 | divmod_internal (0, &remainder_len, remainder_val, | |||
2888 | xi.val, xi.len, precision, | |||
2889 | yi.val, yi.len, yi.precision, sgn, overflow); | |||
2890 | remainder.set_len (remainder_len); | |||
2891 | ||||
2892 | return remainder; | |||
2893 | } | |||
2894 | ||||
2895 | /* Compute X / Y, rouding towards 0, and return the remainder. | |||
2896 | Treat X and Y as signed values. */ | |||
2897 | template <typename T1, typename T2> | |||
2898 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2899 | wi::smod_trunc (const T1 &x, const T2 &y) | |||
2900 | { | |||
2901 | return mod_trunc (x, y, SIGNED); | |||
2902 | } | |||
2903 | ||||
2904 | /* Compute X / Y, rouding towards 0, and return the remainder. | |||
2905 | Treat X and Y as unsigned values. */ | |||
2906 | template <typename T1, typename T2> | |||
2907 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2908 | wi::umod_trunc (const T1 &x, const T2 &y) | |||
2909 | { | |||
2910 | return mod_trunc (x, y, UNSIGNED); | |||
2911 | } | |||
2912 | ||||
2913 | /* Compute X / Y, rouding towards -inf, and return the remainder. | |||
2914 | Treat X and Y as having the signedness given by SGN. Indicate | |||
2915 | in *OVERFLOW if the division overflows. */ | |||
2916 | template <typename T1, typename T2> | |||
2917 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2918 | wi::mod_floor (const T1 &x, const T2 &y, signop sgn, overflow_type *overflow) | |||
2919 | { | |||
2920 | WI_BINARY_RESULT_VAR (quotient, quotient_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type quotient = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *quotient_val = quotient.write_val (); | |||
2921 | WI_BINARY_RESULT_VAR (remainder, remainder_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type remainder = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *remainder_val = remainder.write_val (); | |||
2922 | unsigned int precision = get_precision (quotient); | |||
2923 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2924 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
2925 | ||||
2926 | unsigned int remainder_len; | |||
2927 | quotient.set_len (divmod_internal (quotient_val, | |||
2928 | &remainder_len, remainder_val, | |||
2929 | xi.val, xi.len, precision, | |||
2930 | yi.val, yi.len, yi.precision, sgn, | |||
2931 | overflow)); | |||
2932 | remainder.set_len (remainder_len); | |||
2933 | ||||
2934 | if (wi::neg_p (x, sgn) != wi::neg_p (y, sgn) && remainder != 0) | |||
2935 | return remainder + y; | |||
2936 | return remainder; | |||
2937 | } | |||
2938 | ||||
2939 | /* Compute X / Y, rouding towards -inf, and return the remainder. | |||
2940 | Treat X and Y as unsigned values. */ | |||
2941 | /* ??? Why do we have both this and umod_trunc. Aren't they the same? */ | |||
2942 | template <typename T1, typename T2> | |||
2943 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2944 | wi::umod_floor (const T1 &x, const T2 &y) | |||
2945 | { | |||
2946 | return mod_floor (x, y, UNSIGNED); | |||
2947 | } | |||
2948 | ||||
2949 | /* Compute X / Y, rouding towards +inf, and return the remainder. | |||
2950 | Treat X and Y as having the signedness given by SGN. Indicate | |||
2951 | in *OVERFLOW if the division overflows. */ | |||
2952 | template <typename T1, typename T2> | |||
2953 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2954 | wi::mod_ceil (const T1 &x, const T2 &y, signop sgn, overflow_type *overflow) | |||
2955 | { | |||
2956 | WI_BINARY_RESULT_VAR (quotient, quotient_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type quotient = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *quotient_val = quotient.write_val (); | |||
2957 | WI_BINARY_RESULT_VAR (remainder, remainder_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type remainder = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *remainder_val = remainder.write_val (); | |||
2958 | unsigned int precision = get_precision (quotient); | |||
2959 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2960 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
2961 | ||||
2962 | unsigned int remainder_len; | |||
2963 | quotient.set_len (divmod_internal (quotient_val, | |||
2964 | &remainder_len, remainder_val, | |||
2965 | xi.val, xi.len, precision, | |||
2966 | yi.val, yi.len, yi.precision, sgn, | |||
2967 | overflow)); | |||
2968 | remainder.set_len (remainder_len); | |||
2969 | ||||
2970 | if (wi::neg_p (x, sgn) == wi::neg_p (y, sgn) && remainder != 0) | |||
2971 | return remainder - y; | |||
2972 | return remainder; | |||
2973 | } | |||
2974 | ||||
2975 | /* Compute X / Y, rouding towards nearest with ties away from zero, | |||
2976 | and return the remainder. Treat X and Y as having the signedness | |||
2977 | given by SGN. Indicate in *OVERFLOW if the division overflows. */ | |||
2978 | template <typename T1, typename T2> | |||
2979 | inline WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type | |||
2980 | wi::mod_round (const T1 &x, const T2 &y, signop sgn, overflow_type *overflow) | |||
2981 | { | |||
2982 | WI_BINARY_RESULT_VAR (quotient, quotient_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type quotient = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *quotient_val = quotient.write_val (); | |||
2983 | WI_BINARY_RESULT_VAR (remainder, remainder_val, T1, x, T2, y)typename wi::binary_traits <T1, T2>::result_type remainder = wi::int_traits <typename wi::binary_traits <T1, T2> ::result_type>::get_binary_result (x, y); long *remainder_val = remainder.write_val (); | |||
2984 | unsigned int precision = get_precision (quotient); | |||
2985 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
2986 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
2987 | ||||
2988 | unsigned int remainder_len; | |||
2989 | quotient.set_len (divmod_internal (quotient_val, | |||
2990 | &remainder_len, remainder_val, | |||
2991 | xi.val, xi.len, precision, | |||
2992 | yi.val, yi.len, yi.precision, sgn, | |||
2993 | overflow)); | |||
2994 | remainder.set_len (remainder_len); | |||
2995 | ||||
2996 | if (remainder != 0) | |||
2997 | { | |||
2998 | if (sgn == SIGNED) | |||
2999 | { | |||
3000 | WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type abs_remainder = wi::abs (remainder); | |||
3001 | if (wi::geu_p (abs_remainder, wi::sub (wi::abs (y), abs_remainder))) | |||
3002 | { | |||
3003 | if (wi::neg_p (x, sgn) != wi::neg_p (y, sgn)) | |||
3004 | return remainder + y; | |||
3005 | else | |||
3006 | return remainder - y; | |||
3007 | } | |||
3008 | } | |||
3009 | else | |||
3010 | { | |||
3011 | if (wi::geu_p (remainder, wi::sub (y, remainder))) | |||
3012 | return remainder - y; | |||
3013 | } | |||
3014 | } | |||
3015 | return remainder; | |||
3016 | } | |||
3017 | ||||
3018 | /* Return true if X is a multiple of Y. Treat X and Y as having the | |||
3019 | signedness given by SGN. */ | |||
3020 | template <typename T1, typename T2> | |||
3021 | inline bool | |||
3022 | wi::multiple_of_p (const T1 &x, const T2 &y, signop sgn) | |||
3023 | { | |||
3024 | return wi::mod_trunc (x, y, sgn) == 0; | |||
3025 | } | |||
3026 | ||||
3027 | /* Return true if X is a multiple of Y, storing X / Y in *RES if so. | |||
3028 | Treat X and Y as having the signedness given by SGN. */ | |||
3029 | template <typename T1, typename T2> | |||
3030 | inline bool | |||
3031 | wi::multiple_of_p (const T1 &x, const T2 &y, signop sgn, | |||
3032 | WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type *res) | |||
3033 | { | |||
3034 | WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type remainder; | |||
3035 | WI_BINARY_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::result_type quotient | |||
3036 | = divmod_trunc (x, y, sgn, &remainder); | |||
3037 | if (remainder == 0) | |||
3038 | { | |||
3039 | *res = quotient; | |||
3040 | return true; | |||
3041 | } | |||
3042 | return false; | |||
3043 | } | |||
3044 | ||||
3045 | /* Return X << Y. Return 0 if Y is greater than or equal to | |||
3046 | the precision of X. */ | |||
3047 | template <typename T1, typename T2> | |||
3048 | inline WI_UNARY_RESULT (T1)typename wi::binary_traits <T1, T1>::result_type | |||
3049 | wi::lshift (const T1 &x, const T2 &y) | |||
3050 | { | |||
3051 | WI_UNARY_RESULT_VAR (result, val, T1, x)typename wi::binary_traits <T1, T1>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T1> ::result_type>::get_binary_result (x, x); long *val = result .write_val (); | |||
3052 | unsigned int precision = get_precision (result); | |||
3053 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x, precision); | |||
3054 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
3055 | /* Handle the simple cases quickly. */ | |||
3056 | if (geu_p (yi, precision)) | |||
3057 | { | |||
3058 | val[0] = 0; | |||
3059 | result.set_len (1); | |||
3060 | } | |||
3061 | else | |||
3062 | { | |||
3063 | unsigned int shift = yi.to_uhwi (); | |||
3064 | /* For fixed-precision integers like offset_int and widest_int, | |||
3065 | handle the case where the shift value is constant and the | |||
3066 | result is a single nonnegative HWI (meaning that we don't | |||
3067 | need to worry about val[1]). This is particularly common | |||
3068 | for converting a byte count to a bit count. | |||
3069 | ||||
3070 | For variable-precision integers like wide_int, handle HWI | |||
3071 | and sub-HWI integers inline. */ | |||
3072 | if (STATIC_CONSTANT_P (xi.precision > HOST_BITS_PER_WIDE_INT)(__builtin_constant_p (xi.precision > 64) && (xi.precision > 64)) | |||
3073 | ? (STATIC_CONSTANT_P (shift < HOST_BITS_PER_WIDE_INT - 1)(__builtin_constant_p (shift < 64 - 1) && (shift < 64 - 1)) | |||
3074 | && xi.len == 1 | |||
3075 | && IN_RANGE (xi.val[0], 0, HOST_WIDE_INT_MAX >> shift)((unsigned long) (xi.val[0]) - (unsigned long) (0) <= (unsigned long) ((~((long) (1UL << (64 - 1)))) >> shift) - (unsigned long) (0))) | |||
3076 | : precision <= HOST_BITS_PER_WIDE_INT64) | |||
3077 | { | |||
3078 | val[0] = xi.ulow () << shift; | |||
3079 | result.set_len (1); | |||
3080 | } | |||
3081 | else | |||
3082 | result.set_len (lshift_large (val, xi.val, xi.len, | |||
3083 | precision, shift)); | |||
3084 | } | |||
3085 | return result; | |||
3086 | } | |||
3087 | ||||
3088 | /* Return X >> Y, using a logical shift. Return 0 if Y is greater than | |||
3089 | or equal to the precision of X. */ | |||
3090 | template <typename T1, typename T2> | |||
3091 | inline WI_UNARY_RESULT (T1)typename wi::binary_traits <T1, T1>::result_type | |||
3092 | wi::lrshift (const T1 &x, const T2 &y) | |||
3093 | { | |||
3094 | WI_UNARY_RESULT_VAR (result, val, T1, x)typename wi::binary_traits <T1, T1>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T1> ::result_type>::get_binary_result (x, x); long *val = result .write_val (); | |||
3095 | /* Do things in the precision of the input rather than the output, | |||
3096 | since the result can be no larger than that. */ | |||
3097 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x); | |||
3098 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
3099 | /* Handle the simple cases quickly. */ | |||
3100 | if (geu_p (yi, xi.precision)) | |||
3101 | { | |||
3102 | val[0] = 0; | |||
3103 | result.set_len (1); | |||
3104 | } | |||
3105 | else | |||
3106 | { | |||
3107 | unsigned int shift = yi.to_uhwi (); | |||
3108 | /* For fixed-precision integers like offset_int and widest_int, | |||
3109 | handle the case where the shift value is constant and the | |||
3110 | shifted value is a single nonnegative HWI (meaning that all | |||
3111 | bits above the HWI are zero). This is particularly common | |||
3112 | for converting a bit count to a byte count. | |||
3113 | ||||
3114 | For variable-precision integers like wide_int, handle HWI | |||
3115 | and sub-HWI integers inline. */ | |||
3116 | if (STATIC_CONSTANT_P (xi.precision > HOST_BITS_PER_WIDE_INT)(__builtin_constant_p (xi.precision > 64) && (xi.precision > 64)) | |||
3117 | ? (shift < HOST_BITS_PER_WIDE_INT64 | |||
3118 | && xi.len == 1 | |||
3119 | && xi.val[0] >= 0) | |||
3120 | : xi.precision <= HOST_BITS_PER_WIDE_INT64) | |||
3121 | { | |||
3122 | val[0] = xi.to_uhwi () >> shift; | |||
3123 | result.set_len (1); | |||
3124 | } | |||
3125 | else | |||
3126 | result.set_len (lrshift_large (val, xi.val, xi.len, xi.precision, | |||
3127 | get_precision (result), shift)); | |||
3128 | } | |||
3129 | return result; | |||
3130 | } | |||
3131 | ||||
3132 | /* Return X >> Y, using an arithmetic shift. Return a sign mask if | |||
3133 | Y is greater than or equal to the precision of X. */ | |||
3134 | template <typename T1, typename T2> | |||
3135 | inline WI_UNARY_RESULT (T1)typename wi::binary_traits <T1, T1>::result_type | |||
3136 | wi::arshift (const T1 &x, const T2 &y) | |||
3137 | { | |||
3138 | WI_UNARY_RESULT_VAR (result, val, T1, x)typename wi::binary_traits <T1, T1>::result_type result = wi::int_traits <typename wi::binary_traits <T1, T1> ::result_type>::get_binary_result (x, x); long *val = result .write_val (); | |||
3139 | /* Do things in the precision of the input rather than the output, | |||
3140 | since the result can be no larger than that. */ | |||
3141 | WIDE_INT_REF_FOR (T1)generic_wide_int <wide_int_ref_storage <wi::int_traits < T1>::is_sign_extended, wi::int_traits <T1>::host_dependent_precision > > xi (x); | |||
3142 | WIDE_INT_REF_FOR (T2)generic_wide_int <wide_int_ref_storage <wi::int_traits < T2>::is_sign_extended, wi::int_traits <T2>::host_dependent_precision > > yi (y); | |||
3143 | /* Handle the simple cases quickly. */ | |||
3144 | if (geu_p (yi, xi.precision)) | |||
3145 | { | |||
3146 | val[0] = sign_mask (x); | |||
3147 | result.set_len (1); | |||
3148 | } | |||
3149 | else | |||
3150 | { | |||
3151 | unsigned int shift = yi.to_uhwi (); | |||
3152 | if (xi.precision <= HOST_BITS_PER_WIDE_INT64) | |||
3153 | { | |||
3154 | val[0] = sext_hwi (xi.ulow () >> shift, xi.precision - shift); | |||
3155 | result.set_len (1, true); | |||
3156 | } | |||
3157 | else | |||
3158 | result.set_len (arshift_large (val, xi.val, xi.len, xi.precision, | |||
3159 | get_precision (result), shift)); | |||
3160 | } | |||
3161 | return result; | |||
3162 | } | |||
3163 | ||||
3164 | /* Return X >> Y, using an arithmetic shift if SGN is SIGNED and a | |||
3165 | logical shift otherwise. */ | |||
3166 | template <typename T1, typename T2> | |||
3167 | inline WI_UNARY_RESULT (T1)typename wi::binary_traits <T1, T1>::result_type | |||
3168 | wi::rshift (const T1 &x, const T2 &y, signop sgn) | |||
3169 | { | |||
3170 | if (sgn == UNSIGNED) | |||
3171 | return lrshift (x, y); | |||
3172 | else | |||
3173 | return arshift (x, y); | |||
3174 | } | |||
3175 | ||||
3176 | /* Return the result of rotating the low WIDTH bits of X left by Y | |||
3177 | bits and zero-extending the result. Use a full-width rotate if | |||
3178 | WIDTH is zero. */ | |||
3179 | template <typename T1, typename T2> | |||
3180 | WI_UNARY_RESULT (T1)typename wi::binary_traits <T1, T1>::result_type | |||
3181 | wi::lrotate (const T1 &x, const T2 &y, unsigned int width) | |||
3182 | { | |||
3183 | unsigned int precision = get_binary_precision (x, x); | |||
3184 | if (width == 0) | |||
3185 | width = precision; | |||
3186 | WI_UNARY_RESULT (T2)typename wi::binary_traits <T2, T2>::result_type ymod = umod_trunc (y, width); | |||
3187 | WI_UNARY_RESULT (T1)typename wi::binary_traits <T1, T1>::result_type left = wi::lshift (x, ymod); | |||
3188 | WI_UNARY_RESULT (T1)typename wi::binary_traits <T1, T1>::result_type right = wi::lrshift (x, wi::sub (width, ymod)); | |||
3189 | if (width != precision) | |||
3190 | return wi::zext (left, width) | wi::zext (right, width); | |||
3191 | return left | right; | |||
3192 | } | |||
3193 | ||||
3194 | /* Return the result of rotating the low WIDTH bits of X right by Y | |||
3195 | bits and zero-extending the result. Use a full-width rotate if | |||
3196 | WIDTH is zero. */ | |||
3197 | template <typename T1, typename T2> | |||
3198 | WI_UNARY_RESULT (T1)typename wi::binary_traits <T1, T1>::result_type | |||
3199 | wi::rrotate (const T1 &x, const T2 &y, unsigned int width) | |||
3200 | { | |||
3201 | unsigned int precision = get_binary_precision (x, x); | |||
3202 | if (width == 0) | |||
3203 | width = precision; | |||
3204 | WI_UNARY_RESULT (T2)typename wi::binary_traits <T2, T2>::result_type ymod = umod_trunc (y, width); | |||
3205 | WI_UNARY_RESULT (T1)typename wi::binary_traits <T1, T1>::result_type right = wi::lrshift (x, ymod); | |||
3206 | WI_UNARY_RESULT (T1)typename wi::binary_traits <T1, T1>::result_type left = wi::lshift (x, wi::sub (width, ymod)); | |||
3207 | if (width != precision) | |||
3208 | return wi::zext (left, width) | wi::zext (right, width); | |||
3209 | return left | right; | |||
3210 | } | |||
3211 | ||||
3212 | /* Return 0 if the number of 1s in X is even and 1 if the number of 1s | |||
3213 | is odd. */ | |||
3214 | inline int | |||
3215 | wi::parity (const wide_int_ref &x) | |||
3216 | { | |||
3217 | return popcount (x) & 1; | |||
3218 | } | |||
3219 | ||||
3220 | /* Extract WIDTH bits from X, starting at BITPOS. */ | |||
3221 | template <typename T> | |||
3222 | inline unsigned HOST_WIDE_INTlong | |||
3223 | wi::extract_uhwi (const T &x, unsigned int bitpos, unsigned int width) | |||
3224 | { | |||
3225 | unsigned precision = get_precision (x); | |||
3226 | if (precision < bitpos + width) | |||
3227 | precision = bitpos + width; | |||
3228 | WIDE_INT_REF_FOR (T)generic_wide_int <wide_int_ref_storage <wi::int_traits < T>::is_sign_extended, wi::int_traits <T>::host_dependent_precision > > xi (x, precision); | |||
3229 | ||||
3230 | /* Handle this rare case after the above, so that we assert about | |||
3231 | bogus BITPOS values. */ | |||
3232 | if (width == 0) | |||
3233 | return 0; | |||
3234 | ||||
3235 | unsigned int start = bitpos / HOST_BITS_PER_WIDE_INT64; | |||
3236 | unsigned int shift = bitpos % HOST_BITS_PER_WIDE_INT64; | |||
3237 | unsigned HOST_WIDE_INTlong res = xi.elt (start); | |||
3238 | res >>= shift; | |||
3239 | if (shift + width > HOST_BITS_PER_WIDE_INT64) | |||
3240 | { | |||
3241 | unsigned HOST_WIDE_INTlong upper = xi.elt (start + 1); | |||
3242 | res |= upper << (-shift % HOST_BITS_PER_WIDE_INT64); | |||
3243 | } | |||
3244 | return zext_hwi (res, width); | |||
3245 | } | |||
3246 | ||||
3247 | /* Return the minimum precision needed to store X with sign SGN. */ | |||
3248 | template <typename T> | |||
3249 | inline unsigned int | |||
3250 | wi::min_precision (const T &x, signop sgn) | |||
3251 | { | |||
3252 | if (sgn == SIGNED) | |||
3253 | return get_precision (x) - clrsb (x); | |||
3254 | else | |||
3255 | return get_precision (x) - clz (x); | |||
3256 | } | |||
3257 | ||||
3258 | #define SIGNED_BINARY_PREDICATE(OP, F) \ | |||
3259 | template <typename T1, typename T2> \ | |||
3260 | inline WI_SIGNED_BINARY_PREDICATE_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::signed_predicate_result \ | |||
3261 | OP (const T1 &x, const T2 &y) \ | |||
3262 | { \ | |||
3263 | return wi::F (x, y); \ | |||
3264 | } | |||
3265 | ||||
3266 | SIGNED_BINARY_PREDICATE (operator <, lts_p) | |||
3267 | SIGNED_BINARY_PREDICATE (operator <=, les_p) | |||
3268 | SIGNED_BINARY_PREDICATE (operator >, gts_p) | |||
3269 | SIGNED_BINARY_PREDICATE (operator >=, ges_p) | |||
3270 | ||||
3271 | #undef SIGNED_BINARY_PREDICATE | |||
3272 | ||||
3273 | #define UNARY_OPERATOR(OP, F) \ | |||
3274 | template<typename T> \ | |||
3275 | WI_UNARY_RESULT (generic_wide_int<T>)typename wi::binary_traits <generic_wide_int<T>, generic_wide_int <T> >::result_type \ | |||
3276 | OP (const generic_wide_int<T> &x) \ | |||
3277 | { \ | |||
3278 | return wi::F (x); \ | |||
3279 | } | |||
3280 | ||||
3281 | #define BINARY_PREDICATE(OP, F) \ | |||
3282 | template<typename T1, typename T2> \ | |||
3283 | WI_BINARY_PREDICATE_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::predicate_result \ | |||
3284 | OP (const T1 &x, const T2 &y) \ | |||
3285 | { \ | |||
3286 | return wi::F (x, y); \ | |||
3287 | } | |||
3288 | ||||
3289 | #define BINARY_OPERATOR(OP, F) \ | |||
3290 | template<typename T1, typename T2> \ | |||
3291 | WI_BINARY_OPERATOR_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::operator_result \ | |||
3292 | OP (const T1 &x, const T2 &y) \ | |||
3293 | { \ | |||
3294 | return wi::F (x, y); \ | |||
3295 | } | |||
3296 | ||||
3297 | #define SHIFT_OPERATOR(OP, F) \ | |||
3298 | template<typename T1, typename T2> \ | |||
3299 | WI_BINARY_OPERATOR_RESULT (T1, T1)typename wi::binary_traits <T1, T1>::operator_result \ | |||
3300 | OP (const T1 &x, const T2 &y) \ | |||
3301 | { \ | |||
3302 | return wi::F (x, y); \ | |||
3303 | } | |||
3304 | ||||
3305 | UNARY_OPERATOR (operator ~, bit_not) | |||
3306 | UNARY_OPERATOR (operator -, neg) | |||
3307 | BINARY_PREDICATE (operator ==, eq_p) | |||
3308 | BINARY_PREDICATE (operator !=, ne_p) | |||
3309 | BINARY_OPERATOR (operator &, bit_and) | |||
3310 | BINARY_OPERATOR (operator |, bit_or) | |||
3311 | BINARY_OPERATOR (operator ^, bit_xor) | |||
3312 | BINARY_OPERATOR (operator +, add) | |||
3313 | BINARY_OPERATOR (operator -, sub) | |||
3314 | BINARY_OPERATOR (operator *, mul) | |||
3315 | SHIFT_OPERATOR (operator <<, lshift) | |||
3316 | ||||
3317 | #undef UNARY_OPERATOR | |||
3318 | #undef BINARY_PREDICATE | |||
3319 | #undef BINARY_OPERATOR | |||
3320 | #undef SHIFT_OPERATOR | |||
3321 | ||||
3322 | template <typename T1, typename T2> | |||
3323 | inline WI_SIGNED_SHIFT_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::signed_shift_result_type | |||
3324 | operator >> (const T1 &x, const T2 &y) | |||
3325 | { | |||
3326 | return wi::arshift (x, y); | |||
3327 | } | |||
3328 | ||||
3329 | template <typename T1, typename T2> | |||
3330 | inline WI_SIGNED_SHIFT_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::signed_shift_result_type | |||
3331 | operator / (const T1 &x, const T2 &y) | |||
3332 | { | |||
3333 | return wi::sdiv_trunc (x, y); | |||
3334 | } | |||
3335 | ||||
3336 | template <typename T1, typename T2> | |||
3337 | inline WI_SIGNED_SHIFT_RESULT (T1, T2)typename wi::binary_traits <T1, T2>::signed_shift_result_type | |||
3338 | operator % (const T1 &x, const T2 &y) | |||
3339 | { | |||
3340 | return wi::smod_trunc (x, y); | |||
3341 | } | |||
3342 | ||||
3343 | template<typename T> | |||
3344 | void | |||
3345 | gt_ggc_mx (generic_wide_int <T> *) | |||
3346 | { | |||
3347 | } | |||
3348 | ||||
3349 | template<typename T> | |||
3350 | void | |||
3351 | gt_pch_nx (generic_wide_int <T> *) | |||
3352 | { | |||
3353 | } | |||
3354 | ||||
3355 | template<typename T> | |||
3356 | void | |||
3357 | gt_pch_nx (generic_wide_int <T> *, gt_pointer_operator, void *) | |||
3358 | { | |||
3359 | } | |||
3360 | ||||
3361 | template<int N> | |||
3362 | void | |||
3363 | gt_ggc_mx (trailing_wide_ints <N> *) | |||
3364 | { | |||
3365 | } | |||
3366 | ||||
3367 | template<int N> | |||
3368 | void | |||
3369 | gt_pch_nx (trailing_wide_ints <N> *) | |||
3370 | { | |||
3371 | } | |||
3372 | ||||
3373 | template<int N> | |||
3374 | void | |||
3375 | gt_pch_nx (trailing_wide_ints <N> *, gt_pointer_operator, void *) | |||
3376 | { | |||
3377 | } | |||
3378 | ||||
3379 | namespace wi | |||
3380 | { | |||
3381 | /* Used for overloaded functions in which the only other acceptable | |||
3382 | scalar type is a pointer. It stops a plain 0 from being treated | |||
3383 | as a null pointer. */ | |||
3384 | struct never_used1 {}; | |||
3385 | struct never_used2 {}; | |||
3386 | ||||
3387 | wide_int min_value (unsigned int, signop); | |||
3388 | wide_int min_value (never_used1 *); | |||
3389 | wide_int min_value (never_used2 *); | |||
3390 | wide_int max_value (unsigned int, signop); | |||
3391 | wide_int max_value (never_used1 *); | |||
3392 | wide_int max_value (never_used2 *); | |||
3393 | ||||
3394 | /* FIXME: this is target dependent, so should be elsewhere. | |||
3395 | It also seems to assume that CHAR_BIT == BITS_PER_UNIT. */ | |||
3396 | wide_int from_buffer (const unsigned char *, unsigned int); | |||
3397 | ||||
3398 | #ifndef GENERATOR_FILE | |||
3399 | void to_mpz (const wide_int_ref &, mpz_t, signop); | |||
3400 | #endif | |||
3401 | ||||
3402 | wide_int mask (unsigned int, bool, unsigned int); | |||
3403 | wide_int shifted_mask (unsigned int, unsigned int, bool, unsigned int); | |||
3404 | wide_int set_bit_in_zero (unsigned int, unsigned int); | |||
3405 | wide_int insert (const wide_int &x, const wide_int &y, unsigned int, | |||
3406 | unsigned int); | |||
3407 | wide_int round_down_for_mask (const wide_int &, const wide_int &); | |||
3408 | wide_int round_up_for_mask (const wide_int &, const wide_int &); | |||
3409 | ||||
3410 | wide_int mod_inv (const wide_int &a, const wide_int &b); | |||
3411 | ||||
3412 | template <typename T> | |||
3413 | T mask (unsigned int, bool); | |||
3414 | ||||
3415 | template <typename T> | |||
3416 | T shifted_mask (unsigned int, unsigned int, bool); | |||
3417 | ||||
3418 | template <typename T> | |||
3419 | T set_bit_in_zero (unsigned int); | |||
3420 | ||||
3421 | unsigned int mask (HOST_WIDE_INTlong *, unsigned int, bool, unsigned int); | |||
3422 | unsigned int shifted_mask (HOST_WIDE_INTlong *, unsigned int, unsigned int, | |||
3423 | bool, unsigned int); | |||
3424 | unsigned int from_array (HOST_WIDE_INTlong *, const HOST_WIDE_INTlong *, | |||
3425 | unsigned int, unsigned int, bool); | |||
3426 | } | |||
3427 | ||||
3428 | /* Return a PRECISION-bit integer in which the low WIDTH bits are set | |||
3429 | and the other bits are clear, or the inverse if NEGATE_P. */ | |||
3430 | inline wide_int | |||
3431 | wi::mask (unsigned int width, bool negate_p, unsigned int precision) | |||
3432 | { | |||
3433 | wide_int result = wide_int::create (precision); | |||
3434 | result.set_len (mask (result.write_val (), width, negate_p, precision)); | |||
3435 | return result; | |||
3436 | } | |||
3437 | ||||
3438 | /* Return a PRECISION-bit integer in which the low START bits are clear, | |||
3439 | the next WIDTH bits are set, and the other bits are clear, | |||
3440 | or the inverse if NEGATE_P. */ | |||
3441 | inline wide_int | |||
3442 | wi::shifted_mask (unsigned int start, unsigned int width, bool negate_p, | |||
3443 | unsigned int precision) | |||
3444 | { | |||
3445 | wide_int result = wide_int::create (precision); | |||
3446 | result.set_len (shifted_mask (result.write_val (), start, width, negate_p, | |||
3447 | precision)); | |||
3448 | return result; | |||
3449 | } | |||
3450 | ||||
3451 | /* Return a PRECISION-bit integer in which bit BIT is set and all the | |||
3452 | others are clear. */ | |||
3453 | inline wide_int | |||
3454 | wi::set_bit_in_zero (unsigned int bit, unsigned int precision) | |||
3455 | { | |||
3456 | return shifted_mask (bit, 1, false, precision); | |||
3457 | } | |||
3458 | ||||
3459 | /* Return an integer of type T in which the low WIDTH bits are set | |||
3460 | and the other bits are clear, or the inverse if NEGATE_P. */ | |||
3461 | template <typename T> | |||
3462 | inline T | |||
3463 | wi::mask (unsigned int width, bool negate_p) | |||
3464 | { | |||
3465 | STATIC_ASSERT (wi::int_traits<T>::precision)static_assert ((wi::int_traits<T>::precision), "wi::int_traits<T>::precision" ); | |||
3466 | T result; | |||
3467 | result.set_len (mask (result.write_val (), width, negate_p, | |||
3468 | wi::int_traits <T>::precision)); | |||
3469 | return result; | |||
3470 | } | |||
3471 | ||||
3472 | /* Return an integer of type T in which the low START bits are clear, | |||
3473 | the next WIDTH bits are set, and the other bits are clear, or the | |||
3474 | inverse if NEGATE_P. */ | |||
3475 | template <typename T> | |||
3476 | inline T | |||
3477 | wi::shifted_mask (unsigned int start, unsigned int width, bool negate_p) | |||
3478 | { | |||
3479 | STATIC_ASSERT (wi::int_traits<T>::precision)static_assert ((wi::int_traits<T>::precision), "wi::int_traits<T>::precision" ); | |||
3480 | T result; | |||
3481 | result.set_len (shifted_mask (result.write_val (), start, width, | |||
3482 | negate_p, | |||
3483 | wi::int_traits <T>::precision)); | |||
3484 | return result; | |||
3485 | } | |||
3486 | ||||
3487 | /* Return an integer of type T in which bit BIT is set and all the | |||
3488 | others are clear. */ | |||
3489 | template <typename T> | |||
3490 | inline T | |||
3491 | wi::set_bit_in_zero (unsigned int bit) | |||
3492 | { | |||
3493 | return shifted_mask <T> (bit, 1, false); | |||
3494 | } | |||
3495 | ||||
3496 | /* Accumulate a set of overflows into OVERFLOW. */ | |||
3497 | ||||
3498 | inline void | |||
3499 | wi::accumulate_overflow (wi::overflow_type &overflow, | |||
3500 | wi::overflow_type suboverflow) | |||
3501 | { | |||
3502 | if (!suboverflow) | |||
3503 | return; | |||
3504 | if (!overflow) | |||
3505 | overflow = suboverflow; | |||
3506 | else if (overflow != suboverflow) | |||
3507 | overflow = wi::OVF_UNKNOWN; | |||
3508 | } | |||
3509 | ||||
3510 | #endif /* WIDE_INT_H */ |