gem5  v20.0.0.0
All Classes Namespaces Files Functions Variables Typedefs Enumerations Enumerator Friends Macros Modules Pages
vfp.cc
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2010-2013, 2019 ARM Limited
3  * All rights reserved
4  *
5  * The license below extends only to copyright in the software and shall
6  * not be construed as granting a license to any other intellectual
7  * property including but not limited to intellectual property relating
8  * to a hardware implementation of the functionality of the software
9  * licensed hereunder. You may use the software subject to the license
10  * terms below provided that you ensure that this notice is replicated
11  * unmodified and in its entirety in all distributions of the software,
12  * modified or unmodified, in source code or in binary form.
13  *
14  * Redistribution and use in source and binary forms, with or without
15  * modification, are permitted provided that the following conditions are
16  * met: redistributions of source code must retain the above copyright
17  * notice, this list of conditions and the following disclaimer;
18  * redistributions in binary form must reproduce the above copyright
19  * notice, this list of conditions and the following disclaimer in the
20  * documentation and/or other materials provided with the distribution;
21  * neither the name of the copyright holders nor the names of its
22  * contributors may be used to endorse or promote products derived from
23  * this software without specific prior written permission.
24  *
25  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
26  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
27  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
28  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
29  * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
30  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
31  * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
32  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
33  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
34  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
35  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
36  */
37 
38 #include "arch/arm/insts/vfp.hh"
39 
40 /*
41  * The asm statements below are to keep gcc from reordering code. Otherwise
42  * the rounding mode might be set after the operation it was intended for, the
43  * exception bits read before it, etc.
44  */
45 
46 std::string
47 FpCondCompRegOp::generateDisassembly(
48  Addr pc, const Loader::SymbolTable *symtab) const
49 {
50  std::stringstream ss;
51  printMnemonic(ss, "", false);
52  printIntReg(ss, op1);
53  ccprintf(ss, ", ");
54  printIntReg(ss, op2);
55  ccprintf(ss, ", #%d", defCc);
56  ccprintf(ss, ", ");
57  printCondition(ss, condCode, true);
58  return ss.str();
59 }
60 
61 std::string
63  Addr pc, const Loader::SymbolTable *symtab) const
64 {
65  std::stringstream ss;
66  printMnemonic(ss, "", false);
67  printIntReg(ss, dest);
68  ccprintf(ss, ", ");
69  printIntReg(ss, op1);
70  ccprintf(ss, ", ");
71  printIntReg(ss, op2);
72  ccprintf(ss, ", ");
73  printCondition(ss, condCode, true);
74  return ss.str();
75 }
76 
77 std::string
79  Addr pc, const Loader::SymbolTable *symtab) const
80 {
81  std::stringstream ss;
82  printMnemonic(ss);
83  printFloatReg(ss, dest);
84  ss << ", ";
85  printFloatReg(ss, op1);
86  return ss.str();
87 }
88 
89 std::string
91  Addr pc, const Loader::SymbolTable *symtab) const
92 {
93  std::stringstream ss;
94  printMnemonic(ss);
95  printFloatReg(ss, dest);
96  ccprintf(ss, ", #%d", imm);
97  return ss.str();
98 }
99 
100 std::string
102  Addr pc, const Loader::SymbolTable *symtab) const
103 {
104  std::stringstream ss;
105  printMnemonic(ss);
106  printFloatReg(ss, dest);
107  ss << ", ";
108  printFloatReg(ss, op1);
109  ccprintf(ss, ", #%d", imm);
110  return ss.str();
111 }
112 
113 std::string
115  Addr pc, const Loader::SymbolTable *symtab) const
116 {
117  std::stringstream ss;
118  printMnemonic(ss);
119  printFloatReg(ss, dest);
120  ss << ", ";
121  printFloatReg(ss, op1);
122  ss << ", ";
123  printFloatReg(ss, op2);
124  return ss.str();
125 }
126 
127 std::string
129  Addr pc, const Loader::SymbolTable *symtab)
130  const
131 {
132  std::stringstream ss;
133  printMnemonic(ss);
134  printCondition(ss, cond);
135  printFloatReg(ss, dest);
136  ss << ", ";
137  printFloatReg(ss, op1);
138  ss << ", ";
139  printFloatReg(ss, op2);
140  return ss.str();
141 }
142 
143 std::string
145  Addr pc, const Loader::SymbolTable *symtab) const
146 {
147  std::stringstream ss;
148  printMnemonic(ss);
149  printFloatReg(ss, dest);
150  ss << ", ";
151  printFloatReg(ss, op1);
152  ss << ", ";
153  printFloatReg(ss, op2);
154  ss << ", ";
155  printFloatReg(ss, op3);
156  return ss.str();
157 }
158 
159 std::string
161  Addr pc, const Loader::SymbolTable *symtab) const
162 {
163  std::stringstream ss;
164  printMnemonic(ss);
165  printFloatReg(ss, dest);
166  ss << ", ";
167  printFloatReg(ss, op1);
168  ss << ", ";
169  printFloatReg(ss, op2);
170  ccprintf(ss, ", #%d", imm);
171  return ss.str();
172 }
173 
174 namespace ArmISA
175 {
176 
179 {
180  int roundingMode = fegetround();
181  feclearexcept(FeAllExceptions);
182  switch (rMode) {
183  case VfpRoundNearest:
184  fesetround(FeRoundNearest);
185  break;
186  case VfpRoundUpward:
187  fesetround(FeRoundUpward);
188  break;
189  case VfpRoundDown:
190  fesetround(FeRoundDown);
191  break;
192  case VfpRoundZero:
193  fesetround(FeRoundZero);
194  break;
195  }
196  return roundingMode;
197 }
198 
199 void
200 finishVfp(FPSCR &fpscr, VfpSavedState state, bool flush, FPSCR mask)
201 {
202  int exceptions = fetestexcept(FeAllExceptions);
203  bool underflow = false;
204  if ((exceptions & FeInvalid) && mask.ioc) {
205  fpscr.ioc = 1;
206  }
207  if ((exceptions & FeDivByZero) && mask.dzc) {
208  fpscr.dzc = 1;
209  }
210  if ((exceptions & FeOverflow) && mask.ofc) {
211  fpscr.ofc = 1;
212  }
213  if (exceptions & FeUnderflow) {
214  underflow = true;
215  if (mask.ufc)
216  fpscr.ufc = 1;
217  }
218  if ((exceptions & FeInexact) && !(underflow && flush) && mask.ixc) {
219  fpscr.ixc = 1;
220  }
221  fesetround(state);
222 }
223 
224 template <class fpType>
225 fpType
226 fixDest(bool flush, bool defaultNan, fpType val, fpType op1)
227 {
228  int fpClass = std::fpclassify(val);
229  fpType junk = 0.0;
230  if (fpClass == FP_NAN) {
231  const bool single = (sizeof(val) == sizeof(float));
232  const uint64_t qnan = single ? 0x7fc00000 : ULL(0x7ff8000000000000);
233  const bool nan = std::isnan(op1);
234  if (!nan || defaultNan) {
235  val = bitsToFp(qnan, junk);
236  } else if (nan) {
237  val = bitsToFp(fpToBits(op1) | qnan, junk);
238  }
239  } else if (fpClass == FP_SUBNORMAL && flush == 1) {
240  // Turn val into a zero with the correct sign;
241  uint64_t bitMask = ULL(0x1) << (sizeof(fpType) * 8 - 1);
242  val = bitsToFp(fpToBits(val) & bitMask, junk);
243  feclearexcept(FeInexact);
244  feraiseexcept(FeUnderflow);
245  }
246  return val;
247 }
248 
249 template
250 float fixDest<float>(bool flush, bool defaultNan, float val, float op1);
251 template
252 double fixDest<double>(bool flush, bool defaultNan, double val, double op1);
253 
254 template <class fpType>
255 fpType
256 fixDest(bool flush, bool defaultNan, fpType val, fpType op1, fpType op2)
257 {
258  int fpClass = std::fpclassify(val);
259  fpType junk = 0.0;
260  if (fpClass == FP_NAN) {
261  const bool single = (sizeof(val) == sizeof(float));
262  const uint64_t qnan = single ? 0x7fc00000 : ULL(0x7ff8000000000000);
263  const bool nan1 = std::isnan(op1);
264  const bool nan2 = std::isnan(op2);
265  const bool signal1 = nan1 && ((fpToBits(op1) & qnan) != qnan);
266  const bool signal2 = nan2 && ((fpToBits(op2) & qnan) != qnan);
267  if ((!nan1 && !nan2) || defaultNan) {
268  val = bitsToFp(qnan, junk);
269  } else if (signal1) {
270  val = bitsToFp(fpToBits(op1) | qnan, junk);
271  } else if (signal2) {
272  val = bitsToFp(fpToBits(op2) | qnan, junk);
273  } else if (nan1) {
274  val = op1;
275  } else if (nan2) {
276  val = op2;
277  }
278  } else if (fpClass == FP_SUBNORMAL && flush) {
279  // Turn val into a zero with the correct sign;
280  uint64_t bitMask = ULL(0x1) << (sizeof(fpType) * 8 - 1);
281  val = bitsToFp(fpToBits(val) & bitMask, junk);
282  feclearexcept(FeInexact);
283  feraiseexcept(FeUnderflow);
284  }
285  return val;
286 }
287 
288 template
289 float fixDest<float>(bool flush, bool defaultNan,
290  float val, float op1, float op2);
291 template
292 double fixDest<double>(bool flush, bool defaultNan,
293  double val, double op1, double op2);
294 
295 template <class fpType>
296 fpType
297 fixDivDest(bool flush, bool defaultNan, fpType val, fpType op1, fpType op2)
298 {
299  fpType mid = fixDest(flush, defaultNan, val, op1, op2);
300  const bool single = (sizeof(fpType) == sizeof(float));
301  const fpType junk = 0.0;
302  if ((single && (val == bitsToFp(0x00800000, junk) ||
303  val == bitsToFp(0x80800000, junk))) ||
304  (!single && (val == bitsToFp(ULL(0x0010000000000000), junk) ||
305  val == bitsToFp(ULL(0x8010000000000000), junk)))
306  ) {
307  __asm__ __volatile__("" : "=m" (op1) : "m" (op1));
308  fesetround(FeRoundZero);
309  fpType temp = 0.0;
310  __asm__ __volatile__("" : "=m" (temp) : "m" (temp));
311  temp = op1 / op2;
312  if (flushToZero(temp)) {
313  feraiseexcept(FeUnderflow);
314  if (flush) {
315  feclearexcept(FeInexact);
316  mid = temp;
317  }
318  }
319  __asm__ __volatile__("" :: "m" (temp));
320  }
321  return mid;
322 }
323 
324 template
325 float fixDivDest<float>(bool flush, bool defaultNan,
326  float val, float op1, float op2);
327 template
328 double fixDivDest<double>(bool flush, bool defaultNan,
329  double val, double op1, double op2);
330 
331 float
332 fixFpDFpSDest(FPSCR fpscr, double val)
333 {
334  const float junk = 0.0;
335  float op1 = 0.0;
336  if (std::isnan(val)) {
337  uint64_t valBits = fpToBits(val);
338  uint32_t op1Bits = bits(valBits, 50, 29) |
339  (mask(9) << 22) |
340  (bits(valBits, 63) << 31);
341  op1 = bitsToFp(op1Bits, junk);
342  }
343  float mid = fixDest(fpscr.fz, fpscr.dn, (float)val, op1);
344  if (fpscr.fz && fetestexcept(FeUnderflow | FeInexact) ==
345  (FeUnderflow | FeInexact)) {
346  feclearexcept(FeInexact);
347  }
348  if (mid == bitsToFp(0x00800000, junk) ||
349  mid == bitsToFp(0x80800000, junk)) {
350  __asm__ __volatile__("" : "=m" (val) : "m" (val));
351  fesetround(FeRoundZero);
352  float temp = 0.0;
353  __asm__ __volatile__("" : "=m" (temp) : "m" (temp));
354  temp = val;
355  if (flushToZero(temp)) {
356  feraiseexcept(FeUnderflow);
357  if (fpscr.fz) {
358  feclearexcept(FeInexact);
359  mid = temp;
360  }
361  }
362  __asm__ __volatile__("" :: "m" (temp));
363  }
364  return mid;
365 }
366 
367 double
368 fixFpSFpDDest(FPSCR fpscr, float val)
369 {
370  const double junk = 0.0;
371  double op1 = 0.0;
372  if (std::isnan(val)) {
373  uint32_t valBits = fpToBits(val);
374  uint64_t op1Bits = ((uint64_t)bits(valBits, 21, 0) << 29) |
375  (mask(12) << 51) |
376  ((uint64_t)bits(valBits, 31) << 63);
377  op1 = bitsToFp(op1Bits, junk);
378  }
379  double mid = fixDest(fpscr.fz, fpscr.dn, (double)val, op1);
380  if (mid == bitsToFp(ULL(0x0010000000000000), junk) ||
381  mid == bitsToFp(ULL(0x8010000000000000), junk)) {
382  __asm__ __volatile__("" : "=m" (val) : "m" (val));
383  fesetround(FeRoundZero);
384  double temp = 0.0;
385  __asm__ __volatile__("" : "=m" (temp) : "m" (temp));
386  temp = val;
387  if (flushToZero(temp)) {
388  feraiseexcept(FeUnderflow);
389  if (fpscr.fz) {
390  feclearexcept(FeInexact);
391  mid = temp;
392  }
393  }
394  __asm__ __volatile__("" :: "m" (temp));
395  }
396  return mid;
397 }
398 
399 static inline uint16_t
400 vcvtFpFpH(FPSCR &fpscr, bool flush, bool defaultNan,
401  uint32_t rMode, bool ahp, uint64_t opBits, bool isDouble)
402 {
403  uint32_t mWidth;
404  uint32_t eWidth;
405  uint32_t eHalfRange;
406  uint32_t sBitPos;
407 
408  if (isDouble) {
409  mWidth = 52;
410  eWidth = 11;
411  } else {
412  mWidth = 23;
413  eWidth = 8;
414  }
415  sBitPos = eWidth + mWidth;
416  eHalfRange = (1 << (eWidth-1)) - 1;
417 
418  // Extract the operand.
419  bool neg = bits(opBits, sBitPos);
420  uint32_t exponent = bits(opBits, sBitPos-1, mWidth);
421  uint64_t oldMantissa = bits(opBits, mWidth-1, 0);
422  uint32_t mantissa = oldMantissa >> (mWidth - 10);
423  // Do the conversion.
424  uint64_t extra = oldMantissa & mask(mWidth - 10);
425  if (exponent == mask(eWidth)) {
426  if (oldMantissa != 0) {
427  // Nans.
428  if (bits(mantissa, 9) == 0) {
429  // Signalling nan.
430  fpscr.ioc = 1;
431  }
432  if (ahp) {
433  mantissa = 0;
434  exponent = 0;
435  fpscr.ioc = 1;
436  } else if (defaultNan) {
437  mantissa = (1 << 9);
438  exponent = 0x1f;
439  neg = false;
440  } else {
441  exponent = 0x1f;
442  mantissa |= (1 << 9);
443  }
444  } else {
445  // Infinities.
446  exponent = 0x1F;
447  if (ahp) {
448  fpscr.ioc = 1;
449  mantissa = 0x3ff;
450  } else {
451  mantissa = 0;
452  }
453  }
454  } else if (exponent == 0 && oldMantissa == 0) {
455  // Zero, don't need to do anything.
456  } else {
457  // Normalized or denormalized numbers.
458 
459  bool inexact = (extra != 0);
460 
461  if (exponent == 0) {
462  // Denormalized.
463  // If flush to zero is on, this shouldn't happen.
464  assert(!flush);
465 
466  // Check for underflow
467  if (inexact || fpscr.ufe)
468  fpscr.ufc = 1;
469 
470  // Handle rounding.
471  unsigned mode = rMode;
472  if ((mode == VfpRoundUpward && !neg && extra) ||
473  (mode == VfpRoundDown && neg && extra) ||
474  (mode == VfpRoundNearest &&
475  (extra > (1 << 9) ||
476  (extra == (1 << 9) && bits(mantissa, 0))))) {
477  mantissa++;
478  }
479 
480  // See if the number became normalized after rounding.
481  if (mantissa == (1 << 10)) {
482  mantissa = 0;
483  exponent = 1;
484  }
485  } else {
486  // Normalized.
487 
488  // We need to track the dropped bits differently since
489  // more can be dropped by denormalizing.
490  bool topOne = bits(extra, mWidth - 10 - 1);
491  bool restZeros = bits(extra, mWidth - 10 - 2, 0) == 0;
492 
493  if (exponent <= (eHalfRange - 15)) {
494  // The result is too small. Denormalize.
495  mantissa |= (1 << 10);
496  while (mantissa && exponent <= (eHalfRange - 15)) {
497  restZeros = restZeros && !topOne;
498  topOne = bits(mantissa, 0);
499  mantissa = mantissa >> 1;
500  exponent++;
501  }
502  if (topOne || !restZeros)
503  inexact = true;
504  exponent = 0;
505  } else {
506  // Change bias.
507  exponent -= (eHalfRange - 15);
508  }
509 
510  if (exponent == 0 && (inexact || fpscr.ufe)) {
511  // Underflow
512  fpscr.ufc = 1;
513  }
514 
515  // Handle rounding.
516  unsigned mode = rMode;
517  bool nonZero = topOne || !restZeros;
518  if ((mode == VfpRoundUpward && !neg && nonZero) ||
519  (mode == VfpRoundDown && neg && nonZero) ||
520  (mode == VfpRoundNearest && topOne &&
521  (!restZeros || bits(mantissa, 0)))) {
522  mantissa++;
523  }
524 
525  // See if we rounded up and need to bump the exponent.
526  if (mantissa == (1 << 10)) {
527  mantissa = 0;
528  exponent++;
529  }
530 
531  // Deal with overflow
532  if (ahp) {
533  if (exponent >= 0x20) {
534  exponent = 0x1f;
535  mantissa = 0x3ff;
536  fpscr.ioc = 1;
537  // Supress inexact exception.
538  inexact = false;
539  }
540  } else {
541  if (exponent >= 0x1f) {
542  if ((mode == VfpRoundNearest) ||
543  (mode == VfpRoundUpward && !neg) ||
544  (mode == VfpRoundDown && neg)) {
545  // Overflow to infinity.
546  exponent = 0x1f;
547  mantissa = 0;
548  } else {
549  // Overflow to max normal.
550  exponent = 0x1e;
551  mantissa = 0x3ff;
552  }
553  fpscr.ofc = 1;
554  inexact = true;
555  }
556  }
557  }
558 
559  if (inexact) {
560  fpscr.ixc = 1;
561  }
562  }
563  // Reassemble and install the result.
564  uint32_t result = bits(mantissa, 9, 0);
565  replaceBits(result, 14, 10, exponent);
566  if (neg)
567  result |= (1 << 15);
568  return result;
569 }
570 
571 uint16_t
572 vcvtFpSFpH(FPSCR &fpscr, bool flush, bool defaultNan,
573  uint32_t rMode, bool ahp, float op)
574 {
575  uint64_t opBits = fpToBits(op);
576  return vcvtFpFpH(fpscr, flush, defaultNan, rMode, ahp, opBits, false);
577 }
578 
579 uint16_t
580 vcvtFpDFpH(FPSCR &fpscr, bool flush, bool defaultNan,
581  uint32_t rMode, bool ahp, double op)
582 {
583  uint64_t opBits = fpToBits(op);
584  return vcvtFpFpH(fpscr, flush, defaultNan, rMode, ahp, opBits, true);
585 }
586 
587 static inline uint64_t
588 vcvtFpHFp(FPSCR &fpscr, bool defaultNan, bool ahp, uint16_t op, bool isDouble)
589 {
590  uint32_t mWidth;
591  uint32_t eWidth;
592  uint32_t eHalfRange;
593  uint32_t sBitPos;
594 
595  if (isDouble) {
596  mWidth = 52;
597  eWidth = 11;
598  } else {
599  mWidth = 23;
600  eWidth = 8;
601  }
602  sBitPos = eWidth + mWidth;
603  eHalfRange = (1 << (eWidth-1)) - 1;
604 
605  // Extract the bitfields.
606  bool neg = bits(op, 15);
607  uint32_t exponent = bits(op, 14, 10);
608  uint64_t mantissa = bits(op, 9, 0);
609  // Do the conversion.
610  if (exponent == 0) {
611  if (mantissa != 0) {
612  // Normalize the value.
613  exponent = exponent + (eHalfRange - 15) + 1;
614  while (mantissa < (1 << 10)) {
615  mantissa = mantissa << 1;
616  exponent--;
617  }
618  }
619  mantissa = mantissa << (mWidth - 10);
620  } else if (exponent == 0x1f && !ahp) {
621  // Infinities and nans.
622  exponent = mask(eWidth);
623  if (mantissa != 0) {
624  // Nans.
625  mantissa = mantissa << (mWidth - 10);
626  if (bits(mantissa, mWidth-1) == 0) {
627  // Signalling nan.
628  fpscr.ioc = 1;
629  mantissa |= (((uint64_t) 1) << (mWidth-1));
630  }
631  if (defaultNan) {
632  mantissa &= ~mask(mWidth-1);
633  neg = false;
634  }
635  }
636  } else {
637  exponent = exponent + (eHalfRange - 15);
638  mantissa = mantissa << (mWidth - 10);
639  }
640  // Reassemble the result.
641  uint64_t result = bits(mantissa, mWidth-1, 0);
642  replaceBits(result, sBitPos-1, mWidth, exponent);
643  if (neg) {
644  result |= (((uint64_t) 1) << sBitPos);
645  }
646  return result;
647 }
648 
649 double
650 vcvtFpHFpD(FPSCR &fpscr, bool defaultNan, bool ahp, uint16_t op)
651 {
652  double junk = 0.0;
653  uint64_t result;
654 
655  result = vcvtFpHFp(fpscr, defaultNan, ahp, op, true);
656  return bitsToFp(result, junk);
657 }
658 
659 float
660 vcvtFpHFpS(FPSCR &fpscr, bool defaultNan, bool ahp, uint16_t op)
661 {
662  float junk = 0.0;
663  uint64_t result;
664 
665  result = vcvtFpHFp(fpscr, defaultNan, ahp, op, false);
666  return bitsToFp(result, junk);
667 }
668 
669 float
670 vfpUFixedToFpS(bool flush, bool defaultNan,
671  uint64_t val, uint8_t width, uint8_t imm)
672 {
673  fesetround(FeRoundNearest);
674  if (width == 16)
675  val = (uint16_t)val;
676  else if (width == 32)
677  val = (uint32_t)val;
678  else if (width != 64)
679  panic("Unsupported width %d", width);
680  float scale = powf(2.0, imm);
681  __asm__ __volatile__("" : "=m" (scale) : "m" (scale));
682  feclearexcept(FeAllExceptions);
683  __asm__ __volatile__("" : "=m" (scale) : "m" (scale));
684  return fixDivDest(flush, defaultNan, val / scale, (float)val, scale);
685 }
686 
687 float
688 vfpSFixedToFpS(bool flush, bool defaultNan,
689  int64_t val, uint8_t width, uint8_t imm)
690 {
691  fesetround(FeRoundNearest);
692  if (width == 16)
693  val = sext<16>(val & mask(16));
694  else if (width == 32)
695  val = sext<32>(val & mask(32));
696  else if (width != 64)
697  panic("Unsupported width %d", width);
698 
699  float scale = powf(2.0, imm);
700  __asm__ __volatile__("" : "=m" (scale) : "m" (scale));
701  feclearexcept(FeAllExceptions);
702  __asm__ __volatile__("" : "=m" (scale) : "m" (scale));
703  return fixDivDest(flush, defaultNan, val / scale, (float)val, scale);
704 }
705 
706 
707 double
708 vfpUFixedToFpD(bool flush, bool defaultNan,
709  uint64_t val, uint8_t width, uint8_t imm)
710 {
711  fesetround(FeRoundNearest);
712  if (width == 16)
713  val = (uint16_t)val;
714  else if (width == 32)
715  val = (uint32_t)val;
716  else if (width != 64)
717  panic("Unsupported width %d", width);
718 
719  double scale = pow(2.0, imm);
720  __asm__ __volatile__("" : "=m" (scale) : "m" (scale));
721  feclearexcept(FeAllExceptions);
722  __asm__ __volatile__("" : "=m" (scale) : "m" (scale));
723  return fixDivDest(flush, defaultNan, val / scale, (double)val, scale);
724 }
725 
726 double
727 vfpSFixedToFpD(bool flush, bool defaultNan,
728  int64_t val, uint8_t width, uint8_t imm)
729 {
730  fesetround(FeRoundNearest);
731  if (width == 16)
732  val = sext<16>(val & mask(16));
733  else if (width == 32)
734  val = sext<32>(val & mask(32));
735  else if (width != 64)
736  panic("Unsupported width %d", width);
737 
738  double scale = pow(2.0, imm);
739  __asm__ __volatile__("" : "=m" (scale) : "m" (scale));
740  feclearexcept(FeAllExceptions);
741  __asm__ __volatile__("" : "=m" (scale) : "m" (scale));
742  return fixDivDest(flush, defaultNan, val / scale, (double)val, scale);
743 }
744 
745 // This function implements a magic formula taken from the architecture
746 // reference manual. It was originally called recip_sqrt_estimate.
747 static double
749 {
750  int64_t q0, q1, s;
751  double r;
752  if (a < 0.5) {
753  q0 = (int64_t)(a * 512.0);
754  r = 1.0 / sqrt(((double)q0 + 0.5) / 512.0);
755  } else {
756  q1 = (int64_t)(a * 256.0);
757  r = 1.0 / sqrt(((double)q1 + 0.5) / 256.0);
758  }
759  s = (int64_t)(256.0 * r + 0.5);
760  return (double)s / 256.0;
761 }
762 
763 // This function is only intended for use in Neon instructions because
764 // it ignores certain bits in the FPSCR.
765 float
766 fprSqrtEstimate(FPSCR &fpscr, float op)
767 {
768  const uint32_t qnan = 0x7fc00000;
769  float junk = 0.0;
770  int fpClass = std::fpclassify(op);
771  if (fpClass == FP_NAN) {
772  if ((fpToBits(op) & qnan) != qnan)
773  fpscr.ioc = 1;
774  return bitsToFp(qnan, junk);
775  } else if (fpClass == FP_ZERO) {
776  fpscr.dzc = 1;
777  // Return infinity with the same sign as the operand.
778  return bitsToFp((std::signbit(op) << 31) |
779  (0xFF << 23) | (0 << 0), junk);
780  } else if (std::signbit(op)) {
781  // Set invalid op bit.
782  fpscr.ioc = 1;
783  return bitsToFp(qnan, junk);
784  } else if (fpClass == FP_INFINITE) {
785  return 0.0;
786  } else {
787  uint64_t opBits = fpToBits(op);
788  double scaled;
789  if (bits(opBits, 23)) {
790  scaled = bitsToFp((0 << 0) | (bits(opBits, 22, 0) << 29) |
791  (ULL(0x3fd) << 52) | (bits(opBits, 31) << 63),
792  (double)0.0);
793  } else {
794  scaled = bitsToFp((0 << 0) | (bits(opBits, 22, 0) << 29) |
795  (ULL(0x3fe) << 52) | (bits(opBits, 31) << 63),
796  (double)0.0);
797  }
798  uint64_t resultExp = (380 - bits(opBits, 30, 23)) / 2;
799 
800  uint64_t estimate = fpToBits(recipSqrtEstimate(scaled));
801 
802  return bitsToFp((bits(estimate, 63) << 31) |
803  (bits(resultExp, 7, 0) << 23) |
804  (bits(estimate, 51, 29) << 0), junk);
805  }
806 }
807 
808 uint32_t
810 {
811  if (bits(op, 31, 30) == 0) {
812  return -1;
813  } else {
814  double dpOp;
815  if (bits(op, 31)) {
816  dpOp = bitsToFp((ULL(0) << 63) |
817  (ULL(0x3fe) << 52) |
818  (bits((uint64_t)op, 30, 0) << 21) |
819  (0 << 0), (double)0.0);
820  } else {
821  dpOp = bitsToFp((ULL(0) << 63) |
822  (ULL(0x3fd) << 52) |
823  (bits((uint64_t)op, 29, 0) << 22) |
824  (0 << 0), (double)0.0);
825  }
826  uint64_t estimate = fpToBits(recipSqrtEstimate(dpOp));
827  return (1 << 31) | bits(estimate, 51, 21);
828  }
829 }
830 
831 // This function implements a magic formula taken from the architecture
832 // reference manual. It was originally called recip_estimate.
833 
834 static double
836 {
837  int64_t q, s;
838  double r;
839  q = (int64_t)(a * 512.0);
840  r = 1.0 / (((double)q + 0.5) / 512.0);
841  s = (int64_t)(256.0 * r + 0.5);
842  return (double)s / 256.0;
843 }
844 
845 // This function is only intended for use in Neon instructions because
846 // it ignores certain bits in the FPSCR.
847 float
848 fpRecipEstimate(FPSCR &fpscr, float op)
849 {
850  const uint32_t qnan = 0x7fc00000;
851  float junk = 0.0;
852  int fpClass = std::fpclassify(op);
853  if (fpClass == FP_NAN) {
854  if ((fpToBits(op) & qnan) != qnan)
855  fpscr.ioc = 1;
856  return bitsToFp(qnan, junk);
857  } else if (fpClass == FP_INFINITE) {
858  return bitsToFp(std::signbit(op) << 31, junk);
859  } else if (fpClass == FP_ZERO) {
860  fpscr.dzc = 1;
861  // Return infinity with the same sign as the operand.
862  return bitsToFp((std::signbit(op) << 31) |
863  (0xFF << 23) | (0 << 0), junk);
864  } else if (fabs(op) >= pow(2.0, 126)) {
865  fpscr.ufc = 1;
866  return bitsToFp(std::signbit(op) << 31, junk);
867  } else {
868  uint64_t opBits = fpToBits(op);
869  double scaled;
870  scaled = bitsToFp((0 << 0) | (bits(opBits, 22, 0) << 29) |
871  (ULL(0x3fe) << 52) | (ULL(0) << 63),
872  (double)0.0);
873  uint64_t resultExp = 253 - bits(opBits, 30, 23);
874 
875  uint64_t estimate = fpToBits(recipEstimate(scaled));
876 
877  return bitsToFp((bits(opBits, 31) << 31) |
878  (bits(resultExp, 7, 0) << 23) |
879  (bits(estimate, 51, 29) << 0), junk);
880  }
881 }
882 
883 uint32_t
885 {
886  if (bits(op, 31) == 0) {
887  return -1;
888  } else {
889  double dpOp;
890  dpOp = bitsToFp((ULL(0) << 63) |
891  (ULL(0x3fe) << 52) |
892  (bits((uint64_t)op, 30, 0) << 21) |
893  (0 << 0), (double)0.0);
894  uint64_t estimate = fpToBits(recipEstimate(dpOp));
895  return (1 << 31) | bits(estimate, 51, 21);
896  }
897 }
898 
899 FPSCR
900 fpStandardFPSCRValue(const FPSCR &fpscr)
901 {
902  FPSCR new_fpscr(0);
903  new_fpscr.ahp = fpscr.ahp;
904  new_fpscr.dn = 1;
905  new_fpscr.fz = 1;
906  new_fpscr.fz16 = fpscr.fz16;
907  return new_fpscr;
908 };
909 
910 template <class fpType>
911 fpType
912 FpOp::processNans(FPSCR &fpscr, bool &done, bool defaultNan,
913  fpType op1, fpType op2) const
914 {
915  done = true;
916  fpType junk = 0.0;
917  fpType dest = 0.0;
918  const bool single = (sizeof(fpType) == sizeof(float));
919  const uint64_t qnan =
920  single ? 0x7fc00000 : ULL(0x7ff8000000000000);
921  const bool nan1 = std::isnan(op1);
922  const bool nan2 = std::isnan(op2);
923  const bool signal1 = nan1 && ((fpToBits(op1) & qnan) != qnan);
924  const bool signal2 = nan2 && ((fpToBits(op2) & qnan) != qnan);
925  if (nan1 || nan2) {
926  if (defaultNan) {
927  dest = bitsToFp(qnan, junk);
928  } else if (signal1) {
929  dest = bitsToFp(fpToBits(op1) | qnan, junk);
930  } else if (signal2) {
931  dest = bitsToFp(fpToBits(op2) | qnan, junk);
932  } else if (nan1) {
933  dest = op1;
934  } else if (nan2) {
935  dest = op2;
936  }
937  if (signal1 || signal2) {
938  fpscr.ioc = 1;
939  }
940  } else {
941  done = false;
942  }
943  return dest;
944 }
945 
946 template
947 float FpOp::processNans(FPSCR &fpscr, bool &done, bool defaultNan,
948  float op1, float op2) const;
949 template
950 double FpOp::processNans(FPSCR &fpscr, bool &done, bool defaultNan,
951  double op1, double op2) const;
952 
953 // @TODO remove this function when we've finished switching all FMA code to use the new FPLIB
954 template <class fpType>
955 fpType
956 FpOp::ternaryOp(FPSCR &fpscr, fpType op1, fpType op2, fpType op3,
957  fpType (*func)(fpType, fpType, fpType),
958  bool flush, bool defaultNan, uint32_t rMode) const
959 {
960  const bool single = (sizeof(fpType) == sizeof(float));
961  fpType junk = 0.0;
962 
963  if (flush && (flushToZero(op1, op2) || flushToZero(op3)))
964  fpscr.idc = 1;
965  VfpSavedState state = prepFpState(rMode);
966  __asm__ __volatile__ ("" : "=m" (op1), "=m" (op2), "=m" (op3), "=m" (state)
967  : "m" (op1), "m" (op2), "m" (op3), "m" (state));
968  fpType dest = func(op1, op2, op3);
969  __asm__ __volatile__ ("" : "=m" (dest) : "m" (dest));
970 
971  int fpClass = std::fpclassify(dest);
972  // Get NAN behavior right. This varies between x86 and ARM.
973  if (fpClass == FP_NAN) {
974  const uint64_t qnan =
975  single ? 0x7fc00000 : ULL(0x7ff8000000000000);
976  const bool nan1 = std::isnan(op1);
977  const bool nan2 = std::isnan(op2);
978  const bool nan3 = std::isnan(op3);
979  const bool signal1 = nan1 && ((fpToBits(op1) & qnan) != qnan);
980  const bool signal2 = nan2 && ((fpToBits(op2) & qnan) != qnan);
981  const bool signal3 = nan3 && ((fpToBits(op3) & qnan) != qnan);
982  if ((!nan1 && !nan2 && !nan3) || (defaultNan == 1)) {
983  dest = bitsToFp(qnan, junk);
984  } else if (signal1) {
985  dest = bitsToFp(fpToBits(op1) | qnan, junk);
986  } else if (signal2) {
987  dest = bitsToFp(fpToBits(op2) | qnan, junk);
988  } else if (signal3) {
989  dest = bitsToFp(fpToBits(op3) | qnan, junk);
990  } else if (nan1) {
991  dest = op1;
992  } else if (nan2) {
993  dest = op2;
994  } else if (nan3) {
995  dest = op3;
996  }
997  } else if (flush && flushToZero(dest)) {
998  feraiseexcept(FeUnderflow);
999  } else if ((
1000  (single && (dest == bitsToFp(0x00800000, junk) ||
1001  dest == bitsToFp(0x80800000, junk))) ||
1002  (!single &&
1003  (dest == bitsToFp(ULL(0x0010000000000000), junk) ||
1004  dest == bitsToFp(ULL(0x8010000000000000), junk)))
1005  ) && rMode != VfpRoundZero) {
1006  /*
1007  * Correct for the fact that underflow is detected -before- rounding
1008  * in ARM and -after- rounding in x86.
1009  */
1010  fesetround(FeRoundZero);
1011  __asm__ __volatile__ ("" : "=m" (op1), "=m" (op2), "=m" (op3)
1012  : "m" (op1), "m" (op2), "m" (op3));
1013  fpType temp = func(op1, op2, op2);
1014  __asm__ __volatile__ ("" : "=m" (temp) : "m" (temp));
1015  if (flush && flushToZero(temp)) {
1016  dest = temp;
1017  }
1018  }
1019  finishVfp(fpscr, state, flush);
1020  return dest;
1021 }
1022 
1023 template
1024 float FpOp::ternaryOp(FPSCR &fpscr, float op1, float op2, float op3,
1025  float (*func)(float, float, float),
1026  bool flush, bool defaultNan, uint32_t rMode) const;
1027 template
1028 double FpOp::ternaryOp(FPSCR &fpscr, double op1, double op2, double op3,
1029  double (*func)(double, double, double),
1030  bool flush, bool defaultNan, uint32_t rMode) const;
1031 
1032 template <class fpType>
1033 fpType
1034 FpOp::binaryOp(FPSCR &fpscr, fpType op1, fpType op2,
1035  fpType (*func)(fpType, fpType),
1036  bool flush, bool defaultNan, uint32_t rMode) const
1037 {
1038  const bool single = (sizeof(fpType) == sizeof(float));
1039  fpType junk = 0.0;
1040 
1041  if (flush && flushToZero(op1, op2))
1042  fpscr.idc = 1;
1043  VfpSavedState state = prepFpState(rMode);
1044  __asm__ __volatile__ ("" : "=m" (op1), "=m" (op2), "=m" (state)
1045  : "m" (op1), "m" (op2), "m" (state));
1046  fpType dest = func(op1, op2);
1047  __asm__ __volatile__ ("" : "=m" (dest) : "m" (dest));
1048 
1049  // Get NAN behavior right. This varies between x86 and ARM.
1050  if (std::isnan(dest)) {
1051  const uint64_t qnan =
1052  single ? 0x7fc00000 : ULL(0x7ff8000000000000);
1053  const bool nan1 = std::isnan(op1);
1054  const bool nan2 = std::isnan(op2);
1055  const bool signal1 = nan1 && ((fpToBits(op1) & qnan) != qnan);
1056  const bool signal2 = nan2 && ((fpToBits(op2) & qnan) != qnan);
1057  if ((!nan1 && !nan2) || (defaultNan == 1)) {
1058  dest = bitsToFp(qnan, junk);
1059  } else if (signal1) {
1060  dest = bitsToFp(fpToBits(op1) | qnan, junk);
1061  } else if (signal2) {
1062  dest = bitsToFp(fpToBits(op2) | qnan, junk);
1063  } else if (nan1) {
1064  dest = op1;
1065  } else if (nan2) {
1066  dest = op2;
1067  }
1068  } else if (flush && flushToZero(dest)) {
1069  feraiseexcept(FeUnderflow);
1070  } else if ((
1071  (single && (dest == bitsToFp(0x00800000, junk) ||
1072  dest == bitsToFp(0x80800000, junk))) ||
1073  (!single &&
1074  (dest == bitsToFp(ULL(0x0010000000000000), junk) ||
1075  dest == bitsToFp(ULL(0x8010000000000000), junk)))
1076  ) && rMode != VfpRoundZero) {
1077  /*
1078  * Correct for the fact that underflow is detected -before- rounding
1079  * in ARM and -after- rounding in x86.
1080  */
1081  fesetround(FeRoundZero);
1082  __asm__ __volatile__ ("" : "=m" (op1), "=m" (op2)
1083  : "m" (op1), "m" (op2));
1084  fpType temp = func(op1, op2);
1085  __asm__ __volatile__ ("" : "=m" (temp) : "m" (temp));
1086  if (flush && flushToZero(temp)) {
1087  dest = temp;
1088  }
1089  }
1090  finishVfp(fpscr, state, flush);
1091  return dest;
1092 }
1093 
1094 template
1095 float FpOp::binaryOp(FPSCR &fpscr, float op1, float op2,
1096  float (*func)(float, float),
1097  bool flush, bool defaultNan, uint32_t rMode) const;
1098 template
1099 double FpOp::binaryOp(FPSCR &fpscr, double op1, double op2,
1100  double (*func)(double, double),
1101  bool flush, bool defaultNan, uint32_t rMode) const;
1102 
1103 template <class fpType>
1104 fpType
1105 FpOp::unaryOp(FPSCR &fpscr, fpType op1, fpType (*func)(fpType),
1106  bool flush, uint32_t rMode) const
1107 {
1108  const bool single = (sizeof(fpType) == sizeof(float));
1109  fpType junk = 0.0;
1110 
1111  if (flush && flushToZero(op1))
1112  fpscr.idc = 1;
1113  VfpSavedState state = prepFpState(rMode);
1114  __asm__ __volatile__ ("" : "=m" (op1), "=m" (state)
1115  : "m" (op1), "m" (state));
1116  fpType dest = func(op1);
1117  __asm__ __volatile__ ("" : "=m" (dest) : "m" (dest));
1118 
1119  // Get NAN behavior right. This varies between x86 and ARM.
1120  if (std::isnan(dest)) {
1121  const uint64_t qnan =
1122  single ? 0x7fc00000 : ULL(0x7ff8000000000000);
1123  const bool nan = std::isnan(op1);
1124  if (!nan || fpscr.dn == 1) {
1125  dest = bitsToFp(qnan, junk);
1126  } else if (nan) {
1127  dest = bitsToFp(fpToBits(op1) | qnan, junk);
1128  }
1129  } else if (flush && flushToZero(dest)) {
1130  feraiseexcept(FeUnderflow);
1131  } else if ((
1132  (single && (dest == bitsToFp(0x00800000, junk) ||
1133  dest == bitsToFp(0x80800000, junk))) ||
1134  (!single &&
1135  (dest == bitsToFp(ULL(0x0010000000000000), junk) ||
1136  dest == bitsToFp(ULL(0x8010000000000000), junk)))
1137  ) && rMode != VfpRoundZero) {
1138  /*
1139  * Correct for the fact that underflow is detected -before- rounding
1140  * in ARM and -after- rounding in x86.
1141  */
1142  fesetround(FeRoundZero);
1143  __asm__ __volatile__ ("" : "=m" (op1) : "m" (op1));
1144  fpType temp = func(op1);
1145  __asm__ __volatile__ ("" : "=m" (temp) : "m" (temp));
1146  if (flush && flushToZero(temp)) {
1147  dest = temp;
1148  }
1149  }
1150  finishVfp(fpscr, state, flush);
1151  return dest;
1152 }
1153 
1154 template
1155 float FpOp::unaryOp(FPSCR &fpscr, float op1, float (*func)(float),
1156  bool flush, uint32_t rMode) const;
1157 template
1158 double FpOp::unaryOp(FPSCR &fpscr, double op1, double (*func)(double),
1159  bool flush, uint32_t rMode) const;
1160 
1163 {
1164  if (wide) {
1165  stride *= 2;
1166  }
1167  unsigned offset = idx % 8;
1168  idx = (IntRegIndex)(idx - offset);
1169  offset += stride;
1170  idx = (IntRegIndex)(idx + (offset % 8));
1171  return idx;
1172 }
1173 
1174 void
1176 {
1177  unsigned stride = (machInst.fpscrStride == 0) ? 1 : 2;
1178  assert(!inScalarBank(dest));
1179  dest = addStride(dest, stride);
1180  op1 = addStride(op1, stride);
1181  if (!inScalarBank(op2)) {
1182  op2 = addStride(op2, stride);
1183  }
1184 }
1185 
1186 void
1188 {
1189  unsigned stride = (machInst.fpscrStride == 0) ? 1 : 2;
1190  assert(!inScalarBank(dest));
1191  dest = addStride(dest, stride);
1192  if (!inScalarBank(op1)) {
1193  op1 = addStride(op1, stride);
1194  }
1195 }
1196 
1197 void
1199 {
1200  unsigned stride = (machInst.fpscrStride == 0) ? 1 : 2;
1201  assert(!inScalarBank(dest));
1202  dest = addStride(dest, stride);
1203 }
1204 
1205 }
#define panic(...)
This implements a cprintf based panic() function.
Definition: logging.hh:163
void ccprintf(cp::Print &print)
Definition: cprintf.hh:127
template double fixDest< double >(bool flush, bool defaultNan, double val, double op1)
void printMnemonic(std::ostream &os, const std::string &suffix="", bool withPred=true, bool withCond64=false, ConditionCode cond64=COND_UC) const
Definition: static_inst.cc:372
double fixFpSFpDDest(FPSCR fpscr, float val)
Definition: vfp.cc:368
static float bitsToFp(uint64_t, float)
Definition: vfp.hh:178
fpType fixDivDest(bool flush, bool defaultNan, fpType val, fpType op1, fpType op2)
Definition: vfp.cc:297
std::string generateDisassembly(Addr pc, const Loader::SymbolTable *symtab) const override
Internal function to generate disassembly string.
Definition: vfp.cc:160
IntRegIndex
Definition: intregs.hh:51
uint32_t unsignedRecipEstimate(uint32_t op)
Definition: vfp.cc:884
Bitfield< 21, 20 > stride
Bitfield< 8 > a
float fixFpDFpSDest(FPSCR fpscr, double val)
Definition: vfp.cc:332
bool fpclassify(T src0, uint32_t src1)
Definition: decl.hh:537
template float fixDivDest< float >(bool flush, bool defaultNan, float val, float op1, float op2)
Bitfield< 23, 0 > offset
Definition: types.hh:152
Definition: ccregs.hh:41
static double recipSqrtEstimate(double a)
Definition: vfp.cc:748
template double fixDivDest< double >(bool flush, bool defaultNan, double val, double op1, double op2)
uint16_t vcvtFpDFpH(FPSCR &fpscr, bool flush, bool defaultNan, uint32_t rMode, bool ahp, double op)
Definition: vfp.cc:580
std::string generateDisassembly(Addr pc, const Loader::SymbolTable *symtab) const override
Internal function to generate disassembly string.
Definition: vfp.cc:62
Bitfield< 4, 0 > mode
double vcvtFpHFpD(FPSCR &fpscr, bool defaultNan, bool ahp, uint16_t op)
Definition: vfp.cc:650
uint16_t vcvtFpSFpH(FPSCR &fpscr, bool flush, bool defaultNan, uint32_t rMode, bool ahp, float op)
Definition: vfp.cc:572
FPSCR fpStandardFPSCRValue(const FPSCR &fpscr)
Definition: vfp.cc:900
void neg(sc_fxval &c, const sc_fxnum &a)
Definition: sc_fxnum.hh:2270
Bitfield< 63 > val
Definition: misc.hh:769
uint32_t unsignedRSqrtEstimate(uint32_t op)
Definition: vfp.cc:809
float fprSqrtEstimate(FPSCR &fpscr, float op)
Definition: vfp.cc:766
Bitfield< 26 > ahp
const ExtMachInst machInst
The binary machine instruction.
Definition: static_inst.hh:231
Bitfield< 4 > pc
IntRegIndex op2
Definition: vfp.hh:883
static double recipEstimate(double a)
Definition: vfp.cc:835
static uint32_t fpToBits(float)
Definition: vfp.hh:154
std::string generateDisassembly(Addr pc, const Loader::SymbolTable *symtab) const override
Internal function to generate disassembly string.
Definition: vfp.cc:78
std::string generateDisassembly(Addr pc, const Loader::SymbolTable *symtab) const override
Internal function to generate disassembly string.
Definition: vfp.cc:101
Bitfield< 4 > s
static uint16_t vcvtFpFpH(FPSCR &fpscr, bool flush, bool defaultNan, uint32_t rMode, bool ahp, uint64_t opBits, bool isDouble)
Definition: vfp.cc:400
std::string generateDisassembly(Addr pc, const Loader::SymbolTable *symtab) const override
Internal function to generate disassembly string.
Definition: vfp.cc:90
void printCondition(std::ostream &os, unsigned code, bool noImplicit=false) const
Definition: static_inst.cc:411
void finishVfp(FPSCR &fpscr, VfpSavedState state, bool flush, FPSCR mask)
Definition: vfp.cc:200
Bitfield< 27 > q
void replaceBits(T &val, int first, int last, B bit_val)
A convenience function to replace bits first to last of val with bit_val in place.
Definition: bitfield.hh:156
void printIntReg(std::ostream &os, RegIndex reg_idx, uint8_t opWidth=0) const
Print a register name for disassembly given the unique dependence tag number (FP or int)...
Definition: static_inst.cc:294
fpType processNans(FPSCR &fpscr, bool &done, bool defaultNan, fpType op1, fpType op2) const
Definition: vfp.cc:912
std::string generateDisassembly(Addr pc, const Loader::SymbolTable *symtab) const override
Internal function to generate disassembly string.
Definition: vfp.cc:114
fpType binaryOp(FPSCR &fpscr, fpType op1, fpType op2, fpType(*func)(fpType, fpType), bool flush, bool defaultNan, uint32_t rMode) const
Definition: vfp.cc:1034
double vfpUFixedToFpD(bool flush, bool defaultNan, uint64_t val, uint8_t width, uint8_t imm)
Definition: vfp.cc:708
Bitfield< 21 > ss
uint64_t Addr
Address type This will probably be moved somewhere else in the near future.
Definition: types.hh:140
#define ULL(N)
uint64_t constant
Definition: types.hh:48
Bitfield< 7, 0 > imm
Definition: types.hh:140
fpType unaryOp(FPSCR &fpscr, fpType op1, fpType(*func)(fpType), bool flush, uint32_t rMode) const
Definition: vfp.cc:1105
double vfpSFixedToFpD(bool flush, bool defaultNan, int64_t val, uint8_t width, uint8_t imm)
Definition: vfp.cc:727
void nextIdxs(IntRegIndex &dest, IntRegIndex &op1, IntRegIndex &op2)
Definition: vfp.cc:1175
scale
Definition: types.hh:92
fpType fixDest(bool flush, bool defaultNan, fpType val, fpType op1)
Definition: vfp.cc:226
std::string generateDisassembly(Addr pc, const Loader::SymbolTable *symtab) const override
Internal function to generate disassembly string.
Definition: vfp.cc:144
IntRegIndex op1
Definition: vfp.hh:883
int signbit(uint64_t src, int sz)
Definition: hsail_code.hh:100
fpType ternaryOp(FPSCR &fpscr, fpType op1, fpType op2, fpType op3, fpType(*func)(fpType, fpType, fpType), bool flush, bool defaultNan, uint32_t rMode) const
Definition: vfp.cc:956
float fpRecipEstimate(FPSCR &fpscr, float op)
Definition: vfp.cc:848
VfpSavedState prepFpState(uint32_t rMode)
Definition: vfp.cc:178
Bitfield< 4 > width
float vcvtFpHFpS(FPSCR &fpscr, bool defaultNan, bool ahp, uint16_t op)
Definition: vfp.cc:660
ConditionCode condCode
Definition: vfp.hh:884
void printFloatReg(std::ostream &os, RegIndex reg_idx) const
Definition: static_inst.cc:340
Bitfield< 3, 0 > mask
Definition: types.hh:62
template float fixDest< float >(bool flush, bool defaultNan, float val, float op1)
Bitfield< 4 > op
Definition: types.hh:78
float vfpUFixedToFpS(bool flush, bool defaultNan, uint64_t val, uint8_t width, uint8_t imm)
Definition: vfp.cc:670
T bits(T val, int first, int last)
Extract the bitfield from position &#39;first&#39; to &#39;last&#39; (inclusive) from &#39;val&#39; and right justify it...
Definition: bitfield.hh:71
static uint64_t vcvtFpHFp(FPSCR &fpscr, bool defaultNan, bool ahp, uint16_t op, bool isDouble)
Definition: vfp.cc:588
std::string generateDisassembly(Addr pc, const Loader::SymbolTable *symtab) const override
Internal function to generate disassembly string.
Definition: vfp.cc:128
Bitfield< 23, 22 > rMode
cond
Definition: types.hh:61
float vfpSFixedToFpS(bool flush, bool defaultNan, int64_t val, uint8_t width, uint8_t imm)
Definition: vfp.cc:688
IntRegIndex addStride(IntRegIndex idx, unsigned stride)
Definition: vfp.cc:1162
int VfpSavedState
Definition: vfp.hh:211

Generated on Thu May 28 2020 16:11:02 for gem5 by doxygen 1.8.13