237
321
return op & ~mask;
240
uint64_t helper_mskbl(uint64_t val, uint64_t mask)
242
return byte_zap(val, 0x01 << (mask & 7));
245
uint64_t helper_insbl(uint64_t val, uint64_t mask)
247
val <<= (mask & 7) * 8;
248
return byte_zap(val, ~(0x01 << (mask & 7)));
251
uint64_t helper_mskwl(uint64_t val, uint64_t mask)
253
return byte_zap(val, 0x03 << (mask & 7));
256
uint64_t helper_inswl(uint64_t val, uint64_t mask)
258
val <<= (mask & 7) * 8;
259
return byte_zap(val, ~(0x03 << (mask & 7)));
262
uint64_t helper_mskll(uint64_t val, uint64_t mask)
264
return byte_zap(val, 0x0F << (mask & 7));
267
uint64_t helper_insll(uint64_t val, uint64_t mask)
269
val <<= (mask & 7) * 8;
270
return byte_zap(val, ~(0x0F << (mask & 7)));
273
uint64_t helper_zap(uint64_t val, uint64_t mask)
275
return byte_zap(val, mask);
278
uint64_t helper_zapnot(uint64_t val, uint64_t mask)
280
return byte_zap(val, ~mask);
283
uint64_t helper_mskql(uint64_t val, uint64_t mask)
285
return byte_zap(val, 0xFF << (mask & 7));
288
uint64_t helper_insql(uint64_t val, uint64_t mask)
290
val <<= (mask & 7) * 8;
291
return byte_zap(val, ~(0xFF << (mask & 7)));
294
uint64_t helper_mskwh(uint64_t val, uint64_t mask)
296
return byte_zap(val, (0x03 << (mask & 7)) >> 8);
299
uint64_t helper_inswh(uint64_t val, uint64_t mask)
301
val >>= 64 - ((mask & 7) * 8);
302
return byte_zap(val, ~((0x03 << (mask & 7)) >> 8));
305
uint64_t helper_msklh(uint64_t val, uint64_t mask)
307
return byte_zap(val, (0x0F << (mask & 7)) >> 8);
310
uint64_t helper_inslh(uint64_t val, uint64_t mask)
312
val >>= 64 - ((mask & 7) * 8);
313
return byte_zap(val, ~((0x0F << (mask & 7)) >> 8));
316
uint64_t helper_mskqh(uint64_t val, uint64_t mask)
318
return byte_zap(val, (0xFF << (mask & 7)) >> 8);
321
uint64_t helper_insqh(uint64_t val, uint64_t mask)
323
val >>= 64 - ((mask & 7) * 8);
324
return byte_zap(val, ~((0xFF << (mask & 7)) >> 8));
327
uint64_t helper_cmpbge (uint64_t op1, uint64_t op2)
324
void helper_mskbl (void)
326
T0 = byte_zap(T0, 0x01 << (T1 & 7));
329
void helper_extbl (void)
332
T0 = byte_zap(T0, 0xFE);
335
void helper_insbl (void)
338
T0 = byte_zap(T0, ~(0x01 << (T1 & 7)));
341
void helper_mskwl (void)
343
T0 = byte_zap(T0, 0x03 << (T1 & 7));
346
void helper_extwl (void)
349
T0 = byte_zap(T0, 0xFC);
352
void helper_inswl (void)
355
T0 = byte_zap(T0, ~(0x03 << (T1 & 7)));
358
void helper_mskll (void)
360
T0 = byte_zap(T0, 0x0F << (T1 & 7));
363
void helper_extll (void)
366
T0 = byte_zap(T0, 0xF0);
369
void helper_insll (void)
372
T0 = byte_zap(T0, ~(0x0F << (T1 & 7)));
375
void helper_zap (void)
377
T0 = byte_zap(T0, T1);
380
void helper_zapnot (void)
382
T0 = byte_zap(T0, ~T1);
385
void helper_mskql (void)
387
T0 = byte_zap(T0, 0xFF << (T1 & 7));
390
void helper_extql (void)
393
T0 = byte_zap(T0, 0x00);
396
void helper_insql (void)
399
T0 = byte_zap(T0, ~(0xFF << (T1 & 7)));
402
void helper_mskwh (void)
404
T0 = byte_zap(T0, (0x03 << (T1 & 7)) >> 8);
407
void helper_inswh (void)
409
T0 >>= 64 - ((T1 & 7) * 8);
410
T0 = byte_zap(T0, ~((0x03 << (T1 & 7)) >> 8));
413
void helper_extwh (void)
415
T0 <<= 64 - ((T1 & 7) * 8);
416
T0 = byte_zap(T0, ~0x07);
419
void helper_msklh (void)
421
T0 = byte_zap(T0, (0x0F << (T1 & 7)) >> 8);
424
void helper_inslh (void)
426
T0 >>= 64 - ((T1 & 7) * 8);
427
T0 = byte_zap(T0, ~((0x0F << (T1 & 7)) >> 8));
430
void helper_extlh (void)
432
T0 <<= 64 - ((T1 & 7) * 8);
433
T0 = byte_zap(T0, ~0x0F);
436
void helper_mskqh (void)
438
T0 = byte_zap(T0, (0xFF << (T1 & 7)) >> 8);
441
void helper_insqh (void)
443
T0 >>= 64 - ((T1 & 7) * 8);
444
T0 = byte_zap(T0, ~((0xFF << (T1 & 7)) >> 8));
447
void helper_extqh (void)
449
T0 <<= 64 - ((T1 & 7) * 8);
450
T0 = byte_zap(T0, 0x00);
453
void helper_cmpbge (void)
329
455
uint8_t opa, opb, res;
333
for (i = 0; i < 8; i++) {
334
opa = op1 >> (i * 8);
335
opb = op2 >> (i * 8);
459
for (i = 0; i < 7; i++) {
342
/* Floating point helpers */
344
/* F floating (VAX) */
345
static always_inline uint64_t float32_to_f (float32 fa)
348
uint64_t r, exp, mant, sig;
350
a = *(uint32_t*)(&fa);
351
sig = ((uint64_t)a & 0x80000000) << 32;
352
exp = (a >> 23) & 0xff;
353
mant = ((uint64_t)a & 0x007fffff) << 29;
468
void helper_cmov_fir (int freg)
471
env->fir[freg] = FT1;
474
void helper_sqrts (void)
476
FT0 = float32_sqrt(FT0, &FP_STATUS);
479
void helper_cpys (void)
488
r.i = p.i & 0x8000000000000000ULL;
489
r.i |= q.i & ~0x8000000000000000ULL;
493
void helper_cpysn (void)
502
r.i = (~p.i) & 0x8000000000000000ULL;
503
r.i |= q.i & ~0x8000000000000000ULL;
507
void helper_cpyse (void)
516
r.i = p.i & 0xFFF0000000000000ULL;
517
r.i |= q.i & ~0xFFF0000000000000ULL;
521
void helper_itofs (void)
529
FT0 = int64_to_float32(p.i, &FP_STATUS);
532
void helper_ftois (void)
539
p.i = float32_to_int64(FT0, &FP_STATUS);
543
void helper_sqrtt (void)
545
FT0 = float64_sqrt(FT0, &FP_STATUS);
548
void helper_cmptun (void)
556
if (float64_is_nan(FT0) || float64_is_nan(FT1))
557
p.i = 0x4000000000000000ULL;
561
void helper_cmpteq (void)
569
if (float64_eq(FT0, FT1, &FP_STATUS))
570
p.i = 0x4000000000000000ULL;
574
void helper_cmptle (void)
582
if (float64_le(FT0, FT1, &FP_STATUS))
583
p.i = 0x4000000000000000ULL;
587
void helper_cmptlt (void)
595
if (float64_lt(FT0, FT1, &FP_STATUS))
596
p.i = 0x4000000000000000ULL;
600
void helper_itoft (void)
608
FT0 = int64_to_float64(p.i, &FP_STATUS);
611
void helper_ftoit (void)
618
p.i = float64_to_int64(FT0, &FP_STATUS);
622
static int vaxf_is_valid (float ff)
631
exp = (p.i >> 23) & 0xFF;
632
mant = p.i & 0x007FFFFF;
633
if (exp == 0 && ((p.i & 0x80000000) || mant != 0)) {
634
/* Reserved operands / Dirty zero */
641
static float vaxf_to_ieee32 (float ff)
650
exp = (p.i >> 23) & 0xFF;
661
static float ieee32_to_vaxf (float fi)
670
exp = (p.i >> 23) & 0xFF;
671
mant = p.i & 0x007FFFFF;
355
672
if (exp == 255) {
356
673
/* NaN or infinity */
357
r = 1; /* VAX dirty zero */
358
675
} else if (exp == 0) {
363
680
/* Denormalized */
364
r = sig | ((exp + 1) << 52) | mant;
367
684
if (exp >= 253) {
369
r = 1; /* VAX dirty zero */
371
r = sig | ((exp + 2) << 52);
378
static always_inline float32 f_to_float32 (uint64_t a)
380
uint32_t r, exp, mant_sig;
382
exp = ((a >> 55) & 0x80) | ((a >> 52) & 0x7f);
383
mant_sig = ((a >> 32) & 0x80000000) | ((a >> 29) & 0x007fffff);
385
if (unlikely(!exp && mant_sig)) {
695
void helper_addf (void)
699
if (!vaxf_is_valid(FT0) || !vaxf_is_valid(FT1)) {
702
ft0 = vaxf_to_ieee32(FT0);
703
ft1 = vaxf_to_ieee32(FT1);
704
ft2 = float32_add(ft0, ft1, &FP_STATUS);
705
FT0 = ieee32_to_vaxf(ft2);
708
void helper_subf (void)
712
if (!vaxf_is_valid(FT0) || !vaxf_is_valid(FT1)) {
715
ft0 = vaxf_to_ieee32(FT0);
716
ft1 = vaxf_to_ieee32(FT1);
717
ft2 = float32_sub(ft0, ft1, &FP_STATUS);
718
FT0 = ieee32_to_vaxf(ft2);
721
void helper_mulf (void)
725
if (!vaxf_is_valid(FT0) || !vaxf_is_valid(FT1)) {
728
ft0 = vaxf_to_ieee32(FT0);
729
ft1 = vaxf_to_ieee32(FT1);
730
ft2 = float32_mul(ft0, ft1, &FP_STATUS);
731
FT0 = ieee32_to_vaxf(ft2);
734
void helper_divf (void)
738
if (!vaxf_is_valid(FT0) || !vaxf_is_valid(FT1)) {
741
ft0 = vaxf_to_ieee32(FT0);
742
ft1 = vaxf_to_ieee32(FT1);
743
ft2 = float32_div(ft0, ft1, &FP_STATUS);
744
FT0 = ieee32_to_vaxf(ft2);
747
void helper_sqrtf (void)
751
if (!vaxf_is_valid(FT0) || !vaxf_is_valid(FT1)) {
754
ft0 = vaxf_to_ieee32(FT0);
755
ft1 = float32_sqrt(ft0, &FP_STATUS);
756
FT0 = ieee32_to_vaxf(ft1);
759
void helper_itoff (void)
764
static int vaxg_is_valid (double ff)
773
exp = (p.i >> 52) & 0x7FF;
774
mant = p.i & 0x000FFFFFFFFFFFFFULL;
775
if (exp == 0 && ((p.i & 0x8000000000000000ULL) || mant != 0)) {
386
776
/* Reserved operands / Dirty zero */
387
helper_excp(EXCP_OPCDEC, 0);
783
static double vaxg_to_ieee64 (double fg)
792
exp = (p.i >> 52) & 0x7FF;
394
r = ((exp - 2) << 23) | mant_sig;
397
return *(float32*)(&a);
400
uint32_t helper_f_to_memory (uint64_t a)
403
r = (a & 0x00001fffe0000000ull) >> 13;
404
r |= (a & 0x07ffe00000000000ull) >> 45;
405
r |= (a & 0xc000000000000000ull) >> 48;
409
uint64_t helper_memory_to_f (uint32_t a)
412
r = ((uint64_t)(a & 0x0000c000)) << 48;
413
r |= ((uint64_t)(a & 0x003fffff)) << 45;
414
r |= ((uint64_t)(a & 0xffff0000)) << 13;
415
if (!(a & 0x00004000))
420
uint64_t helper_addf (uint64_t a, uint64_t b)
424
fa = f_to_float32(a);
425
fb = f_to_float32(b);
426
fr = float32_add(fa, fb, &FP_STATUS);
427
return float32_to_f(fr);
430
uint64_t helper_subf (uint64_t a, uint64_t b)
434
fa = f_to_float32(a);
435
fb = f_to_float32(b);
436
fr = float32_sub(fa, fb, &FP_STATUS);
437
return float32_to_f(fr);
440
uint64_t helper_mulf (uint64_t a, uint64_t b)
444
fa = f_to_float32(a);
445
fb = f_to_float32(b);
446
fr = float32_mul(fa, fb, &FP_STATUS);
447
return float32_to_f(fr);
450
uint64_t helper_divf (uint64_t a, uint64_t b)
454
fa = f_to_float32(a);
455
fb = f_to_float32(b);
456
fr = float32_div(fa, fb, &FP_STATUS);
457
return float32_to_f(fr);
460
uint64_t helper_sqrtf (uint64_t t)
464
ft = f_to_float32(t);
465
fr = float32_sqrt(ft, &FP_STATUS);
466
return float32_to_f(fr);
470
/* G floating (VAX) */
471
static always_inline uint64_t float64_to_g (float64 fa)
473
uint64_t a, r, exp, mant, sig;
475
a = *(uint64_t*)(&fa);
476
sig = a & 0x8000000000000000ull;
477
exp = (a >> 52) & 0x7ff;
478
mant = a & 0x000fffffffffffffull;
803
static double ieee64_to_vaxg (double fi)
813
exp = (p.i >> 52) & 0x7FF;
814
mant = p.i & 0x000FFFFFFFFFFFFFULL;
481
816
/* NaN or infinity */
482
r = 1; /* VAX dirty zero */
817
p.i = 1; /* VAX dirty zero */
483
818
} else if (exp == 0) {
488
823
/* Denormalized */
489
r = sig | ((exp + 1) << 52) | mant;
492
827
if (exp >= 2045) {
494
r = 1; /* VAX dirty zero */
829
p.i = 1; /* VAX dirty zero */
496
r = sig | ((exp + 2) << 52);
503
static always_inline float64 g_to_float64 (uint64_t a)
505
uint64_t r, exp, mant_sig;
507
exp = (a >> 52) & 0x7ff;
508
mant_sig = a & 0x800fffffffffffffull;
510
if (!exp && mant_sig) {
511
/* Reserved operands / Dirty zero */
512
helper_excp(EXCP_OPCDEC, 0);
519
r = ((exp - 2) << 52) | mant_sig;
522
return *(float64*)(&a);
525
uint64_t helper_g_to_memory (uint64_t a)
528
r = (a & 0x000000000000ffffull) << 48;
529
r |= (a & 0x00000000ffff0000ull) << 16;
530
r |= (a & 0x0000ffff00000000ull) >> 16;
531
r |= (a & 0xffff000000000000ull) >> 48;
535
uint64_t helper_memory_to_g (uint64_t a)
538
r = (a & 0x000000000000ffffull) << 48;
539
r |= (a & 0x00000000ffff0000ull) << 16;
540
r |= (a & 0x0000ffff00000000ull) >> 16;
541
r |= (a & 0xffff000000000000ull) >> 48;
545
uint64_t helper_addg (uint64_t a, uint64_t b)
549
fa = g_to_float64(a);
550
fb = g_to_float64(b);
551
fr = float64_add(fa, fb, &FP_STATUS);
552
return float64_to_g(fr);
555
uint64_t helper_subg (uint64_t a, uint64_t b)
559
fa = g_to_float64(a);
560
fb = g_to_float64(b);
561
fr = float64_sub(fa, fb, &FP_STATUS);
562
return float64_to_g(fr);
565
uint64_t helper_mulg (uint64_t a, uint64_t b)
569
fa = g_to_float64(a);
570
fb = g_to_float64(b);
571
fr = float64_mul(fa, fb, &FP_STATUS);
572
return float64_to_g(fr);
575
uint64_t helper_divg (uint64_t a, uint64_t b)
579
fa = g_to_float64(a);
580
fb = g_to_float64(b);
581
fr = float64_div(fa, fb, &FP_STATUS);
582
return float64_to_g(fr);
585
uint64_t helper_sqrtg (uint64_t a)
589
fa = g_to_float64(a);
590
fr = float64_sqrt(fa, &FP_STATUS);
591
return float64_to_g(fr);
595
/* S floating (single) */
596
static always_inline uint64_t float32_to_s (float32 fa)
601
a = *(uint32_t*)(&fa);
603
r = (((uint64_t)(a & 0xc0000000)) << 32) | (((uint64_t)(a & 0x3fffffff)) << 29);
604
if (((a & 0x7f800000) != 0x7f800000) && (!(a & 0x40000000)))
609
static always_inline float32 s_to_float32 (uint64_t a)
611
uint32_t r = ((a >> 32) & 0xc0000000) | ((a >> 29) & 0x3fffffff);
612
return *(float32*)(&r);
615
uint32_t helper_s_to_memory (uint64_t a)
617
/* Memory format is the same as float32 */
618
float32 fa = s_to_float32(a);
619
return *(uint32_t*)(&fa);
622
uint64_t helper_memory_to_s (uint32_t a)
624
/* Memory format is the same as float32 */
625
return float32_to_s(*(float32*)(&a));
628
uint64_t helper_adds (uint64_t a, uint64_t b)
632
fa = s_to_float32(a);
633
fb = s_to_float32(b);
634
fr = float32_add(fa, fb, &FP_STATUS);
635
return float32_to_s(fr);
638
uint64_t helper_subs (uint64_t a, uint64_t b)
642
fa = s_to_float32(a);
643
fb = s_to_float32(b);
644
fr = float32_sub(fa, fb, &FP_STATUS);
645
return float32_to_s(fr);
648
uint64_t helper_muls (uint64_t a, uint64_t b)
652
fa = s_to_float32(a);
653
fb = s_to_float32(b);
654
fr = float32_mul(fa, fb, &FP_STATUS);
655
return float32_to_s(fr);
658
uint64_t helper_divs (uint64_t a, uint64_t b)
662
fa = s_to_float32(a);
663
fb = s_to_float32(b);
664
fr = float32_div(fa, fb, &FP_STATUS);
665
return float32_to_s(fr);
668
uint64_t helper_sqrts (uint64_t a)
672
fa = s_to_float32(a);
673
fr = float32_sqrt(fa, &FP_STATUS);
674
return float32_to_s(fr);
678
/* T floating (double) */
679
static always_inline float64 t_to_float64 (uint64_t a)
681
/* Memory format is the same as float64 */
682
return *(float64*)(&a);
685
static always_inline uint64_t float64_to_t (float64 fa)
687
/* Memory format is the same as float64 */
688
return *(uint64*)(&fa);
691
uint64_t helper_addt (uint64_t a, uint64_t b)
695
fa = t_to_float64(a);
696
fb = t_to_float64(b);
697
fr = float64_add(fa, fb, &FP_STATUS);
698
return float64_to_t(fr);
701
uint64_t helper_subt (uint64_t a, uint64_t b)
705
fa = t_to_float64(a);
706
fb = t_to_float64(b);
707
fr = float64_sub(fa, fb, &FP_STATUS);
708
return float64_to_t(fr);
711
uint64_t helper_mult (uint64_t a, uint64_t b)
715
fa = t_to_float64(a);
716
fb = t_to_float64(b);
717
fr = float64_mul(fa, fb, &FP_STATUS);
718
return float64_to_t(fr);
721
uint64_t helper_divt (uint64_t a, uint64_t b)
725
fa = t_to_float64(a);
726
fb = t_to_float64(b);
727
fr = float64_div(fa, fb, &FP_STATUS);
728
return float64_to_t(fr);
731
uint64_t helper_sqrtt (uint64_t a)
735
fa = t_to_float64(a);
736
fr = float64_sqrt(fa, &FP_STATUS);
737
return float64_to_t(fr);
742
uint64_t helper_cpys(uint64_t a, uint64_t b)
744
return (a & 0x8000000000000000ULL) | (b & ~0x8000000000000000ULL);
747
uint64_t helper_cpysn(uint64_t a, uint64_t b)
749
return ((~a) & 0x8000000000000000ULL) | (b & ~0x8000000000000000ULL);
752
uint64_t helper_cpyse(uint64_t a, uint64_t b)
754
return (a & 0xFFF0000000000000ULL) | (b & ~0xFFF0000000000000ULL);
759
uint64_t helper_cmptun (uint64_t a, uint64_t b)
763
fa = t_to_float64(a);
764
fb = t_to_float64(b);
766
if (float64_is_nan(fa) || float64_is_nan(fb))
767
return 0x4000000000000000ULL;
772
uint64_t helper_cmpteq(uint64_t a, uint64_t b)
776
fa = t_to_float64(a);
777
fb = t_to_float64(b);
779
if (float64_eq(fa, fb, &FP_STATUS))
780
return 0x4000000000000000ULL;
785
uint64_t helper_cmptle(uint64_t a, uint64_t b)
789
fa = t_to_float64(a);
790
fb = t_to_float64(b);
792
if (float64_le(fa, fb, &FP_STATUS))
793
return 0x4000000000000000ULL;
798
uint64_t helper_cmptlt(uint64_t a, uint64_t b)
802
fa = t_to_float64(a);
803
fb = t_to_float64(b);
805
if (float64_lt(fa, fb, &FP_STATUS))
806
return 0x4000000000000000ULL;
811
uint64_t helper_cmpgeq(uint64_t a, uint64_t b)
815
fa = g_to_float64(a);
816
fb = g_to_float64(b);
818
if (float64_eq(fa, fb, &FP_STATUS))
819
return 0x4000000000000000ULL;
824
uint64_t helper_cmpgle(uint64_t a, uint64_t b)
828
fa = g_to_float64(a);
829
fb = g_to_float64(b);
831
if (float64_le(fa, fb, &FP_STATUS))
832
return 0x4000000000000000ULL;
837
uint64_t helper_cmpglt(uint64_t a, uint64_t b)
841
fa = g_to_float64(a);
842
fb = g_to_float64(b);
844
if (float64_lt(fa, fb, &FP_STATUS))
845
return 0x4000000000000000ULL;
850
uint64_t helper_cmpfeq (uint64_t a)
852
return !(a & 0x7FFFFFFFFFFFFFFFULL);
855
uint64_t helper_cmpfne (uint64_t a)
857
return (a & 0x7FFFFFFFFFFFFFFFULL);
860
uint64_t helper_cmpflt (uint64_t a)
862
return (a & 0x8000000000000000ULL) && (a & 0x7FFFFFFFFFFFFFFFULL);
865
uint64_t helper_cmpfle (uint64_t a)
867
return (a & 0x8000000000000000ULL) || !(a & 0x7FFFFFFFFFFFFFFFULL);
870
uint64_t helper_cmpfgt (uint64_t a)
872
return !(a & 0x8000000000000000ULL) && (a & 0x7FFFFFFFFFFFFFFFULL);
875
uint64_t helper_cmpfge (uint64_t a)
877
return !(a & 0x8000000000000000ULL) || !(a & 0x7FFFFFFFFFFFFFFFULL);
881
/* Floating point format conversion */
882
uint64_t helper_cvtts (uint64_t a)
887
fa = t_to_float64(a);
888
fr = float64_to_float32(fa, &FP_STATUS);
889
return float32_to_s(fr);
892
uint64_t helper_cvtst (uint64_t a)
897
fa = s_to_float32(a);
898
fr = float32_to_float64(fa, &FP_STATUS);
899
return float64_to_t(fr);
902
uint64_t helper_cvtqs (uint64_t a)
904
float32 fr = int64_to_float32(a, &FP_STATUS);
905
return float32_to_s(fr);
908
uint64_t helper_cvttq (uint64_t a)
910
float64 fa = t_to_float64(a);
911
return float64_to_int64_round_to_zero(fa, &FP_STATUS);
914
uint64_t helper_cvtqt (uint64_t a)
916
float64 fr = int64_to_float64(a, &FP_STATUS);
917
return float64_to_t(fr);
920
uint64_t helper_cvtqf (uint64_t a)
922
float32 fr = int64_to_float32(a, &FP_STATUS);
923
return float32_to_f(fr);
926
uint64_t helper_cvtgf (uint64_t a)
931
fa = g_to_float64(a);
932
fr = float64_to_float32(fa, &FP_STATUS);
933
return float32_to_f(fr);
936
uint64_t helper_cvtgq (uint64_t a)
938
float64 fa = g_to_float64(a);
939
return float64_to_int64_round_to_zero(fa, &FP_STATUS);
942
uint64_t helper_cvtqg (uint64_t a)
945
fr = int64_to_float64(a, &FP_STATUS);
946
return float64_to_g(fr);
949
uint64_t helper_cvtlq (uint64_t a)
951
return (int64_t)((int32_t)((a >> 32) | ((a >> 29) & 0x3FFFFFFF)));
954
static always_inline uint64_t __helper_cvtql (uint64_t a, int s, int v)
958
r = ((uint64_t)(a & 0xC0000000)) << 32;
959
r |= ((uint64_t)(a & 0x7FFFFFFF)) << 29;
961
if (v && (int64_t)((int32_t)r) != (int64_t)r) {
838
void helper_addg (void)
840
double ft0, ft1, ft2;
842
if (!vaxg_is_valid(FT0) || !vaxg_is_valid(FT1)) {
845
ft0 = vaxg_to_ieee64(FT0);
846
ft1 = vaxg_to_ieee64(FT1);
847
ft2 = float64_add(ft0, ft1, &FP_STATUS);
848
FT0 = ieee64_to_vaxg(ft2);
851
void helper_subg (void)
853
double ft0, ft1, ft2;
855
if (!vaxg_is_valid(FT0) || !vaxg_is_valid(FT1)) {
858
ft0 = vaxg_to_ieee64(FT0);
859
ft1 = vaxg_to_ieee64(FT1);
860
ft2 = float64_sub(ft0, ft1, &FP_STATUS);
861
FT0 = ieee64_to_vaxg(ft2);
864
void helper_mulg (void)
866
double ft0, ft1, ft2;
868
if (!vaxg_is_valid(FT0) || !vaxg_is_valid(FT1)) {
871
ft0 = vaxg_to_ieee64(FT0);
872
ft1 = vaxg_to_ieee64(FT1);
873
ft2 = float64_mul(ft0, ft1, &FP_STATUS);
874
FT0 = ieee64_to_vaxg(ft2);
877
void helper_divg (void)
879
double ft0, ft1, ft2;
881
if (!vaxg_is_valid(FT0) || !vaxg_is_valid(FT1)) {
884
ft0 = vaxg_to_ieee64(FT0);
885
ft1 = vaxg_to_ieee64(FT1);
886
ft2 = float64_div(ft0, ft1, &FP_STATUS);
887
FT0 = ieee64_to_vaxg(ft2);
890
void helper_sqrtg (void)
894
if (!vaxg_is_valid(FT0) || !vaxg_is_valid(FT1)) {
897
ft0 = vaxg_to_ieee64(FT0);
898
ft1 = float64_sqrt(ft0, &FP_STATUS);
899
FT0 = ieee64_to_vaxg(ft1);
902
void helper_cmpgeq (void)
910
if (!vaxg_is_valid(FT0) || !vaxg_is_valid(FT1)) {
913
ft0 = vaxg_to_ieee64(FT0);
914
ft1 = vaxg_to_ieee64(FT1);
916
if (float64_eq(ft0, ft1, &FP_STATUS))
917
p.u = 0x4000000000000000ULL;
921
void helper_cmpglt (void)
929
if (!vaxg_is_valid(FT0) || !vaxg_is_valid(FT1)) {
932
ft0 = vaxg_to_ieee64(FT0);
933
ft1 = vaxg_to_ieee64(FT1);
935
if (float64_lt(ft0, ft1, &FP_STATUS))
936
p.u = 0x4000000000000000ULL;
940
void helper_cmpgle (void)
948
if (!vaxg_is_valid(FT0) || !vaxg_is_valid(FT1)) {
951
ft0 = vaxg_to_ieee64(FT0);
952
ft1 = vaxg_to_ieee64(FT1);
954
if (float64_le(ft0, ft1, &FP_STATUS))
955
p.u = 0x4000000000000000ULL;
959
void helper_cvtqs (void)
970
void helper_cvttq (void)
981
void helper_cvtqt (void)
992
void helper_cvtqf (void)
1000
FT0 = ieee32_to_vaxf(p.u);
1003
void helper_cvtgf (void)
1007
ft0 = vaxg_to_ieee64(FT0);
1008
FT0 = ieee32_to_vaxf(ft0);
1011
void helper_cvtgd (void)
1016
void helper_cvtgq (void)
1023
p.u = vaxg_to_ieee64(FT0);
1027
void helper_cvtqg (void)
1035
FT0 = ieee64_to_vaxg(p.u);
1038
void helper_cvtdg (void)
1043
void helper_cvtlq (void)
1051
q.u = (p.u >> 29) & 0x3FFFFFFF;
1053
q.u = (int64_t)((int32_t)q.u);
1057
static inline void __helper_cvtql (int s, int v)
1065
q.u = ((uint64_t)(p.u & 0xC0000000)) << 32;
1066
q.u |= ((uint64_t)(p.u & 0x7FFFFFFF)) << 29;
1068
if (v && (int64_t)((int32_t)p.u) != (int64_t)p.u) {
962
1069
helper_excp(EXCP_ARITH, EXCP_ARITH_OVERFLOW);
970
uint64_t helper_cvtql (uint64_t a)
972
return __helper_cvtql(a, 0, 0);
975
uint64_t helper_cvtqlv (uint64_t a)
977
return __helper_cvtql(a, 0, 1);
980
uint64_t helper_cvtqlsv (uint64_t a)
982
return __helper_cvtql(a, 1, 1);
985
/* PALcode support special instructions */
1076
void helper_cvtql (void)
1078
__helper_cvtql(0, 0);
1081
void helper_cvtqlv (void)
1083
__helper_cvtql(0, 1);
1086
void helper_cvtqlsv (void)
1088
__helper_cvtql(1, 1);
1091
void helper_cmpfeq (void)
1093
if (float64_eq(FT0, FT1, &FP_STATUS))
1099
void helper_cmpfne (void)
1101
if (float64_eq(FT0, FT1, &FP_STATUS))
1107
void helper_cmpflt (void)
1109
if (float64_lt(FT0, FT1, &FP_STATUS))
1115
void helper_cmpfle (void)
1117
if (float64_lt(FT0, FT1, &FP_STATUS))
1123
void helper_cmpfgt (void)
1125
if (float64_le(FT0, FT1, &FP_STATUS))
1131
void helper_cmpfge (void)
1133
if (float64_lt(FT0, FT1, &FP_STATUS))
986
1139
#if !defined (CONFIG_USER_ONLY)
987
void helper_hw_rei (void)
989
env->pc = env->ipr[IPR_EXC_ADDR] & ~3;
990
env->ipr[IPR_EXC_ADDR] = env->ipr[IPR_EXC_ADDR] & 1;
991
/* XXX: re-enable interrupts and memory mapping */
994
void helper_hw_ret (uint64_t a)
997
env->ipr[IPR_EXC_ADDR] = a & 1;
998
/* XXX: re-enable interrupts and memory mapping */
1001
uint64_t helper_mfpr (int iprn, uint64_t val)
1005
if (cpu_alpha_mfpr(env, iprn, &tmp) == 0)
1011
void helper_mtpr (int iprn, uint64_t val)
1013
cpu_alpha_mtpr(env, iprn, val, NULL);
1016
void helper_set_alt_mode (void)
1018
env->saved_mode = env->ps & 0xC;
1019
env->ps = (env->ps & ~0xC) | (env->ipr[IPR_ALT_MODE] & 0xC);
1022
void helper_restore_mode (void)
1024
env->ps = (env->ps & ~0xC) | env->saved_mode;
1140
void helper_mfpr (int iprn)
1144
if (cpu_alpha_mfpr(env, iprn, &val) == 0)
1148
void helper_mtpr (int iprn)
1150
cpu_alpha_mtpr(env, iprn, T0, NULL);
1029
1154
/*****************************************************************************/
1030
1155
/* Softmmu support */
1031
1156
#if !defined (CONFIG_USER_ONLY)
1158
#define GETPC() (__builtin_return_address(0))
1033
1160
/* XXX: the two following helpers are pure hacks.
1034
1161
* Hopefully, we emulate the PALcode, then we should never see
1035
1162
* HW_LD / HW_ST instructions.
1037
uint64_t helper_ld_virt_to_phys (uint64_t virtaddr)
1039
uint64_t tlb_addr, physaddr;
1043
mmu_idx = cpu_mmu_index(env);
1044
index = (virtaddr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1);
1046
tlb_addr = env->tlb_table[mmu_idx][index].addr_read;
1047
if ((virtaddr & TARGET_PAGE_MASK) ==
1048
(tlb_addr & (TARGET_PAGE_MASK | TLB_INVALID_MASK))) {
1049
physaddr = virtaddr + env->tlb_table[mmu_idx][index].addend;
1051
/* the page is not in the TLB : fill it */
1053
tlb_fill(virtaddr, 0, mmu_idx, retaddr);
1059
uint64_t helper_st_virt_to_phys (uint64_t virtaddr)
1061
uint64_t tlb_addr, physaddr;
1065
mmu_idx = cpu_mmu_index(env);
1066
index = (virtaddr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1);
1068
tlb_addr = env->tlb_table[mmu_idx][index].addr_write;
1069
if ((virtaddr & TARGET_PAGE_MASK) ==
1070
(tlb_addr & (TARGET_PAGE_MASK | TLB_INVALID_MASK))) {
1071
physaddr = virtaddr + env->tlb_table[mmu_idx][index].addend;
1073
/* the page is not in the TLB : fill it */
1075
tlb_fill(virtaddr, 1, mmu_idx, retaddr);
1081
void helper_ldl_raw(uint64_t t0, uint64_t t1)
1086
void helper_ldq_raw(uint64_t t0, uint64_t t1)
1091
void helper_ldl_l_raw(uint64_t t0, uint64_t t1)
1097
void helper_ldq_l_raw(uint64_t t0, uint64_t t1)
1103
void helper_ldl_kernel(uint64_t t0, uint64_t t1)
1108
void helper_ldq_kernel(uint64_t t0, uint64_t t1)
1113
void helper_ldl_data(uint64_t t0, uint64_t t1)
1118
void helper_ldq_data(uint64_t t0, uint64_t t1)
1123
void helper_stl_raw(uint64_t t0, uint64_t t1)
1128
void helper_stq_raw(uint64_t t0, uint64_t t1)
1133
uint64_t helper_stl_c_raw(uint64_t t0, uint64_t t1)
1137
if (t1 == env->lock) {
1148
uint64_t helper_stq_c_raw(uint64_t t0, uint64_t t1)
1152
if (t1 == env->lock) {
1164
void helper_ld_phys_to_virt (void)
1166
uint64_t tlb_addr, physaddr;
1170
is_user = (env->ps >> 3) & 3;
1171
index = (T0 >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1);
1173
tlb_addr = env->tlb_table[is_user][index].addr_read;
1174
if ((T0 & TARGET_PAGE_MASK) ==
1175
(tlb_addr & (TARGET_PAGE_MASK | TLB_INVALID_MASK))) {
1176
physaddr = T0 + env->tlb_table[is_user][index].addend;
1178
/* the page is not in the TLB : fill it */
1180
tlb_fill(T0, 0, is_user, retaddr);
1186
void helper_st_phys_to_virt (void)
1188
uint64_t tlb_addr, physaddr;
1192
is_user = (env->ps >> 3) & 3;
1193
index = (T0 >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1);
1195
tlb_addr = env->tlb_table[is_user][index].addr_write;
1196
if ((T0 & TARGET_PAGE_MASK) ==
1197
(tlb_addr & (TARGET_PAGE_MASK | TLB_INVALID_MASK))) {
1198
physaddr = T0 + env->tlb_table[is_user][index].addend;
1200
/* the page is not in the TLB : fill it */
1202
tlb_fill(T0, 1, is_user, retaddr);
1163
1208
#define MMUSUFFIX _mmu