PageRenderTime 360ms CodeModel.GetById 27ms RepoModel.GetById 1ms app.codeStats 0ms

/rpython/jit/backend/zarch/test/test_assembler.py

https://bitbucket.org/pypy/pypy/
Python | 638 lines | 538 code | 81 blank | 19 comment | 36 complexity | dc850b0a5166becce09867b878f56e93 MD5 | raw file
Possible License(s): AGPL-3.0, BSD-3-Clause, Apache-2.0
  1. import py
  2. import struct
  3. import math
  4. from rpython.jit.backend.zarch import conditions as con
  5. from rpython.jit.backend.zarch import masks as msk
  6. from rpython.jit.backend.zarch import registers as r
  7. from rpython.jit.backend.zarch.assembler import AssemblerZARCH
  8. from rpython.jit.backend.zarch import locations as loc
  9. from rpython.jit.backend.zarch.test.support import run_asm
  10. from rpython.jit.backend.detect_cpu import getcpuclass
  11. from rpython.jit.metainterp.resoperation import rop
  12. from rpython.jit.codewriter import longlong
  13. from rpython.rtyper.annlowlevel import llhelper
  14. from rpython.rtyper.lltypesystem import lltype, rffi, ll2ctypes
  15. from rpython.jit.metainterp.history import JitCellToken
  16. from rpython.jit.backend.model import CompiledLoopToken
  17. from rpython.rtyper.lltypesystem import lltype, llmemory, rffi
  18. from rpython.rtyper.annlowlevel import llhelper
  19. from rpython.rlib.objectmodel import specialize
  20. from rpython.rlib.debug import ll_assert
  21. from rpython.rlib.longlong2float import (float2longlong,
  22. DOUBLE_ARRAY_PTR, singlefloat2uint_emulator)
  23. from rpython.rlib.rarithmetic import r_uint, intmask
  24. import ctypes
  25. CPU = getcpuclass()
  26. def BFL(value, short=False):
  27. if short:
  28. return struct.pack('f', value)
  29. return struct.pack('>q', float2longlong(value))
  30. def ADDR(value):
  31. ptr = ll2ctypes.lltype2ctypes(value)
  32. addr = ctypes.addressof(ptr.contents.items)
  33. return struct.pack('>Q', addr)
  34. def gen_func_prolog(mc):
  35. STACK_FRAME_SIZE = 40
  36. mc.STMG(r.r11, r.r15, loc.addr(-STACK_FRAME_SIZE, r.SP))
  37. mc.AHI(r.SP, loc.imm(-STACK_FRAME_SIZE))
  38. def gen_func_epilog(mc):
  39. mc.LMG(r.r11, r.r15, loc.addr(0, r.SP))
  40. mc.BCR_rr(0xf, r.r14.value) # jmp to
  41. def isclose(a,b, rel_tol=1e-9, abs_tol=0.0):
  42. if math.isnan(a) and math.isnan(b):
  43. return True
  44. if a == b:
  45. return True
  46. # from PEP 485, added in python 3.5
  47. return abs(a-b) <= max( rel_tol * max(abs(a), abs(b)), abs_tol )
  48. class LiteralPoolCtx(object):
  49. def __init__(self, asm):
  50. self.asm = asm
  51. self.lit_label = LabelCtx(asm, 'lit')
  52. def __enter__(self):
  53. self.lit_label.__enter__()
  54. self.asm.mc.BRAS(r.r13, loc.imm(0))
  55. return self
  56. def __exit__(self, a, b, c):
  57. self.lit_label.__exit__(None, None, None)
  58. self.asm.jump_here(self.asm.mc.BRAS, 'lit')
  59. def addr(self, mem):
  60. self.asm.mc.write(ADDR(mem))
  61. def float(self, val):
  62. self.asm.mc.write(BFL(val))
  63. def single_float(self, val):
  64. self.asm.mc.write(BFL(val, short=True))
  65. def int64(self, val):
  66. self.asm.mc.write(struct.pack('>q', val))
  67. class LabelCtx(object):
  68. def __init__(self, asm, name):
  69. self.asm = asm
  70. self.name = name
  71. def __enter__(self):
  72. self.asm.mc.mark_op(self.name)
  73. return self
  74. def __exit__(self, a, b, c):
  75. self.asm.mc.mark_op(self.name + '.end')
  76. class ActivationRecordCtx(object):
  77. def __init__(self, asm, name='func'):
  78. self.asm = asm
  79. self.name = name
  80. self.asm.mc.mark_op(self.name)
  81. def __enter__(self):
  82. gen_func_prolog(self.asm.mc)
  83. return self
  84. def __exit__(self, a, b, c):
  85. gen_func_epilog(self.asm.a.mc)
  86. self.asm.mc.mark_op(self.name + '.end')
  87. class TestRunningAssembler(object):
  88. def setup_method(self, method):
  89. cpu = CPU(None, None)
  90. self.a = AssemblerZARCH(cpu)
  91. self.a.setup_once()
  92. token = JitCellToken()
  93. clt = CompiledLoopToken(cpu, 0)
  94. clt.allgcrefs = []
  95. token.compiled_loop_token = clt
  96. self.a.setup(token)
  97. self.mc = self.a.mc
  98. def test_make_operation_list(self):
  99. i = rop.INT_ADD
  100. from rpython.jit.backend.zarch import assembler
  101. assert assembler.asm_operations[i] \
  102. is AssemblerZARCH.emit_int_add.im_func
  103. def test_sync(self):
  104. self.a.mc.XGR(r.r2, r.r2)
  105. self.a.mc.sync()
  106. self.a.mc.BCR(con.ANY, r.r14)
  107. assert run_asm(self.a) == 0
  108. def test_load_64bit(self):
  109. self.a.mc.load_imm(r.r2, 0x0fffFFFF)
  110. self.a.mc.BCR(con.ANY, r.r14)
  111. assert run_asm(self.a) == 0x0fffFFFF
  112. def test_load_64bit_2(self):
  113. self.a.mc.load_imm(r.r2, 0xffffFFFF)
  114. self.a.mc.BCR(con.ANY, r.r14)
  115. assert run_asm(self.a) == 0xffffFFFF
  116. def test_load_64bit_3(self):
  117. self.a.mc.load_imm(r.r2, 2177165728)
  118. self.a.mc.BCR(con.ANY, r.r14)
  119. assert run_asm(self.a) == 2177165728
  120. def test_byte_count_instr(self):
  121. assert self.mc.BRC_byte_count == 4
  122. assert self.mc.LG_byte_count == 6
  123. def test_facility(self):
  124. adr = self.a.datablockwrapper.malloc_aligned(16, 16)
  125. self.a.mc.load_imm(r.r2, adr)
  126. self.a.mc.STFLE(loc.addr(0,r.r2))
  127. self.a.mc.BCR(con.ANY, r.r14)
  128. run_asm(self.a)
  129. fac_data = rffi.cast(rffi.CArrayPtr(rffi.ULONG), adr)
  130. f64 = bin(fac_data[0])[2:]
  131. s64 = bin(fac_data[1])[2:]
  132. print(f64)
  133. print(s64)
  134. for i,c in enumerate(f64):
  135. print('index: %d is set? %s' % (i,c))
  136. assert f64[1] == '1' # The z/Architecture architectural mode is installed.
  137. assert f64[2] == '1' # The z/Architecture architectural mode is active.
  138. assert f64[18] == '1' # long displacement facility
  139. assert f64[21] == '1' # extended immediate facility
  140. assert f64[34] == '1' # general instruction facility
  141. assert f64[41] == '1' # floating-point-support-enhancement
  142. def test_load_byte_zero_extend(self):
  143. adr = self.a.datablockwrapper.malloc_aligned(16, 16)
  144. data = rffi.cast(rffi.CArrayPtr(rffi.ULONG), adr)
  145. data[0] = rffi.cast(rffi.ULONG, intmask(0xffffFFFFffffFF02))
  146. self.a.mc.load_imm(r.r3, adr+7)
  147. self.a.mc.LLGC(r.r2, loc.addr(0,r.r3))
  148. self.a.mc.BCR(con.ANY, r.r14)
  149. assert run_asm(self.a) == 2
  150. def test_load_byte_and_imm(self):
  151. adr = self.a.datablockwrapper.malloc_aligned(16, 16)
  152. data = rffi.cast(rffi.CArrayPtr(rffi.ULONG), adr)
  153. data[0] = rffi.cast(rffi.ULONG, intmask(0xffffFFFFffff0001))
  154. self.a.mc.load_imm(r.r3, adr)
  155. self.a.mc.LG(r.r2, loc.addr(0,r.r3))
  156. self.a.mc.LLGC(r.r2, loc.addr(7,r.r3))
  157. self.a.mc.NILL(r.r2, loc.imm(0x0))
  158. self.a.mc.BCR(con.ANY, r.r14)
  159. assert run_asm(self.a) == 0
  160. @py.test.mark.parametrize('p,v', [(0,0),(8,8),(7,0),(4,0),(1,0),(9,8)])
  161. def test_align(self, p, v):
  162. WORD = 8
  163. self.a.mc.load_imm(r.r2, p)
  164. self.a.mc.LGHI(r.r0, loc.imm(~(WORD-1)))
  165. self.a.mc.NGR(r.r2, r.r0)
  166. self.a.mc.BCR(con.ANY, r.r14)
  167. assert run_asm(self.a) == v
  168. @py.test.mark.parametrize('p', [2**32,2**32+1,2**63-1,2**63-2,0,1,2,3,4,5,6,7,8,10001])
  169. def test_align_withroll(self, p):
  170. self.a.mc.load_imm(r.r2, p & 0xffffFFFFffffFFFF)
  171. self.a.mc.RISBG(r.r2, r.r2, loc.imm(0), loc.imm(0x80 | 60), loc.imm(0))
  172. self.a.mc.BCR(con.ANY, r.r14)
  173. assert run_asm(self.a) == rffi.cast(rffi.ULONG,p) & ~(7)
  174. def test_nill(self):
  175. self.a.mc.load_imm(r.r2, 1)
  176. self.a.mc.load_imm(r.r3, 0x010001)
  177. self.a.mc.NILL(r.r3, loc.imm(0xFFFF))
  178. self.a.mc.BCR(con.EQ, r.r14) # should not branch
  179. self.a.mc.load_imm(r.r2, 0) # should return here
  180. self.a.mc.BCR(con.ANY, r.r14)
  181. assert run_asm(self.a) == 0
  182. def test_complement(self):
  183. self.a.mc.load_imm(r.r2, 0)
  184. #self.a.mc.LCGR(r.r2, r.r2)
  185. self.a.mc.XIHF(r.r2, loc.imm(0xffffFFFF))
  186. self.a.mc.XILF(r.r2, loc.imm(0xffffFFFF))
  187. self.a.mc.BCR(con.ANY, r.r14)
  188. assert run_asm(self.a) == -1
  189. def test_and_7_with_risbgn(self):
  190. n = 13
  191. l = loc
  192. self.a.mc.load_imm(r.r2, 7<<n)
  193. self.a.mc.RISBG(r.r2, r.r2, l.imm(61), l.imm(0x80 | 63), l.imm(64-n))
  194. self.a.mc.BCR(con.ANY, r.r14)
  195. assert run_asm(self.a) == 7
  196. def test_risbgn(self):
  197. n = 16
  198. l = loc
  199. self.a.mc.load_imm(r.r2, 0xffFFffFF)
  200. self.a.mc.RISBG(r.r2, r.r2, l.imm(60), l.imm(0x80 | 63), l.imm(64-n))
  201. self.a.mc.BCR(con.ANY, r.r14)
  202. assert run_asm(self.a) == 15
  203. def test_shift_same_register(self):
  204. self.a.mc.load_imm(r.r3, 0x1)
  205. self.a.mc.SLLG(r.r2, r.r3, loc.addr(1))
  206. self.a.mc.BCR(con.ANY, r.r14)
  207. assert run_asm(self.a) == 2
  208. def test_shift_arith(self):
  209. self.a.mc.load_imm(r.r2, -14)
  210. self.a.mc.SLAG(r.r2, r.r2, loc.addr(1))
  211. self.a.mc.BCR(con.ANY, r.r14)
  212. assert run_asm(self.a) == -28
  213. def test_shift_negative_logical(self):
  214. self.a.mc.load_imm(r.r2, -14)
  215. self.a.mc.SLLG(r.r2, r.r2, loc.addr(1))
  216. self.a.mc.BCR(con.ANY, r.r14)
  217. assert run_asm(self.a) == -28
  218. def test_shift_negative_logical_2(self):
  219. self.a.mc.load_imm(r.r2, -2)
  220. self.a.mc.SLLG(r.r2, r.r2, loc.addr(63))
  221. self.a.mc.BCR(con.ANY, r.r14)
  222. assert run_asm(self.a) == 0
  223. def test_shift_negative_logical_3(self):
  224. self.a.mc.load_imm(r.r2, -2)
  225. self.a.mc.SLLG(r.r3, r.r2, loc.addr(1))
  226. self.a.mc.BCR(con.ANY, r.r14)
  227. assert run_asm(self.a) == -2
  228. def test_load_small_int_to_reg(self):
  229. self.a.mc.LGHI(r.r2, loc.imm(123))
  230. self.a.jmpto(r.r14)
  231. assert run_asm(self.a) == 123
  232. def test_prolog_epilog(self):
  233. gen_func_prolog(self.a.mc)
  234. self.a.mc.LGHI(r.r2, loc.imm(123))
  235. gen_func_epilog(self.a.mc)
  236. assert run_asm(self.a) == 123
  237. def test_simple_func(self):
  238. # enter
  239. self.a.mc.STMG(r.r11, r.r15, loc.addr(-96, r.SP))
  240. self.a.mc.AHI(r.SP, loc.imm(-96))
  241. # from the start of BRASL to end of jmpto there are 8+6 bytes
  242. self.a.mc.BRASL(r.r14, loc.imm(8+6))
  243. self.a.mc.LMG(r.r11, r.r15, loc.addr(0, r.SP))
  244. self.a.jmpto(r.r14)
  245. addr = self.a.mc.get_relative_pos()
  246. assert addr & 0x1 == 0
  247. gen_func_prolog(self.a.mc)
  248. self.a.mc.LGHI(r.r2, loc.imm(321))
  249. gen_func_epilog(self.a.mc)
  250. assert run_asm(self.a) == 321
  251. def test_simple_loop(self):
  252. self.a.mc.LGHI(r.r3, loc.imm(2**15-1))
  253. self.a.mc.LGHI(r.r4, loc.imm(1))
  254. L1 = self.a.mc.get_relative_pos()
  255. self.a.mc.SGR(r.r3, r.r4)
  256. LJ = self.a.mc.get_relative_pos()
  257. self.a.mc.BRCL(con.GT, loc.imm(L1-LJ))
  258. self.a.mc.LGR(r.r2, r.r3)
  259. self.a.jmpto(r.r14)
  260. assert run_asm(self.a) == 0
  261. def test_and_imm(self):
  262. self.a.mc.NIHH(r.r2, loc.imm(0))
  263. self.a.mc.NIHL(r.r2, loc.imm(0))
  264. self.a.mc.NILL(r.r2, loc.imm(0))
  265. self.a.mc.NILH(r.r2, loc.imm(0))
  266. self.a.jmpto(r.r14)
  267. assert run_asm(self.a) == 0
  268. def test_or_imm(self):
  269. self.a.mc.OIHH(r.r2, loc.imm(0xffff))
  270. self.a.mc.OIHL(r.r2, loc.imm(0xffff))
  271. self.a.mc.OILL(r.r2, loc.imm(0xffff))
  272. self.a.mc.OILH(r.r2, loc.imm(0xffff))
  273. self.a.jmpto(r.r14)
  274. assert run_asm(self.a) == -1
  275. def test_or_bitpos_0to15(self):
  276. self.a.mc.XGR(r.r2, r.r2)
  277. self.a.mc.OIHH(r.r2, loc.imm(0x0000))
  278. self.a.mc.OIHL(r.r2, loc.imm(0x0000))
  279. self.a.mc.OILL(r.r2, loc.imm(0x0000))
  280. self.a.mc.OILH(r.r2, loc.imm(0x300c))
  281. self.a.jmpto(r.r14)
  282. res = run_asm(self.a)
  283. assert res == 0x00000000300c0000
  284. def test_uint_rshift(self):
  285. self.a.mc.XGR(r.r4, r.r4)
  286. self.a.mc.LGFI(r.r5, loc.imm(63))
  287. self.a.mc.NGR(r.r4, r.r5)
  288. self.a.mc.LGFI(r.r3, loc.imm(18))
  289. self.a.mc.LGFI(r.r2, loc.imm(-1))
  290. self.a.mc.SRLG(r.r2, r.r3, loc.addr(18))
  291. self.a.jmpto(r.r14)
  292. assert run_asm(self.a) == 0
  293. def test_generate_max_integral_64bit(self):
  294. self.a.mc.LGHI(r.r2, loc.imm(-1))
  295. self.a.mc.RISBG(r.r2, r.r2, loc.imm(1), loc.imm(0x80 | 63), loc.imm(0))
  296. self.a.jmpto(r.r14)
  297. assert run_asm(self.a) == 2**63-1
  298. def test_generate_sign_bit(self):
  299. self.a.mc.LGHI(r.r2, loc.imm(-1))
  300. self.a.mc.RISBG(r.r2, r.r2, loc.imm(0), loc.imm(0x80 | 0), loc.imm(0))
  301. self.a.jmpto(r.r14)
  302. assert run_asm(self.a) == -2**63
  303. def test_ag_overflow(self):
  304. self.a.mc.BRC(con.ANY, loc.imm(4+8+8))
  305. self.a.mc.write('\x7f' + '\xff' * 7)
  306. self.a.mc.write('\x7f' + '\xff' * 7)
  307. self.a.mc.LARL(r.r5, loc.imm(-8))
  308. self.a.mc.LG(r.r4, loc.addr(8,r.r5))
  309. self.a.mc.AG(r.r4, loc.addr(0,r.r5))
  310. self.a.mc.LGR(r.r2, r.r4)
  311. self.a.jmpto(r.r14)
  312. assert run_asm(self.a) == -2
  313. def test_xor(self):
  314. self.a.mc.XGR(r.r2, r.r2)
  315. self.a.jmpto(r.r14)
  316. assert run_asm(self.a) == 0
  317. def test_literal_pool(self):
  318. gen_func_prolog(self.a.mc)
  319. self.a.mc.BRAS(r.r13, loc.imm(8 + self.mc.BRAS_byte_count))
  320. self.a.mc.write('\x08\x07\x06\x05\x04\x03\x02\x01')
  321. self.a.mc.LG(r.r2, loc.addr(0, r.r13))
  322. gen_func_epilog(self.a.mc)
  323. assert run_asm(self.a) == 0x0807060504030201
  324. def label(self, name, func=False):
  325. if not func:
  326. return LabelCtx(self, name)
  327. return ActivationRecordCtx(self, name)
  328. def patch_branch_imm16(self, base, imm):
  329. imm = (imm & 0xffff) >> 1
  330. self.mc.overwrite(base, chr((imm >> 8) & 0xFF))
  331. self.mc.overwrite(base+1, chr(imm & 0xFF))
  332. def pos(self, name):
  333. return self.mc.ops_offset[name]
  334. def cur(self):
  335. return self.mc.get_relative_pos()
  336. def jump_here(self, func, name):
  337. if func.__name__ == 'BRAS':
  338. self.patch_branch_imm16(self.pos(name)+2, self.cur() - self.pos(name))
  339. else:
  340. raise NotImplementedError
  341. def jump_to(self, reg, label):
  342. val = (self.pos(label) - self.cur())
  343. self.mc.BRAS(reg, loc.imm(val))
  344. def test_stmg(self):
  345. self.mc.LGR(r.r2, r.r15)
  346. self.a.jmpto(r.r14)
  347. print hex(run_asm(self.a))
  348. def test_recursion(self):
  349. with ActivationRecordCtx(self):
  350. with self.label('lit'):
  351. self.mc.BRAS(r.r13, loc.imm(0))
  352. self.mc.write('\x00\x00\x00\x00\x00\x00\x00\x00')
  353. self.jump_here(self.mc.BRAS, 'lit')
  354. # recurse X times
  355. self.mc.XGR(r.r2, r.r2)
  356. self.mc.LGHI(r.r9, loc.imm(15))
  357. with self.label('L1'):
  358. self.mc.BRAS(r.r14, loc.imm(0))
  359. with ActivationRecordCtx(self, 'rec'):
  360. self.mc.AGR(r.r2, r.r9)
  361. self.mc.AHI(r.r9, loc.imm(-1))
  362. # if not entered recursion, return from activation record
  363. # implicitly generated here by with statement
  364. self.mc.BRC(con.GT, loc.imm(self.pos('rec') - self.cur()))
  365. self.jump_here(self.mc.BRAS, 'L1')
  366. # call rec... recursivly
  367. self.jump_to(r.r14, 'rec')
  368. self.a.jmpto(r.r14)
  369. assert run_asm(self.a) == 120
  370. def test_printf(self):
  371. with ActivationRecordCtx(self):
  372. with self.label('lit'):
  373. self.mc.BRAS(r.r13, loc.imm(0))
  374. for c in "hello syscall\n":
  375. self.mc.writechar(c)
  376. self.jump_here(self.mc.BRAS, 'lit')
  377. self.mc.LGHI(r.r2, loc.imm(1)) # stderr
  378. self.mc.LA(r.r3, loc.addr(0, r.r13)) # char*
  379. self.mc.LGHI(r.r4, loc.imm(14)) # length
  380. # write sys call
  381. self.mc.SVC(loc.imm(4))
  382. self.a.jmpto(r.r14)
  383. assert run_asm(self.a) == 14
  384. def test_float(self):
  385. with ActivationRecordCtx(self):
  386. with self.label('lit'):
  387. self.mc.BRAS(r.r13, loc.imm(0))
  388. self.mc.write(BFL(-15.0))
  389. self.jump_here(self.mc.BRAS, 'lit')
  390. self.mc.LD(r.f0, loc.addr(0, r.r13))
  391. self.mc.CGDBR(r.r2, msk.RND_CURMODE, r.f0)
  392. self.a.jmpto(r.r14)
  393. assert run_asm(self.a) == -15
  394. @py.test.mark.parametrize("v1,v2,res", [
  395. ( 0.0, 0.0, 0.0),
  396. ( -15.0, -15.0, -30.0),
  397. ( 1.5, -3.22, -1.72),
  398. ( 0.5, 0.0, 0.5),
  399. ( 0.0001, -0.0002, -0.0001),
  400. (float('nan'), 1.0, float('nan')),
  401. ])
  402. def test_float_to_memory(self, v1, v2, res):
  403. with lltype.scoped_alloc(DOUBLE_ARRAY_PTR.TO, 16) as mem:
  404. with ActivationRecordCtx(self):
  405. with self.label('lit'):
  406. self.mc.BRAS(r.r13, loc.imm(0))
  407. self.mc.write(BFL(v1))
  408. self.mc.write(BFL(v2))
  409. self.mc.write(ADDR(mem))
  410. self.jump_here(self.mc.BRAS, 'lit')
  411. self.mc.LD(r.f0, loc.addr(0, r.r13))
  412. self.mc.LD(r.f1, loc.addr(8, r.r13))
  413. self.mc.ADBR(r.f0, r.f1)
  414. self.mc.LG(r.r11, loc.addr(16, r.r13))
  415. self.mc.STD(r.f0, loc.addr(0, r.r11))
  416. self.a.jmpto(r.r14)
  417. run_asm(self.a)
  418. assert isclose(mem[0],res)
  419. @py.test.mark.parametrize("v1,v2,res", [
  420. ( 0.0, 0.0, 0.0),
  421. ( -15.0, -15.0, 225.0),
  422. ( 0.0, 9876543.21, 0.0),
  423. ( -0.5, 14.5, -7.25),
  424. ( 0.0001, 2.0, 0.0002),
  425. (float('nan'), 1.0, float('nan')),
  426. ])
  427. def test_float_mul_to_memory(self, v1, v2, res):
  428. with lltype.scoped_alloc(DOUBLE_ARRAY_PTR.TO, 16) as mem:
  429. with ActivationRecordCtx(self):
  430. with LiteralPoolCtx(self) as pool:
  431. pool.float(v1)
  432. pool.float(v2)
  433. pool.addr(mem)
  434. self.mc.LD(r.f0, loc.addr(0, r.r13))
  435. self.mc.MDB(r.f0, loc.addr(8, r.r13))
  436. self.mc.LG(r.r11, loc.addr(16, r.r13))
  437. self.mc.STD(r.f0, loc.addr(0, r.r11))
  438. self.a.jmpto(r.r14)
  439. run_asm(self.a)
  440. assert isclose(mem[0],res)
  441. def test_float_load_zero(self):
  442. with lltype.scoped_alloc(DOUBLE_ARRAY_PTR.TO, 16) as mem:
  443. with ActivationRecordCtx(self):
  444. with LiteralPoolCtx(self) as pool:
  445. pool.addr(mem)
  446. self.mc.LZDR(r.f0)
  447. self.mc.LG(r.r11, loc.addr(0, r.r13))
  448. self.mc.STD(r.f0, loc.addr(0, r.r11))
  449. run_asm(self.a)
  450. assert isclose(mem[0], 0.0)
  451. def test_cast_single_float_to_float(self):
  452. with lltype.scoped_alloc(DOUBLE_ARRAY_PTR.TO, 16) as mem:
  453. with ActivationRecordCtx(self):
  454. with LiteralPoolCtx(self) as pool:
  455. pool.single_float(6.66)
  456. pool.addr(mem)
  457. self.mc.LEY(r.f1, loc.addr(0, r.r13))
  458. ## cast short to long!
  459. self.mc.LDEBR(r.f0, r.f1)
  460. self.mc.LG(r.r11, loc.addr(4, r.r13))
  461. self.mc.STD(r.f0, loc.addr(0, r.r11))
  462. run_asm(self.a)
  463. assert isclose(mem[0], 6.66, abs_tol=0.05)
  464. def test_cast_int64_to_float(self):
  465. with lltype.scoped_alloc(DOUBLE_ARRAY_PTR.TO, 16) as mem:
  466. with ActivationRecordCtx(self):
  467. with LiteralPoolCtx(self) as pool:
  468. pool.int64(12345)
  469. pool.addr(mem)
  470. self.mc.LG(r.r12, loc.addr(0, r.r13))
  471. # cast int to float!
  472. self.mc.CDGBR(r.f0, r.r12)
  473. self.mc.LG(r.r11, loc.addr(8, r.r13))
  474. self.mc.STD(r.f0, loc.addr(0, r.r11))
  475. run_asm(self.a)
  476. assert isclose(mem[0], 12345.0)
  477. def test_float_cmp(self):
  478. with ActivationRecordCtx(self):
  479. with LiteralPoolCtx(self) as pool:
  480. pool.float(1.0)
  481. pool.float(2.0)
  482. self.mc.LD(r.f0, loc.addr(0, r.r13))
  483. self.mc.LD(r.f1, loc.addr(8, r.r13))
  484. self.mc.CDBR(r.f0, r.f1)
  485. self.mc.LGHI(r.r2, loc.imm(0))
  486. self.mc.BCR(con.EQ, r.r14) # must not branch
  487. self.mc.LGHI(r.r2, loc.imm(1))
  488. self.a.jmpto(r.r14)
  489. assert run_asm(self.a) == 1
  490. def pushpop_jitframe(self, registers):
  491. self.a._push_core_regs_to_jitframe(self.mc, registers)
  492. self.a._pop_core_regs_from_jitframe(self.mc, registers)
  493. def test_pushpop_jitframe_multiple_optimization(self):
  494. stored = []
  495. loaded = []
  496. def STMG(start, end, addr):
  497. stored.append((start, end))
  498. def STG(reg, addr):
  499. stored.append((reg,))
  500. def LMG(start, end, addr):
  501. loaded.append((start, end))
  502. def LG(reg, addr):
  503. loaded.append((reg,))
  504. self.mc.STMG = STMG
  505. self.mc.STG = STG
  506. self.mc.LMG = LMG
  507. self.mc.LG = LG
  508. # 2-6
  509. self.pushpop_jitframe([r.r2, r.r3, r.r4, r.r5, r.r6, r.r8, r.r10])
  510. assert stored == [(r.r2, r.r6), (r.r8,), (r.r10,)]
  511. assert stored == loaded
  512. stored = []
  513. loaded = []
  514. # two sequences 10-11, 13-14
  515. self.pushpop_jitframe([r.r2, r.r3, r.r10, r.r11])
  516. assert stored == [(r.r2, r.r3), (r.r10, r.r11)]
  517. assert stored == loaded
  518. stored = []
  519. loaded = []
  520. # one sequence and on single
  521. self.pushpop_jitframe([r.r2, r.r3, r.r5])
  522. assert stored == [(r.r2, r.r3), (r.r5,)]
  523. assert stored == loaded
  524. stored = []
  525. loaded = []
  526. # single items
  527. self.pushpop_jitframe(r.MANAGED_REGS[::2])
  528. assert stored == [(x,) for x in r.MANAGED_REGS[::2]]
  529. assert stored == loaded
  530. stored = []
  531. loaded = []
  532. # large sequence 0-5 and one hole between
  533. self.pushpop_jitframe([r.r2, r.r3,
  534. r.r4, r.r5, r.r10, r.r11])
  535. assert stored == [(r.r2, r.r5), (r.r10, r.r11)]
  536. assert stored == loaded
  537. stored = []
  538. loaded = []
  539. # ensure there is just on instruction for the 'best case'
  540. self.pushpop_jitframe(r.MANAGED_REGS)
  541. assert stored == [(r.r2, r.r11)]
  542. assert stored == loaded
  543. stored = []
  544. loaded = []
  545. # just one single
  546. for x in [r.r10, r.r3, r.r2, r.r11]:
  547. self.pushpop_jitframe([x])
  548. assert stored == [(x,)]
  549. assert stored == loaded
  550. stored = []
  551. loaded = []
  552. # unordered
  553. self.pushpop_jitframe([r.r11, r.r8, r.r4, r.r2])
  554. assert stored == [(r.r11,), (r.r8,), (r.r4,), (r.r2,)]
  555. assert stored == loaded
  556. stored = []
  557. loaded = []