diff --git a/include/bx/simd128_langext.inl b/include/bx/simd128_langext.inl index 8b23fe2..c89e612 100644 --- a/include/bx/simd128_langext.inl +++ b/include/bx/simd128_langext.inl @@ -6,37 +6,37 @@ #ifndef BX_SIMD128_LANGEXT_H_HEADER_GUARD #define BX_SIMD128_LANGEXT_H_HEADER_GUARD -#define float4_rcp float4_rcp_ni -#define float4_orx float4_orx_ni -#define float4_orc float4_orc_ni -#define float4_neg float4_neg_ni -#define float4_madd float4_madd_ni -#define float4_nmsub float4_nmsub_ni -#define float4_div_nr float4_div_nr_ni -#define float4_selb float4_selb_ni -#define float4_sels float4_sels_ni -#define float4_not float4_not_ni -#define float4_abs float4_abs_ni -#define float4_clamp float4_clamp_ni -#define float4_lerp float4_lerp_ni -#define float4_rcp_est float4_rcp_ni -#define float4_rsqrt float4_rsqrt_ni -#define float4_rsqrt_nr float4_rsqrt_nr_ni -#define float4_rsqrt_carmack float4_rsqrt_carmack_ni -#define float4_sqrt_nr float4_sqrt_nr_ni -#define float4_log2 float4_log2_ni -#define float4_exp2 float4_exp2_ni -#define float4_pow float4_pow_ni -#define float4_cross3 float4_cross3_ni -#define float4_normalize3 float4_normalize3_ni -#define float4_dot3 float4_dot3_ni -#define float4_dot float4_dot_ni -#define float4_ceil float4_ceil_ni -#define float4_floor float4_floor_ni -#define float4_min float4_min_ni -#define float4_max float4_max_ni -#define float4_imin float4_imin_ni -#define float4_imax float4_imax_ni +#define simd_rcp simd_rcp_ni +#define simd_orx simd_orx_ni +#define simd_orc simd_orc_ni +#define simd_neg simd_neg_ni +#define simd_madd simd_madd_ni +#define simd_nmsub simd_nmsub_ni +#define simd_div_nr simd_div_nr_ni +#define simd_selb simd_selb_ni +#define simd_sels simd_sels_ni +#define simd_not simd_not_ni +#define simd_abs simd_abs_ni +#define simd_clamp simd_clamp_ni +#define simd_lerp simd_lerp_ni +#define simd_rcp_est simd_rcp_ni +#define simd_rsqrt simd_rsqrt_ni +#define simd_rsqrt_nr simd_rsqrt_nr_ni +#define simd_rsqrt_carmack simd_rsqrt_carmack_ni +#define simd_sqrt_nr simd_sqrt_nr_ni +#define simd_log2 simd_log2_ni +#define simd_exp2 simd_exp2_ni +#define simd_pow simd_pow_ni +#define simd_cross3 simd_cross3_ni +#define simd_normalize3 simd_normalize3_ni +#define simd_dot3 simd_dot3_ni +#define simd_dot simd_dot_ni +#define simd_ceil simd_ceil_ni +#define simd_floor simd_floor_ni +#define simd_min simd_min_ni +#define simd_max simd_max_ni +#define simd_imin simd_imin_ni +#define simd_imax simd_imax_ni #include "simd_ni.inl" @@ -48,14 +48,14 @@ namespace bx #define ELEMw 3 #define BX_SIMD128_IMPLEMENT_SWIZZLE(_x, _y, _z, _w) \ template<> \ - BX_SIMD_FORCE_INLINE simd_langext_t float4_swiz_##_x##_y##_z##_w(simd_langext_t _a) \ + BX_SIMD_FORCE_INLINE simd128_langext_t simd_swiz_##_x##_y##_z##_w(simd128_langext_t _a) \ { \ - simd_langext_t result; \ + simd128_langext_t result; \ result.vf = __builtin_shufflevector(_a.vf, _a.vf, ELEM##_x, ELEM##_y, ELEM##_z, ELEM##_w); \ return result; \ } -#include "float4_swizzle.inl" +#include "simd128_swizzle.inl" #undef BX_SIMD128_IMPLEMENT_SWIZZLE #undef ELEMw @@ -65,7 +65,7 @@ namespace bx #define BX_SIMD128_IMPLEMENT_TEST(_xyzw, _mask) \ template<> \ - BX_SIMD_FORCE_INLINE bool simd_test_any_##_xyzw(simd_langext_t _test) \ + BX_SIMD_FORCE_INLINE bool simd_test_any_##_xyzw(simd128_langext_t _test) \ { \ uint32_t tmp = ( (_test.uxyzw[3]>>31)<<3) \ | ( (_test.uxyzw[2]>>31)<<2) \ @@ -76,7 +76,7 @@ namespace bx } \ \ template<> \ - BX_SIMD_FORCE_INLINE bool simd_test_all_##_xyzw(simd_langext_t _test) \ + BX_SIMD_FORCE_INLINE bool simd_test_all_##_xyzw(simd128_langext_t _test) \ { \ uint32_t tmp = ( (_test.uxyzw[3]>>31)<<3) \ | ( (_test.uxyzw[2]>>31)<<2) \ @@ -105,114 +105,114 @@ BX_SIMD128_IMPLEMENT_TEST(xyzw , 0xf); #undef BX_SIMD128_IMPLEMENT_TEST template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_shuf_xyAB(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_shuf_xyAB(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = __builtin_shufflevector(_a.vf, _b.vf, 0, 1, 4, 5); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_shuf_ABxy(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_shuf_ABxy(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = __builtin_shufflevector(_a.vf, _b.vf, 4, 5, 0, 1); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_shuf_CDzw(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_shuf_CDzw(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = __builtin_shufflevector(_a.vf, _b.vf, 6, 7, 2, 3); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_shuf_zwCD(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_shuf_zwCD(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = __builtin_shufflevector(_a.vf, _b.vf, 2, 3, 6, 7); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_shuf_xAyB(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_shuf_xAyB(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = __builtin_shufflevector(_a.vf, _b.vf, 0, 4, 1, 5); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_shuf_yBxA(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_shuf_yBxA(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = __builtin_shufflevector(_a.vf, _b.vf, 1, 5, 0, 4); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_shuf_zCwD(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_shuf_zCwD(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = __builtin_shufflevector(_a.vf, _b.vf, 2, 6, 3, 7); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_shuf_CzDw(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_shuf_CzDw(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = __builtin_shufflevector(_a.vf, _b.vf, 6, 2, 7, 3); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_shuf_xAzC(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_shuf_xAzC(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = __builtin_shufflevector(_a.vf, _b.vf, 0, 4, 2, 6); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_shuf_yBwD(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_shuf_yBwD(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = __builtin_shufflevector(_a.vf, _b.vf, 1, 5, 3, 7); return result; } template<> - BX_SIMD_FORCE_INLINE float float4_x(simd_langext_t _a) + BX_SIMD_FORCE_INLINE float simd_x(simd128_langext_t _a) { return _a.fxyzw[0]; } template<> - BX_SIMD_FORCE_INLINE float float4_y(simd_langext_t _a) + BX_SIMD_FORCE_INLINE float simd_y(simd128_langext_t _a) { return _a.fxyzw[1]; } template<> - BX_SIMD_FORCE_INLINE float float4_z(simd_langext_t _a) + BX_SIMD_FORCE_INLINE float simd_z(simd128_langext_t _a) { return _a.fxyzw[2]; } template<> - BX_SIMD_FORCE_INLINE float float4_w(simd_langext_t _a) + BX_SIMD_FORCE_INLINE float simd_w(simd128_langext_t _a) { return _a.fxyzw[3]; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_ld(const void* _ptr) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_ld(const void* _ptr) { const uint32_t* input = reinterpret_cast(_ptr); - simd_langext_t result; + simd128_langext_t result; result.uxyzw[0] = input[0]; result.uxyzw[1] = input[1]; result.uxyzw[2] = input[2]; @@ -221,7 +221,7 @@ BX_SIMD128_IMPLEMENT_TEST(xyzw , 0xf); } template<> - BX_SIMD_FORCE_INLINE void float4_st(void* _ptr, simd_langext_t _a) + BX_SIMD_FORCE_INLINE void simd_st(void* _ptr, simd128_langext_t _a) { uint32_t* result = reinterpret_cast(_ptr); result[0] = _a.uxyzw[0]; @@ -231,14 +231,14 @@ BX_SIMD128_IMPLEMENT_TEST(xyzw , 0xf); } template<> - BX_SIMD_FORCE_INLINE void float4_stx(void* _ptr, simd_langext_t _a) + BX_SIMD_FORCE_INLINE void simd_stx(void* _ptr, simd128_langext_t _a) { uint32_t* result = reinterpret_cast(_ptr); result[0] = _a.uxyzw[0]; } template<> - BX_SIMD_FORCE_INLINE void float4_stream(void* _ptr, simd_langext_t _a) + BX_SIMD_FORCE_INLINE void simd_stream(void* _ptr, simd128_langext_t _a) { uint32_t* result = reinterpret_cast(_ptr); result[0] = _a.uxyzw[0]; @@ -248,109 +248,109 @@ BX_SIMD128_IMPLEMENT_TEST(xyzw , 0xf); } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_ld(float _x, float _y, float _z, float _w) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_ld(float _x, float _y, float _z, float _w) { - simd_langext_t result; + simd128_langext_t result; result.vf = (float __attribute__((vector_size(16)))){ _x, _y, _z, _w }; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_ild(uint32_t _x, uint32_t _y, uint32_t _z, uint32_t _w) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_ild(uint32_t _x, uint32_t _y, uint32_t _z, uint32_t _w) { - simd_langext_t result; + simd128_langext_t result; result.vu = (uint32_t __attribute__((vector_size(16)))){ _x, _y, _z, _w }; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_splat(const void* _ptr) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_splat(const void* _ptr) { const uint32_t val = *reinterpret_cast(_ptr); - simd_langext_t result; + simd128_langext_t result; result.vu = (uint32_t __attribute__((vector_size(16)))){ val, val, val, val }; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_splat(float _a) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_splat(float _a) { - return float4_ld(_a, _a, _a, _a); + return simd_ld(_a, _a, _a, _a); } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_isplat(uint32_t _a) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_isplat(uint32_t _a) { - return float4_ild(_a, _a, _a, _a); + return simd_ild(_a, _a, _a, _a); } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_zero() + BX_SIMD_FORCE_INLINE simd128_langext_t simd_zero() { - return float4_ild(0, 0, 0, 0); + return simd_ild(0, 0, 0, 0); } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_itof(simd_langext_t _a) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_itof(simd128_langext_t _a) { - simd_langext_t result; + simd128_langext_t result; result.vf = __builtin_convertvector(_a.vi, float __attribute__((vector_size(16))) ); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_ftoi(simd_langext_t _a) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_ftoi(simd128_langext_t _a) { - simd_langext_t result; + simd128_langext_t result; result.vi = __builtin_convertvector(_a.vf, int32_t __attribute__((vector_size(16))) ); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_round(simd_langext_t _a) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_round(simd128_langext_t _a) { - const simd_langext_t tmp = float4_ftoi(_a); - const simd_langext_t result = float4_itof(tmp); + const simd128_langext_t tmp = simd_ftoi(_a); + const simd128_langext_t result = simd_itof(tmp); return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_add(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_add(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = _a.vf + _b.vf; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_sub(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_sub(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = _a.vf - _b.vf; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_mul(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_mul(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = _a.vf * _b.vf; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_div(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_div(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vf = _a.vf / _b.vf; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_sqrt(simd_langext_t _a) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_sqrt(simd128_langext_t _a) { - simd_langext_t result; + simd128_langext_t result; result.vf[0] = sqrtf(_a.vf[0]); result.vf[1] = sqrtf(_a.vf[1]); result.vf[2] = sqrtf(_a.vf[2]); @@ -359,9 +359,9 @@ BX_SIMD128_IMPLEMENT_TEST(xyzw , 0xf); } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_rsqrt_est(simd_langext_t _a) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_rsqrt_est(simd128_langext_t _a) { - simd_langext_t result; + simd128_langext_t result; result.vf[0] = 1.0f / sqrtf(_a.vf[0]); result.vf[1] = 1.0f / sqrtf(_a.vf[1]); result.vf[2] = 1.0f / sqrtf(_a.vf[2]); @@ -370,145 +370,145 @@ BX_SIMD128_IMPLEMENT_TEST(xyzw , 0xf); } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_cmpeq(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_cmpeq(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vi = _a.vf == _b.vf; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_cmplt(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_cmplt(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vi = _a.vf < _b.vf; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_cmple(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_cmple(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vi = _a.vf <= _b.vf; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_cmpgt(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_cmpgt(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vi = _a.vf > _b.vf; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_cmpge(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_cmpge(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vi = _a.vf >= _b.vf; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_and(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_and(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vu = _a.vu & _b.vu; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_andc(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_andc(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vu = _a.vu & ~_b.vu; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_or(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_or(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vu = _a.vu | _b.vu; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_xor(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_xor(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vu = _a.vu ^ _b.vu; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_sll(simd_langext_t _a, int _count) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_sll(simd128_langext_t _a, int _count) { - simd_langext_t result; - const simd_langext_t count = float4_isplat(_count); + simd128_langext_t result; + const simd128_langext_t count = simd_isplat(_count); result.vu = _a.vu << count.vi; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_srl(simd_langext_t _a, int _count) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_srl(simd128_langext_t _a, int _count) { - simd_langext_t result; - const simd_langext_t count = float4_isplat(_count); + simd128_langext_t result; + const simd128_langext_t count = simd_isplat(_count); result.vu = _a.vu >> count.vi; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_sra(simd_langext_t _a, int _count) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_sra(simd128_langext_t _a, int _count) { - simd_langext_t result; - const simd_langext_t count = float4_isplat(_count); + simd128_langext_t result; + const simd128_langext_t count = simd_isplat(_count); result.vi = _a.vi >> count.vi; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_icmpeq(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_icmpeq(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vi = _a.vi == _b.vi; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_icmplt(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_icmplt(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vi = _a.vi < _b.vi; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_icmpgt(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_icmpgt(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vi = _a.vi > _b.vi; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_iadd(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_iadd(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vi = _a.vi + _b.vi; return result; } template<> - BX_SIMD_FORCE_INLINE simd_langext_t float4_isub(simd_langext_t _a, simd_langext_t _b) + BX_SIMD_FORCE_INLINE simd128_langext_t simd_isub(simd128_langext_t _a, simd128_langext_t _b) { - simd_langext_t result; + simd128_langext_t result; result.vi = _a.vi - _b.vi; return result; } - typedef simd_langext_t simd128_t; + typedef simd128_langext_t simd128_t; } // namespace bx