From e8816f124cd1665b325d86d190f05ad9332f9370 Mon Sep 17 00:00:00 2001 From: Kim Morrison Date: Tue, 10 Sep 2024 16:59:07 +1000 Subject: [PATCH] chore: update stage0 --- stage0/stdlib/Init/Data.c | 10 +- stage0/stdlib/Init/Data/Fin/Basic.c | 33 +- stage0/stdlib/Init/Data/List/Nat/TakeDrop.c | 6 +- stage0/stdlib/Init/Data/Nat/Basic.c | 6 +- stage0/stdlib/Init/Data/NeZero.c | 29 + stage0/stdlib/Init/Data/Zero.c | 83 + stage0/stdlib/Lean/Meta/LitValues.c | 127 +- stage0/stdlib/Lean/Meta/Match/Match.c | 1042 +++++----- .../Meta/Tactic/Simp/BuiltinSimprocs/BitVec.c | 225 +- .../Meta/Tactic/Simp/BuiltinSimprocs/Fin.c | 1843 +++++++++-------- stage0/stdlib/Lean/ToExpr.c | 54 +- .../Tactic/BVDecide/Bitblast/BVExpr/Basic.c | 28 +- 12 files changed, 1955 insertions(+), 1531 deletions(-) create mode 100644 stage0/stdlib/Init/Data/NeZero.c create mode 100644 stage0/stdlib/Init/Data/Zero.c diff --git a/stage0/stdlib/Init/Data.c b/stage0/stdlib/Init/Data.c index 4657b9650ef3..b353c251ae26 100644 --- a/stage0/stdlib/Init/Data.c +++ b/stage0/stdlib/Init/Data.c @@ -1,6 +1,6 @@ // Lean compiler output // Module: Init.Data -// Imports: Init.Data.Basic Init.Data.Nat Init.Data.Bool Init.Data.BitVec Init.Data.Cast Init.Data.Char Init.Data.String Init.Data.List Init.Data.Int Init.Data.Array Init.Data.Array.Subarray.Split Init.Data.ByteArray Init.Data.FloatArray Init.Data.Fin Init.Data.UInt Init.Data.Float Init.Data.Option Init.Data.Ord Init.Data.Random Init.Data.ToString Init.Data.Range Init.Data.Hashable Init.Data.OfScientific Init.Data.Format Init.Data.Stream Init.Data.Prod Init.Data.AC Init.Data.Queue Init.Data.Channel Init.Data.Cast Init.Data.Sum Init.Data.BEq Init.Data.Subtype Init.Data.ULift Init.Data.PLift +// Imports: Init.Data.Basic Init.Data.Nat Init.Data.Bool Init.Data.BitVec Init.Data.Cast Init.Data.Char Init.Data.String Init.Data.List Init.Data.Int Init.Data.Array Init.Data.Array.Subarray.Split Init.Data.ByteArray Init.Data.FloatArray Init.Data.Fin Init.Data.UInt Init.Data.Float Init.Data.Option Init.Data.Ord Init.Data.Random Init.Data.ToString Init.Data.Range Init.Data.Hashable Init.Data.OfScientific Init.Data.Format Init.Data.Stream Init.Data.Prod Init.Data.AC Init.Data.Queue Init.Data.Channel Init.Data.Cast Init.Data.Sum Init.Data.BEq Init.Data.Subtype Init.Data.ULift Init.Data.PLift Init.Data.Zero Init.Data.NeZero #include #if defined(__clang__) #pragma clang diagnostic ignored "-Wunused-parameter" @@ -48,6 +48,8 @@ lean_object* initialize_Init_Data_BEq(uint8_t builtin, lean_object*); lean_object* initialize_Init_Data_Subtype(uint8_t builtin, lean_object*); lean_object* initialize_Init_Data_ULift(uint8_t builtin, lean_object*); lean_object* initialize_Init_Data_PLift(uint8_t builtin, lean_object*); +lean_object* initialize_Init_Data_Zero(uint8_t builtin, lean_object*); +lean_object* initialize_Init_Data_NeZero(uint8_t builtin, lean_object*); static bool _G_initialized = false; LEAN_EXPORT lean_object* initialize_Init_Data(uint8_t builtin, lean_object* w) { lean_object * res; @@ -158,6 +160,12 @@ lean_dec_ref(res); res = initialize_Init_Data_PLift(builtin, lean_io_mk_world()); if (lean_io_result_is_error(res)) return res; lean_dec_ref(res); +res = initialize_Init_Data_Zero(builtin, lean_io_mk_world()); +if (lean_io_result_is_error(res)) return res; +lean_dec_ref(res); +res = initialize_Init_Data_NeZero(builtin, lean_io_mk_world()); +if (lean_io_result_is_error(res)) return res; +lean_dec_ref(res); return lean_io_result_mk_ok(lean_box(0)); } #ifdef __cplusplus diff --git a/stage0/stdlib/Init/Data/Fin/Basic.c b/stage0/stdlib/Init/Data/Fin/Basic.c index 92562d545c5d..72337a4be0d7 100644 --- a/stage0/stdlib/Init/Data/Fin/Basic.c +++ b/stage0/stdlib/Init/Data/Fin/Basic.c @@ -46,7 +46,7 @@ LEAN_EXPORT lean_object* l_Fin_ofNat_x27___boxed(lean_object*, lean_object*, lea LEAN_EXPORT lean_object* l_Fin_instDiv___boxed(lean_object*); LEAN_EXPORT lean_object* l_Fin_castLT___rarg(lean_object*, lean_object*); LEAN_EXPORT lean_object* l_Fin_instDiv(lean_object*); -LEAN_EXPORT lean_object* l_Fin_instOfNat(lean_object*, lean_object*); +LEAN_EXPORT lean_object* l_Fin_instOfNat(lean_object*, lean_object*, lean_object*); LEAN_EXPORT lean_object* l_Fin_ofNat___boxed(lean_object*, lean_object*); LEAN_EXPORT lean_object* l_Fin_instMod___boxed(lean_object*); LEAN_EXPORT lean_object* l_Fin_castSucc___rarg___boxed(lean_object*); @@ -86,7 +86,7 @@ LEAN_EXPORT lean_object* l_Fin_subNat___boxed(lean_object*); LEAN_EXPORT lean_object* l_Fin_subNat(lean_object*); LEAN_EXPORT lean_object* l_Fin_castLT___boxed(lean_object*, lean_object*); lean_object* lean_nat_lxor(lean_object*, lean_object*); -LEAN_EXPORT lean_object* l_Fin_instOfNat___boxed(lean_object*, lean_object*); +LEAN_EXPORT lean_object* l_Fin_instOfNat___boxed(lean_object*, lean_object*, lean_object*); LEAN_EXPORT lean_object* l_Fin_cast___rarg___boxed(lean_object*); LEAN_EXPORT lean_object* l_Fin_modn___rarg(lean_object*, lean_object*); LEAN_EXPORT lean_object* l_Fin_shiftRight(lean_object*, lean_object*, lean_object*); @@ -645,31 +645,34 @@ lean_closure_set(x_2, 0, x_1); return x_2; } } -LEAN_EXPORT lean_object* l_Fin_instOfNat(lean_object* x_1, lean_object* x_2) { +LEAN_EXPORT lean_object* l_Fin_instOfNat(lean_object* x_1, lean_object* x_2, lean_object* x_3) { _start: { -lean_object* x_3; -x_3 = l_Fin_ofNat(x_1, x_2); -return x_3; +lean_object* x_4; +x_4 = lean_nat_mod(x_3, x_1); +return x_4; } } -LEAN_EXPORT lean_object* l_Fin_instOfNat___boxed(lean_object* x_1, lean_object* x_2) { +LEAN_EXPORT lean_object* l_Fin_instOfNat___boxed(lean_object* x_1, lean_object* x_2, lean_object* x_3) { _start: { -lean_object* x_3; -x_3 = l_Fin_instOfNat(x_1, x_2); -lean_dec(x_2); +lean_object* x_4; +x_4 = l_Fin_instOfNat(x_1, x_2, x_3); +lean_dec(x_3); lean_dec(x_1); -return x_3; +return x_4; } } LEAN_EXPORT lean_object* l_Fin_instInhabitedHAddNatOfNat(lean_object* x_1) { _start: { -lean_object* x_2; lean_object* x_3; -x_2 = lean_unsigned_to_nat(0u); -x_3 = l_Fin_ofNat(x_1, x_2); -return x_3; +lean_object* x_2; lean_object* x_3; lean_object* x_4; lean_object* x_5; +x_2 = lean_unsigned_to_nat(1u); +x_3 = lean_nat_add(x_1, x_2); +x_4 = lean_unsigned_to_nat(0u); +x_5 = lean_nat_mod(x_4, x_3); +lean_dec(x_3); +return x_5; } } LEAN_EXPORT lean_object* l_Fin_instInhabitedHAddNatOfNat___boxed(lean_object* x_1) { diff --git a/stage0/stdlib/Init/Data/List/Nat/TakeDrop.c b/stage0/stdlib/Init/Data/List/Nat/TakeDrop.c index cf3797481d0c..7ee63f378fd9 100644 --- a/stage0/stdlib/Init/Data/List/Nat/TakeDrop.c +++ b/stage0/stdlib/Init/Data/List/Nat/TakeDrop.c @@ -1,6 +1,6 @@ // Lean compiler output // Module: Init.Data.List.Nat.TakeDrop -// Imports: Init.Data.List.Zip Init.Data.List.Sublist Init.Data.Nat.Lemmas +// Imports: Init.Data.List.Zip Init.Data.List.Sublist Init.Data.List.Find Init.Data.Nat.Lemmas #include #if defined(__clang__) #pragma clang diagnostic ignored "-Wunused-parameter" @@ -15,6 +15,7 @@ extern "C" { #endif lean_object* initialize_Init_Data_List_Zip(uint8_t builtin, lean_object*); lean_object* initialize_Init_Data_List_Sublist(uint8_t builtin, lean_object*); +lean_object* initialize_Init_Data_List_Find(uint8_t builtin, lean_object*); lean_object* initialize_Init_Data_Nat_Lemmas(uint8_t builtin, lean_object*); static bool _G_initialized = false; LEAN_EXPORT lean_object* initialize_Init_Data_List_Nat_TakeDrop(uint8_t builtin, lean_object* w) { @@ -27,6 +28,9 @@ lean_dec_ref(res); res = initialize_Init_Data_List_Sublist(builtin, lean_io_mk_world()); if (lean_io_result_is_error(res)) return res; lean_dec_ref(res); +res = initialize_Init_Data_List_Find(builtin, lean_io_mk_world()); +if (lean_io_result_is_error(res)) return res; +lean_dec_ref(res); res = initialize_Init_Data_Nat_Lemmas(builtin, lean_io_mk_world()); if (lean_io_result_is_error(res)) return res; lean_dec_ref(res); diff --git a/stage0/stdlib/Init/Data/Nat/Basic.c b/stage0/stdlib/Init/Data/Nat/Basic.c index fcd8fd7315d7..fa2acfa65f99 100644 --- a/stage0/stdlib/Init/Data/Nat/Basic.c +++ b/stage0/stdlib/Init/Data/Nat/Basic.c @@ -1,6 +1,6 @@ // Lean compiler output // Module: Init.Data.Nat.Basic -// Imports: Init.SimpLemmas +// Imports: Init.SimpLemmas Init.Data.NeZero #include #if defined(__clang__) #pragma clang diagnostic ignored "-Wunused-parameter" @@ -1020,6 +1020,7 @@ return x_4; } } lean_object* initialize_Init_SimpLemmas(uint8_t builtin, lean_object*); +lean_object* initialize_Init_Data_NeZero(uint8_t builtin, lean_object*); static bool _G_initialized = false; LEAN_EXPORT lean_object* initialize_Init_Data_Nat_Basic(uint8_t builtin, lean_object* w) { lean_object * res; @@ -1028,6 +1029,9 @@ _G_initialized = true; res = initialize_Init_SimpLemmas(builtin, lean_io_mk_world()); if (lean_io_result_is_error(res)) return res; lean_dec_ref(res); +res = initialize_Init_Data_NeZero(builtin, lean_io_mk_world()); +if (lean_io_result_is_error(res)) return res; +lean_dec_ref(res); l_Nat_instTransLt = _init_l_Nat_instTransLt(); l_Nat_instTransLe = _init_l_Nat_instTransLe(); l_Nat_instTransLtLe = _init_l_Nat_instTransLtLe(); diff --git a/stage0/stdlib/Init/Data/NeZero.c b/stage0/stdlib/Init/Data/NeZero.c new file mode 100644 index 000000000000..927a11258f7a --- /dev/null +++ b/stage0/stdlib/Init/Data/NeZero.c @@ -0,0 +1,29 @@ +// Lean compiler output +// Module: Init.Data.NeZero +// Imports: Init.Data.Zero +#include +#if defined(__clang__) +#pragma clang diagnostic ignored "-Wunused-parameter" +#pragma clang diagnostic ignored "-Wunused-label" +#elif defined(__GNUC__) && !defined(__CLANG__) +#pragma GCC diagnostic ignored "-Wunused-parameter" +#pragma GCC diagnostic ignored "-Wunused-label" +#pragma GCC diagnostic ignored "-Wunused-but-set-variable" +#endif +#ifdef __cplusplus +extern "C" { +#endif +lean_object* initialize_Init_Data_Zero(uint8_t builtin, lean_object*); +static bool _G_initialized = false; +LEAN_EXPORT lean_object* initialize_Init_Data_NeZero(uint8_t builtin, lean_object* w) { +lean_object * res; +if (_G_initialized) return lean_io_result_mk_ok(lean_box(0)); +_G_initialized = true; +res = initialize_Init_Data_Zero(builtin, lean_io_mk_world()); +if (lean_io_result_is_error(res)) return res; +lean_dec_ref(res); +return lean_io_result_mk_ok(lean_box(0)); +} +#ifdef __cplusplus +} +#endif diff --git a/stage0/stdlib/Init/Data/Zero.c b/stage0/stdlib/Init/Data/Zero.c new file mode 100644 index 000000000000..72e97e19dc85 --- /dev/null +++ b/stage0/stdlib/Init/Data/Zero.c @@ -0,0 +1,83 @@ +// Lean compiler output +// Module: Init.Data.Zero +// Imports: Init.Core +#include +#if defined(__clang__) +#pragma clang diagnostic ignored "-Wunused-parameter" +#pragma clang diagnostic ignored "-Wunused-label" +#elif defined(__GNUC__) && !defined(__CLANG__) +#pragma GCC diagnostic ignored "-Wunused-parameter" +#pragma GCC diagnostic ignored "-Wunused-label" +#pragma GCC diagnostic ignored "-Wunused-but-set-variable" +#endif +#ifdef __cplusplus +extern "C" { +#endif +LEAN_EXPORT lean_object* l_Zero_ofOfNat0___rarg(lean_object*); +LEAN_EXPORT lean_object* l_Zero_ofOfNat0(lean_object*); +LEAN_EXPORT lean_object* l_Zero_toOfNat0___rarg(lean_object*); +LEAN_EXPORT lean_object* l_Zero_toOfNat0___rarg___boxed(lean_object*); +LEAN_EXPORT lean_object* l_Zero_toOfNat0(lean_object*); +LEAN_EXPORT lean_object* l_Zero_ofOfNat0___rarg___boxed(lean_object*); +LEAN_EXPORT lean_object* l_Zero_toOfNat0___rarg(lean_object* x_1) { +_start: +{ +lean_inc(x_1); +return x_1; +} +} +LEAN_EXPORT lean_object* l_Zero_toOfNat0(lean_object* x_1) { +_start: +{ +lean_object* x_2; +x_2 = lean_alloc_closure((void*)(l_Zero_toOfNat0___rarg___boxed), 1, 0); +return x_2; +} +} +LEAN_EXPORT lean_object* l_Zero_toOfNat0___rarg___boxed(lean_object* x_1) { +_start: +{ +lean_object* x_2; +x_2 = l_Zero_toOfNat0___rarg(x_1); +lean_dec(x_1); +return x_2; +} +} +LEAN_EXPORT lean_object* l_Zero_ofOfNat0___rarg(lean_object* x_1) { +_start: +{ +lean_inc(x_1); +return x_1; +} +} +LEAN_EXPORT lean_object* l_Zero_ofOfNat0(lean_object* x_1) { +_start: +{ +lean_object* x_2; +x_2 = lean_alloc_closure((void*)(l_Zero_ofOfNat0___rarg___boxed), 1, 0); +return x_2; +} +} +LEAN_EXPORT lean_object* l_Zero_ofOfNat0___rarg___boxed(lean_object* x_1) { +_start: +{ +lean_object* x_2; +x_2 = l_Zero_ofOfNat0___rarg(x_1); +lean_dec(x_1); +return x_2; +} +} +lean_object* initialize_Init_Core(uint8_t builtin, lean_object*); +static bool _G_initialized = false; +LEAN_EXPORT lean_object* initialize_Init_Data_Zero(uint8_t builtin, lean_object* w) { +lean_object * res; +if (_G_initialized) return lean_io_result_mk_ok(lean_box(0)); +_G_initialized = true; +res = initialize_Init_Core(builtin, lean_io_mk_world()); +if (lean_io_result_is_error(res)) return res; +lean_dec_ref(res); +return lean_io_result_mk_ok(lean_box(0)); +} +#ifdef __cplusplus +} +#endif diff --git a/stage0/stdlib/Lean/Meta/LitValues.c b/stage0/stdlib/Lean/Meta/LitValues.c index 345cc3b0e222..d9408920e488 100644 --- a/stage0/stdlib/Lean/Meta/LitValues.c +++ b/stage0/stdlib/Lean/Meta/LitValues.c @@ -60,10 +60,12 @@ LEAN_EXPORT lean_object* l_Lean_Meta_getCharValue_x3f___lambda__1___boxed(lean_o LEAN_EXPORT lean_object* l_Lean_Meta_normLitValue___lambda__1(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); static lean_object* l_Lean_Meta_getBitVecValue_x3f___closed__2; static lean_object* l_Lean_Loop_forIn_loop___at_Lean_Meta_getListLitOf_x3f___spec__1___rarg___closed__1; +static lean_object* l_Lean_Meta_normLitValue___lambda__9___closed__5; static lean_object* l_Lean_Meta_normLitValue___lambda__2___closed__2; LEAN_EXPORT lean_object* l_Lean_Meta_getUInt32Value_x3f(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); static lean_object* l_Lean_Meta_litToCtor___lambda__2___closed__1; LEAN_EXPORT lean_object* l_Lean_Meta_normLitValue(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); +static lean_object* l_Lean_Meta_normLitValue___lambda__9___closed__6; static lean_object* l_Lean_Meta_litToCtor___lambda__1___closed__25; static lean_object* l_Lean_Meta_normLitValue___lambda__2___closed__6; static lean_object* l_Lean_Meta_litToCtor___lambda__1___closed__9; @@ -186,6 +188,7 @@ LEAN_EXPORT lean_object* l_Lean_Meta_getIntValue_x3f___lambda__1___boxed(lean_ob static lean_object* l_Lean_Meta_normLitValue___lambda__9___closed__3; LEAN_EXPORT lean_object* l_Lean_Loop_forIn_loop___at_Lean_Meta_getListLitOf_x3f___spec__1(lean_object*); LEAN_EXPORT lean_object* l_Lean_Meta_isLitValue___lambda__1(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); +static lean_object* l_Lean_Meta_normLitValue___lambda__9___closed__4; static lean_object* l_Lean_Meta_litToCtor___lambda__1___closed__17; lean_object* l_Lean_Expr_app___override(lean_object*, lean_object*); static lean_object* l_Lean_Meta_normLitValue___lambda__9___closed__1; @@ -4439,6 +4442,34 @@ x_3 = l_Lean_Expr_const___override(x_2, x_1); return x_3; } } +static lean_object* _init_l_Lean_Meta_normLitValue___lambda__9___closed__4() { +_start: +{ +lean_object* x_1; +x_1 = lean_mk_string_unchecked("instNeZeroSucc", 14, 14); +return x_1; +} +} +static lean_object* _init_l_Lean_Meta_normLitValue___lambda__9___closed__5() { +_start: +{ +lean_object* x_1; lean_object* x_2; lean_object* x_3; +x_1 = l_Lean_Meta_getNatValue_x3f___lambda__1___closed__1; +x_2 = l_Lean_Meta_normLitValue___lambda__9___closed__4; +x_3 = l_Lean_Name_mkStr2(x_1, x_2); +return x_3; +} +} +static lean_object* _init_l_Lean_Meta_normLitValue___lambda__9___closed__6() { +_start: +{ +lean_object* x_1; lean_object* x_2; lean_object* x_3; +x_1 = lean_box(0); +x_2 = l_Lean_Meta_normLitValue___lambda__9___closed__5; +x_3 = l_Lean_Expr_const___override(x_2, x_1); +return x_3; +} +} LEAN_EXPORT lean_object* l_Lean_Meta_normLitValue___lambda__9(lean_object* x_1, lean_object* x_2, lean_object* x_3, lean_object* x_4, lean_object* x_5, lean_object* x_6, lean_object* x_7) { _start: { @@ -4478,7 +4509,7 @@ lean_dec(x_9); x_14 = !lean_is_exclusive(x_8); if (x_14 == 0) { -lean_object* x_15; lean_object* x_16; lean_object* x_17; lean_object* x_18; lean_object* x_19; lean_object* x_20; lean_object* x_21; lean_object* x_22; lean_object* x_23; lean_object* x_24; lean_object* x_25; lean_object* x_26; lean_object* x_27; lean_object* x_28; +lean_object* x_15; lean_object* x_16; lean_object* x_17; lean_object* x_18; lean_object* x_19; lean_object* x_20; lean_object* x_21; lean_object* x_22; lean_object* x_23; lean_object* x_24; lean_object* x_25; lean_object* x_26; lean_object* x_27; lean_object* x_28; lean_object* x_29; lean_object* x_30; x_15 = lean_ctor_get(x_8, 0); lean_dec(x_15); x_16 = lean_ctor_get(x_13, 0); @@ -4490,76 +4521,82 @@ x_18 = l_Lean_mkRawNatLit(x_17); lean_inc(x_16); x_19 = l_Lean_mkNatLit(x_16); x_20 = l_Lean_Meta_normLitValue___lambda__9___closed__1; +lean_inc(x_19); x_21 = l_Lean_Expr_app___override(x_20, x_19); x_22 = lean_unsigned_to_nat(1u); x_23 = lean_nat_sub(x_16, x_22); lean_dec(x_16); x_24 = l_Lean_mkNatLit(x_23); -x_25 = l_Lean_Meta_normLitValue___lambda__9___closed__3; +x_25 = l_Lean_Meta_normLitValue___lambda__9___closed__6; +x_26 = l_Lean_Expr_app___override(x_25, x_24); +x_27 = l_Lean_Meta_normLitValue___lambda__9___closed__3; lean_inc(x_18); -x_26 = l_Lean_mkAppB(x_25, x_24, x_18); -x_27 = l_Lean_Meta_normLitValue___lambda__2___closed__3; -x_28 = l_Lean_mkApp3(x_27, x_21, x_18, x_26); -lean_ctor_set(x_8, 0, x_28); +x_28 = l_Lean_mkApp3(x_27, x_19, x_26, x_18); +x_29 = l_Lean_Meta_normLitValue___lambda__2___closed__3; +x_30 = l_Lean_mkApp3(x_29, x_21, x_18, x_28); +lean_ctor_set(x_8, 0, x_30); return x_8; } else { -lean_object* x_29; lean_object* x_30; lean_object* x_31; lean_object* x_32; lean_object* x_33; lean_object* x_34; lean_object* x_35; lean_object* x_36; lean_object* x_37; lean_object* x_38; lean_object* x_39; lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; -x_29 = lean_ctor_get(x_8, 1); -lean_inc(x_29); -lean_dec(x_8); -x_30 = lean_ctor_get(x_13, 0); -lean_inc(x_30); -x_31 = lean_ctor_get(x_13, 1); +lean_object* x_31; lean_object* x_32; lean_object* x_33; lean_object* x_34; lean_object* x_35; lean_object* x_36; lean_object* x_37; lean_object* x_38; lean_object* x_39; lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; +x_31 = lean_ctor_get(x_8, 1); lean_inc(x_31); +lean_dec(x_8); +x_32 = lean_ctor_get(x_13, 0); +lean_inc(x_32); +x_33 = lean_ctor_get(x_13, 1); +lean_inc(x_33); lean_dec(x_13); -x_32 = l_Lean_mkRawNatLit(x_31); -lean_inc(x_30); -x_33 = l_Lean_mkNatLit(x_30); -x_34 = l_Lean_Meta_normLitValue___lambda__9___closed__1; -x_35 = l_Lean_Expr_app___override(x_34, x_33); -x_36 = lean_unsigned_to_nat(1u); -x_37 = lean_nat_sub(x_30, x_36); -lean_dec(x_30); -x_38 = l_Lean_mkNatLit(x_37); -x_39 = l_Lean_Meta_normLitValue___lambda__9___closed__3; +x_34 = l_Lean_mkRawNatLit(x_33); lean_inc(x_32); -x_40 = l_Lean_mkAppB(x_39, x_38, x_32); -x_41 = l_Lean_Meta_normLitValue___lambda__2___closed__3; -x_42 = l_Lean_mkApp3(x_41, x_35, x_32, x_40); -x_43 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_43, 0, x_42); -lean_ctor_set(x_43, 1, x_29); -return x_43; +x_35 = l_Lean_mkNatLit(x_32); +x_36 = l_Lean_Meta_normLitValue___lambda__9___closed__1; +lean_inc(x_35); +x_37 = l_Lean_Expr_app___override(x_36, x_35); +x_38 = lean_unsigned_to_nat(1u); +x_39 = lean_nat_sub(x_32, x_38); +lean_dec(x_32); +x_40 = l_Lean_mkNatLit(x_39); +x_41 = l_Lean_Meta_normLitValue___lambda__9___closed__6; +x_42 = l_Lean_Expr_app___override(x_41, x_40); +x_43 = l_Lean_Meta_normLitValue___lambda__9___closed__3; +lean_inc(x_34); +x_44 = l_Lean_mkApp3(x_43, x_35, x_42, x_34); +x_45 = l_Lean_Meta_normLitValue___lambda__2___closed__3; +x_46 = l_Lean_mkApp3(x_45, x_37, x_34, x_44); +x_47 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_47, 0, x_46); +lean_ctor_set(x_47, 1, x_31); +return x_47; } } } else { -uint8_t x_44; +uint8_t x_48; lean_dec(x_6); lean_dec(x_5); lean_dec(x_4); lean_dec(x_3); lean_dec(x_1); -x_44 = !lean_is_exclusive(x_8); -if (x_44 == 0) +x_48 = !lean_is_exclusive(x_8); +if (x_48 == 0) { return x_8; } else { -lean_object* x_45; lean_object* x_46; lean_object* x_47; -x_45 = lean_ctor_get(x_8, 0); -x_46 = lean_ctor_get(x_8, 1); -lean_inc(x_46); -lean_inc(x_45); +lean_object* x_49; lean_object* x_50; lean_object* x_51; +x_49 = lean_ctor_get(x_8, 0); +x_50 = lean_ctor_get(x_8, 1); +lean_inc(x_50); +lean_inc(x_49); lean_dec(x_8); -x_47 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_47, 0, x_45); -lean_ctor_set(x_47, 1, x_46); -return x_47; +x_51 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_51, 0, x_49); +lean_ctor_set(x_51, 1, x_50); +return x_51; } } } @@ -7943,6 +7980,12 @@ l_Lean_Meta_normLitValue___lambda__9___closed__2 = _init_l_Lean_Meta_normLitValu lean_mark_persistent(l_Lean_Meta_normLitValue___lambda__9___closed__2); l_Lean_Meta_normLitValue___lambda__9___closed__3 = _init_l_Lean_Meta_normLitValue___lambda__9___closed__3(); lean_mark_persistent(l_Lean_Meta_normLitValue___lambda__9___closed__3); +l_Lean_Meta_normLitValue___lambda__9___closed__4 = _init_l_Lean_Meta_normLitValue___lambda__9___closed__4(); +lean_mark_persistent(l_Lean_Meta_normLitValue___lambda__9___closed__4); +l_Lean_Meta_normLitValue___lambda__9___closed__5 = _init_l_Lean_Meta_normLitValue___lambda__9___closed__5(); +lean_mark_persistent(l_Lean_Meta_normLitValue___lambda__9___closed__5); +l_Lean_Meta_normLitValue___lambda__9___closed__6 = _init_l_Lean_Meta_normLitValue___lambda__9___closed__6(); +lean_mark_persistent(l_Lean_Meta_normLitValue___lambda__9___closed__6); l_Lean_Meta_normLitValue___lambda__10___closed__1 = _init_l_Lean_Meta_normLitValue___lambda__10___closed__1(); lean_mark_persistent(l_Lean_Meta_normLitValue___lambda__10___closed__1); l_Lean_Meta_normLitValue___lambda__10___closed__2 = _init_l_Lean_Meta_normLitValue___lambda__10___closed__2(); diff --git a/stage0/stdlib/Lean/Meta/Match/Match.c b/stage0/stdlib/Lean/Meta/Match/Match.c index 00cd8df4e62e..c8b6fbb30f50 100644 --- a/stage0/stdlib/Lean/Meta/Match/Match.c +++ b/stage0/stdlib/Lean/Meta/Match/Match.c @@ -110,7 +110,6 @@ LEAN_EXPORT uint8_t l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_inLocal uint8_t l_Lean_Exception_isInterrupt(lean_object*); LEAN_EXPORT lean_object* l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_processVariable___spec__2___lambda__1(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); uint64_t lean_uint64_of_nat(lean_object*); -lean_object* l_Lean_mkAppB(lean_object*, lean_object*, lean_object*); uint64_t lean_uint64_mix_hash(uint64_t, uint64_t); static lean_object* l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_traceStep___closed__2; LEAN_EXPORT lean_object* l_Array_foldrMUnsafe_fold___at_Lean_Meta_Match_withCleanLCtxFor___spec__23___boxed(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); @@ -123,6 +122,7 @@ lean_object* l_Lean_mkDefinitionValInferrringUnsafe___at_Lean_Meta_mkAuxDefiniti LEAN_EXPORT lean_object* l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_throwCasesException___rarg___boxed(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); size_t lean_uint64_to_usize(uint64_t); static lean_object* l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_isConstructorTransition___closed__1; +static lean_object* l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__15; static lean_object* l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_withAlts_loop___rarg___closed__10; static lean_object* l_List_filterMapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_processNonVariable___spec__2___closed__1; LEAN_EXPORT lean_object* l_Lean_Meta_Match_mkMatcherAuxDefinition___lambda__2___boxed(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); @@ -505,6 +505,7 @@ LEAN_EXPORT lean_object* l_Lean_Meta_Match_mkMatcherAuxDefinition___lambda__3(le LEAN_EXPORT lean_object* l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_checkNumPatterns(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); static lean_object* l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_processValue___lambda__1___closed__1; lean_object* l_EStateM_pure___rarg(lean_object*, lean_object*); +static lean_object* l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__16; static lean_object* l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_throwCasesException___rarg___closed__16; LEAN_EXPORT lean_object* l_Array_mapMUnsafe_map___at_Lean_Meta_Match_getMkMatcherInputInContext___spec__5___boxed(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); LEAN_EXPORT lean_object* l_Lean_Meta_Match_mkMatcherAuxDefinition___lambda__3___boxed(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); @@ -576,6 +577,7 @@ LEAN_EXPORT lean_object* l_Lean_Meta_Match_mkMatcher___lambda__13___boxed(lean_o static lean_object* l_Lean_Meta_Match_getMkMatcherInputInContext___closed__2; LEAN_EXPORT lean_object* l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandVarIntoArrayLit_loop___lambda__1(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); LEAN_EXPORT lean_object* l_List_mapTR_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_reorientCnstrs___spec__1(lean_object*, lean_object*, lean_object*); +static lean_object* l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__14; static lean_object* l_Lean_Meta_Match_initFn____x40_Lean_Meta_Match_Match___hyg_10721____closed__2; LEAN_EXPORT lean_object* l_Array_forInUnsafe_loop___at_Lean_Meta_Match_mkMatcher___spec__12(lean_object*, lean_object*, size_t, size_t, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); LEAN_EXPORT lean_object* l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_solveCnstrs_go___lambda__3___boxed(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); @@ -27291,6 +27293,34 @@ x_3 = l_Lean_Expr_const___override(x_2, x_1); return x_3; } } +static lean_object* _init_l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__14() { +_start: +{ +lean_object* x_1; +x_1 = lean_mk_string_unchecked("instNeZeroSucc", 14, 14); +return x_1; +} +} +static lean_object* _init_l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__15() { +_start: +{ +lean_object* x_1; lean_object* x_2; lean_object* x_3; +x_1 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandNatValuePattern___spec__1___closed__1; +x_2 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__14; +x_3 = l_Lean_Name_mkStr2(x_1, x_2); +return x_3; +} +} +static lean_object* _init_l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__16() { +_start: +{ +lean_object* x_1; lean_object* x_2; lean_object* x_3; +x_1 = lean_box(0); +x_2 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__15; +x_3 = l_Lean_Expr_const___override(x_2, x_1); +return x_3; +} +} LEAN_EXPORT lean_object* l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1(lean_object* x_1, lean_object* x_2, lean_object* x_3, lean_object* x_4, lean_object* x_5, lean_object* x_6, lean_object* x_7) { _start: { @@ -27443,7 +27473,7 @@ lean_dec(x_33); x_46 = !lean_is_exclusive(x_44); if (x_46 == 0) { -lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; lean_object* x_54; lean_object* x_55; lean_object* x_56; lean_object* x_57; lean_object* x_58; lean_object* x_59; lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; +lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; lean_object* x_54; lean_object* x_55; lean_object* x_56; lean_object* x_57; lean_object* x_58; lean_object* x_59; lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; x_47 = lean_ctor_get(x_44, 0); x_48 = lean_ctor_get(x_44, 1); x_49 = lean_box(0); @@ -27454,26 +27484,29 @@ x_52 = l_Lean_mkRawNatLit(x_48); lean_inc(x_51); x_53 = l_Lean_mkNatLit(x_51); x_54 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__10; +lean_inc(x_53); x_55 = l_Lean_Expr_app___override(x_54, x_53); x_56 = lean_unsigned_to_nat(1u); x_57 = lean_nat_sub(x_51, x_56); lean_dec(x_51); x_58 = l_Lean_mkNatLit(x_57); -x_59 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; +x_59 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__16; +x_60 = l_Lean_Expr_app___override(x_59, x_58); +x_61 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; lean_inc(x_52); -x_60 = l_Lean_mkAppB(x_59, x_58, x_52); -x_61 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; -x_62 = l_Lean_mkApp3(x_61, x_55, x_52, x_60); -lean_ctor_set(x_19, 0, x_62); +x_62 = l_Lean_mkApp3(x_61, x_53, x_60, x_52); +x_63 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; +x_64 = l_Lean_mkApp3(x_63, x_55, x_52, x_62); +lean_ctor_set(x_19, 0, x_64); lean_ctor_set(x_11, 1, x_49); -x_63 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; -x_64 = lean_alloc_ctor(2, 4, 0); -lean_ctor_set(x_64, 0, x_63); -lean_ctor_set(x_64, 1, x_49); -lean_ctor_set(x_64, 2, x_49); -lean_ctor_set(x_64, 3, x_11); +x_65 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; +x_66 = lean_alloc_ctor(2, 4, 0); +lean_ctor_set(x_66, 0, x_65); +lean_ctor_set(x_66, 1, x_49); +lean_ctor_set(x_66, 2, x_49); +lean_ctor_set(x_66, 3, x_11); lean_ctor_set(x_1, 1, x_29); -lean_ctor_set(x_1, 0, x_64); +lean_ctor_set(x_1, 0, x_66); lean_ctor_set(x_10, 4, x_1); lean_ctor_set_tag(x_44, 1); lean_ctor_set(x_44, 1, x_2); @@ -27485,125 +27518,131 @@ goto _start; } else { -lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; lean_object* x_78; lean_object* x_79; lean_object* x_80; lean_object* x_81; lean_object* x_82; lean_object* x_83; lean_object* x_84; -x_66 = lean_ctor_get(x_44, 0); -x_67 = lean_ctor_get(x_44, 1); -lean_inc(x_67); -lean_inc(x_66); +lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; lean_object* x_78; lean_object* x_79; lean_object* x_80; lean_object* x_81; lean_object* x_82; lean_object* x_83; lean_object* x_84; lean_object* x_85; lean_object* x_86; lean_object* x_87; lean_object* x_88; +x_68 = lean_ctor_get(x_44, 0); +x_69 = lean_ctor_get(x_44, 1); +lean_inc(x_69); +lean_inc(x_68); lean_dec(x_44); -x_68 = lean_box(0); -x_69 = lean_unsigned_to_nat(2u); -x_70 = lean_nat_pow(x_69, x_66); -lean_dec(x_66); -x_71 = l_Lean_mkRawNatLit(x_67); -lean_inc(x_70); -x_72 = l_Lean_mkNatLit(x_70); -x_73 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__10; -x_74 = l_Lean_Expr_app___override(x_73, x_72); -x_75 = lean_unsigned_to_nat(1u); -x_76 = lean_nat_sub(x_70, x_75); -lean_dec(x_70); -x_77 = l_Lean_mkNatLit(x_76); -x_78 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; -lean_inc(x_71); -x_79 = l_Lean_mkAppB(x_78, x_77, x_71); -x_80 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; -x_81 = l_Lean_mkApp3(x_80, x_74, x_71, x_79); -lean_ctor_set(x_19, 0, x_81); -lean_ctor_set(x_11, 1, x_68); -x_82 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; -x_83 = lean_alloc_ctor(2, 4, 0); -lean_ctor_set(x_83, 0, x_82); -lean_ctor_set(x_83, 1, x_68); -lean_ctor_set(x_83, 2, x_68); -lean_ctor_set(x_83, 3, x_11); +x_70 = lean_box(0); +x_71 = lean_unsigned_to_nat(2u); +x_72 = lean_nat_pow(x_71, x_68); +lean_dec(x_68); +x_73 = l_Lean_mkRawNatLit(x_69); +lean_inc(x_72); +x_74 = l_Lean_mkNatLit(x_72); +x_75 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__10; +lean_inc(x_74); +x_76 = l_Lean_Expr_app___override(x_75, x_74); +x_77 = lean_unsigned_to_nat(1u); +x_78 = lean_nat_sub(x_72, x_77); +lean_dec(x_72); +x_79 = l_Lean_mkNatLit(x_78); +x_80 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__16; +x_81 = l_Lean_Expr_app___override(x_80, x_79); +x_82 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; +lean_inc(x_73); +x_83 = l_Lean_mkApp3(x_82, x_74, x_81, x_73); +x_84 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; +x_85 = l_Lean_mkApp3(x_84, x_76, x_73, x_83); +lean_ctor_set(x_19, 0, x_85); +lean_ctor_set(x_11, 1, x_70); +x_86 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; +x_87 = lean_alloc_ctor(2, 4, 0); +lean_ctor_set(x_87, 0, x_86); +lean_ctor_set(x_87, 1, x_70); +lean_ctor_set(x_87, 2, x_70); +lean_ctor_set(x_87, 3, x_11); lean_ctor_set(x_1, 1, x_29); -lean_ctor_set(x_1, 0, x_83); +lean_ctor_set(x_1, 0, x_87); lean_ctor_set(x_10, 4, x_1); -x_84 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_84, 0, x_10); -lean_ctor_set(x_84, 1, x_2); +x_88 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_88, 0, x_10); +lean_ctor_set(x_88, 1, x_2); x_1 = x_21; -x_2 = x_84; +x_2 = x_88; x_7 = x_45; goto _start; } } else { -lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; lean_object* x_98; lean_object* x_99; lean_object* x_100; lean_object* x_101; lean_object* x_102; lean_object* x_103; lean_object* x_104; lean_object* x_105; lean_object* x_106; lean_object* x_107; lean_object* x_108; +lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; lean_object* x_98; lean_object* x_99; lean_object* x_100; lean_object* x_101; lean_object* x_102; lean_object* x_103; lean_object* x_104; lean_object* x_105; lean_object* x_106; lean_object* x_107; lean_object* x_108; lean_object* x_109; lean_object* x_110; lean_object* x_111; lean_object* x_112; lean_object* x_113; lean_object* x_114; lean_dec(x_10); -x_86 = lean_ctor_get(x_34, 0); -lean_inc(x_86); +x_90 = lean_ctor_get(x_34, 0); +lean_inc(x_90); lean_dec(x_34); -x_87 = lean_ctor_get(x_33, 1); -lean_inc(x_87); +x_91 = lean_ctor_get(x_33, 1); +lean_inc(x_91); lean_dec(x_33); -x_88 = lean_ctor_get(x_86, 0); -lean_inc(x_88); -x_89 = lean_ctor_get(x_86, 1); -lean_inc(x_89); -if (lean_is_exclusive(x_86)) { - lean_ctor_release(x_86, 0); - lean_ctor_release(x_86, 1); - x_90 = x_86; +x_92 = lean_ctor_get(x_90, 0); +lean_inc(x_92); +x_93 = lean_ctor_get(x_90, 1); +lean_inc(x_93); +if (lean_is_exclusive(x_90)) { + lean_ctor_release(x_90, 0); + lean_ctor_release(x_90, 1); + x_94 = x_90; } else { - lean_dec_ref(x_86); - x_90 = lean_box(0); + lean_dec_ref(x_90); + x_94 = lean_box(0); } -x_91 = lean_box(0); -x_92 = lean_unsigned_to_nat(2u); -x_93 = lean_nat_pow(x_92, x_88); -lean_dec(x_88); -x_94 = l_Lean_mkRawNatLit(x_89); -lean_inc(x_93); -x_95 = l_Lean_mkNatLit(x_93); -x_96 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__10; -x_97 = l_Lean_Expr_app___override(x_96, x_95); -x_98 = lean_unsigned_to_nat(1u); -x_99 = lean_nat_sub(x_93, x_98); -lean_dec(x_93); -x_100 = l_Lean_mkNatLit(x_99); -x_101 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; -lean_inc(x_94); -x_102 = l_Lean_mkAppB(x_101, x_100, x_94); -x_103 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; -x_104 = l_Lean_mkApp3(x_103, x_97, x_94, x_102); -lean_ctor_set(x_19, 0, x_104); -lean_ctor_set(x_11, 1, x_91); -x_105 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; -x_106 = lean_alloc_ctor(2, 4, 0); -lean_ctor_set(x_106, 0, x_105); -lean_ctor_set(x_106, 1, x_91); -lean_ctor_set(x_106, 2, x_91); -lean_ctor_set(x_106, 3, x_11); +x_95 = lean_box(0); +x_96 = lean_unsigned_to_nat(2u); +x_97 = lean_nat_pow(x_96, x_92); +lean_dec(x_92); +x_98 = l_Lean_mkRawNatLit(x_93); +lean_inc(x_97); +x_99 = l_Lean_mkNatLit(x_97); +x_100 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__10; +lean_inc(x_99); +x_101 = l_Lean_Expr_app___override(x_100, x_99); +x_102 = lean_unsigned_to_nat(1u); +x_103 = lean_nat_sub(x_97, x_102); +lean_dec(x_97); +x_104 = l_Lean_mkNatLit(x_103); +x_105 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__16; +x_106 = l_Lean_Expr_app___override(x_105, x_104); +x_107 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; +lean_inc(x_98); +x_108 = l_Lean_mkApp3(x_107, x_99, x_106, x_98); +x_109 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; +x_110 = l_Lean_mkApp3(x_109, x_101, x_98, x_108); +lean_ctor_set(x_19, 0, x_110); +lean_ctor_set(x_11, 1, x_95); +x_111 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; +x_112 = lean_alloc_ctor(2, 4, 0); +lean_ctor_set(x_112, 0, x_111); +lean_ctor_set(x_112, 1, x_95); +lean_ctor_set(x_112, 2, x_95); +lean_ctor_set(x_112, 3, x_11); lean_ctor_set(x_1, 1, x_29); -lean_ctor_set(x_1, 0, x_106); -x_107 = lean_alloc_ctor(0, 6, 0); -lean_ctor_set(x_107, 0, x_23); -lean_ctor_set(x_107, 1, x_24); -lean_ctor_set(x_107, 2, x_25); -lean_ctor_set(x_107, 3, x_26); -lean_ctor_set(x_107, 4, x_1); -lean_ctor_set(x_107, 5, x_27); -if (lean_is_scalar(x_90)) { - x_108 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_1, 0, x_112); +x_113 = lean_alloc_ctor(0, 6, 0); +lean_ctor_set(x_113, 0, x_23); +lean_ctor_set(x_113, 1, x_24); +lean_ctor_set(x_113, 2, x_25); +lean_ctor_set(x_113, 3, x_26); +lean_ctor_set(x_113, 4, x_1); +lean_ctor_set(x_113, 5, x_27); +if (lean_is_scalar(x_94)) { + x_114 = lean_alloc_ctor(1, 2, 0); } else { - x_108 = x_90; - lean_ctor_set_tag(x_108, 1); + x_114 = x_94; + lean_ctor_set_tag(x_114, 1); } -lean_ctor_set(x_108, 0, x_107); -lean_ctor_set(x_108, 1, x_2); +lean_ctor_set(x_114, 0, x_113); +lean_ctor_set(x_114, 1, x_2); x_1 = x_21; -x_2 = x_108; -x_7 = x_87; +x_2 = x_114; +x_7 = x_91; goto _start; } } } else { -uint8_t x_110; +uint8_t x_116; lean_free_object(x_19); lean_free_object(x_11); lean_dec(x_29); @@ -27620,45 +27659,45 @@ lean_dec(x_5); lean_dec(x_4); lean_dec(x_3); lean_dec(x_2); -x_110 = !lean_is_exclusive(x_33); -if (x_110 == 0) +x_116 = !lean_is_exclusive(x_33); +if (x_116 == 0) { return x_33; } else { -lean_object* x_111; lean_object* x_112; lean_object* x_113; -x_111 = lean_ctor_get(x_33, 0); -x_112 = lean_ctor_get(x_33, 1); -lean_inc(x_112); -lean_inc(x_111); +lean_object* x_117; lean_object* x_118; lean_object* x_119; +x_117 = lean_ctor_get(x_33, 0); +x_118 = lean_ctor_get(x_33, 1); +lean_inc(x_118); +lean_inc(x_117); lean_dec(x_33); -x_113 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_113, 0, x_111); -lean_ctor_set(x_113, 1, x_112); -return x_113; +x_119 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_119, 0, x_117); +lean_ctor_set(x_119, 1, x_118); +return x_119; } } } else { -lean_object* x_114; lean_object* x_115; -x_114 = lean_ctor_get(x_19, 0); -lean_inc(x_114); +lean_object* x_120; lean_object* x_121; +x_120 = lean_ctor_get(x_19, 0); +lean_inc(x_120); lean_dec(x_19); lean_inc(x_6); lean_inc(x_5); lean_inc(x_4); lean_inc(x_3); -x_115 = l_Lean_Meta_getBitVecValue_x3f(x_114, x_3, x_4, x_5, x_6, x_7); -if (lean_obj_tag(x_115) == 0) +x_121 = l_Lean_Meta_getBitVecValue_x3f(x_120, x_3, x_4, x_5, x_6, x_7); +if (lean_obj_tag(x_121) == 0) { -lean_object* x_116; -x_116 = lean_ctor_get(x_115, 0); -lean_inc(x_116); -if (lean_obj_tag(x_116) == 0) +lean_object* x_122; +x_122 = lean_ctor_get(x_121, 0); +lean_inc(x_122); +if (lean_obj_tag(x_122) == 0) { -lean_object* x_117; +lean_object* x_123; lean_dec(x_29); lean_dec(x_27); lean_dec(x_26); @@ -27666,19 +27705,19 @@ lean_dec(x_25); lean_dec(x_24); lean_dec(x_23); lean_free_object(x_1); -x_117 = lean_ctor_get(x_115, 1); -lean_inc(x_117); -lean_dec(x_115); +x_123 = lean_ctor_get(x_121, 1); +lean_inc(x_123); +lean_dec(x_121); lean_ctor_set(x_11, 1, x_2); lean_ctor_set(x_11, 0, x_10); x_1 = x_21; x_2 = x_11; -x_7 = x_117; +x_7 = x_123; goto _start; } else { -lean_object* x_119; lean_object* x_120; lean_object* x_121; lean_object* x_122; lean_object* x_123; lean_object* x_124; lean_object* x_125; lean_object* x_126; lean_object* x_127; lean_object* x_128; lean_object* x_129; lean_object* x_130; lean_object* x_131; lean_object* x_132; lean_object* x_133; lean_object* x_134; lean_object* x_135; lean_object* x_136; lean_object* x_137; lean_object* x_138; lean_object* x_139; lean_object* x_140; lean_object* x_141; lean_object* x_142; lean_object* x_143; +lean_object* x_125; lean_object* x_126; lean_object* x_127; lean_object* x_128; lean_object* x_129; lean_object* x_130; lean_object* x_131; lean_object* x_132; lean_object* x_133; lean_object* x_134; lean_object* x_135; lean_object* x_136; lean_object* x_137; lean_object* x_138; lean_object* x_139; lean_object* x_140; lean_object* x_141; lean_object* x_142; lean_object* x_143; lean_object* x_144; lean_object* x_145; lean_object* x_146; lean_object* x_147; lean_object* x_148; lean_object* x_149; lean_object* x_150; lean_object* x_151; if (lean_is_exclusive(x_10)) { lean_ctor_release(x_10, 0); lean_ctor_release(x_10, 1); @@ -27686,87 +27725,90 @@ if (lean_is_exclusive(x_10)) { lean_ctor_release(x_10, 3); lean_ctor_release(x_10, 4); lean_ctor_release(x_10, 5); - x_119 = x_10; + x_125 = x_10; } else { lean_dec_ref(x_10); - x_119 = lean_box(0); -} -x_120 = lean_ctor_get(x_116, 0); -lean_inc(x_120); -lean_dec(x_116); -x_121 = lean_ctor_get(x_115, 1); -lean_inc(x_121); -lean_dec(x_115); -x_122 = lean_ctor_get(x_120, 0); -lean_inc(x_122); -x_123 = lean_ctor_get(x_120, 1); -lean_inc(x_123); -if (lean_is_exclusive(x_120)) { - lean_ctor_release(x_120, 0); - lean_ctor_release(x_120, 1); - x_124 = x_120; -} else { - lean_dec_ref(x_120); - x_124 = lean_box(0); + x_125 = lean_box(0); } -x_125 = lean_box(0); -x_126 = lean_unsigned_to_nat(2u); -x_127 = lean_nat_pow(x_126, x_122); +x_126 = lean_ctor_get(x_122, 0); +lean_inc(x_126); lean_dec(x_122); -x_128 = l_Lean_mkRawNatLit(x_123); +x_127 = lean_ctor_get(x_121, 1); lean_inc(x_127); -x_129 = l_Lean_mkNatLit(x_127); -x_130 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__10; -x_131 = l_Lean_Expr_app___override(x_130, x_129); -x_132 = lean_unsigned_to_nat(1u); -x_133 = lean_nat_sub(x_127, x_132); -lean_dec(x_127); -x_134 = l_Lean_mkNatLit(x_133); -x_135 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; +lean_dec(x_121); +x_128 = lean_ctor_get(x_126, 0); lean_inc(x_128); -x_136 = l_Lean_mkAppB(x_135, x_134, x_128); -x_137 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; -x_138 = l_Lean_mkApp3(x_137, x_131, x_128, x_136); -x_139 = lean_alloc_ctor(3, 1, 0); -lean_ctor_set(x_139, 0, x_138); -lean_ctor_set(x_11, 1, x_125); -lean_ctor_set(x_11, 0, x_139); -x_140 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; -x_141 = lean_alloc_ctor(2, 4, 0); -lean_ctor_set(x_141, 0, x_140); -lean_ctor_set(x_141, 1, x_125); -lean_ctor_set(x_141, 2, x_125); -lean_ctor_set(x_141, 3, x_11); +x_129 = lean_ctor_get(x_126, 1); +lean_inc(x_129); +if (lean_is_exclusive(x_126)) { + lean_ctor_release(x_126, 0); + lean_ctor_release(x_126, 1); + x_130 = x_126; +} else { + lean_dec_ref(x_126); + x_130 = lean_box(0); +} +x_131 = lean_box(0); +x_132 = lean_unsigned_to_nat(2u); +x_133 = lean_nat_pow(x_132, x_128); +lean_dec(x_128); +x_134 = l_Lean_mkRawNatLit(x_129); +lean_inc(x_133); +x_135 = l_Lean_mkNatLit(x_133); +x_136 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__10; +lean_inc(x_135); +x_137 = l_Lean_Expr_app___override(x_136, x_135); +x_138 = lean_unsigned_to_nat(1u); +x_139 = lean_nat_sub(x_133, x_138); +lean_dec(x_133); +x_140 = l_Lean_mkNatLit(x_139); +x_141 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__16; +x_142 = l_Lean_Expr_app___override(x_141, x_140); +x_143 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; +lean_inc(x_134); +x_144 = l_Lean_mkApp3(x_143, x_135, x_142, x_134); +x_145 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; +x_146 = l_Lean_mkApp3(x_145, x_137, x_134, x_144); +x_147 = lean_alloc_ctor(3, 1, 0); +lean_ctor_set(x_147, 0, x_146); +lean_ctor_set(x_11, 1, x_131); +lean_ctor_set(x_11, 0, x_147); +x_148 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; +x_149 = lean_alloc_ctor(2, 4, 0); +lean_ctor_set(x_149, 0, x_148); +lean_ctor_set(x_149, 1, x_131); +lean_ctor_set(x_149, 2, x_131); +lean_ctor_set(x_149, 3, x_11); lean_ctor_set(x_1, 1, x_29); -lean_ctor_set(x_1, 0, x_141); -if (lean_is_scalar(x_119)) { - x_142 = lean_alloc_ctor(0, 6, 0); +lean_ctor_set(x_1, 0, x_149); +if (lean_is_scalar(x_125)) { + x_150 = lean_alloc_ctor(0, 6, 0); } else { - x_142 = x_119; + x_150 = x_125; } -lean_ctor_set(x_142, 0, x_23); -lean_ctor_set(x_142, 1, x_24); -lean_ctor_set(x_142, 2, x_25); -lean_ctor_set(x_142, 3, x_26); -lean_ctor_set(x_142, 4, x_1); -lean_ctor_set(x_142, 5, x_27); -if (lean_is_scalar(x_124)) { - x_143 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_150, 0, x_23); +lean_ctor_set(x_150, 1, x_24); +lean_ctor_set(x_150, 2, x_25); +lean_ctor_set(x_150, 3, x_26); +lean_ctor_set(x_150, 4, x_1); +lean_ctor_set(x_150, 5, x_27); +if (lean_is_scalar(x_130)) { + x_151 = lean_alloc_ctor(1, 2, 0); } else { - x_143 = x_124; - lean_ctor_set_tag(x_143, 1); + x_151 = x_130; + lean_ctor_set_tag(x_151, 1); } -lean_ctor_set(x_143, 0, x_142); -lean_ctor_set(x_143, 1, x_2); +lean_ctor_set(x_151, 0, x_150); +lean_ctor_set(x_151, 1, x_2); x_1 = x_21; -x_2 = x_143; -x_7 = x_121; +x_2 = x_151; +x_7 = x_127; goto _start; } } else { -lean_object* x_145; lean_object* x_146; lean_object* x_147; lean_object* x_148; +lean_object* x_153; lean_object* x_154; lean_object* x_155; lean_object* x_156; lean_free_object(x_11); lean_dec(x_29); lean_dec(x_27); @@ -27782,79 +27824,79 @@ lean_dec(x_5); lean_dec(x_4); lean_dec(x_3); lean_dec(x_2); -x_145 = lean_ctor_get(x_115, 0); -lean_inc(x_145); -x_146 = lean_ctor_get(x_115, 1); -lean_inc(x_146); -if (lean_is_exclusive(x_115)) { - lean_ctor_release(x_115, 0); - lean_ctor_release(x_115, 1); - x_147 = x_115; +x_153 = lean_ctor_get(x_121, 0); +lean_inc(x_153); +x_154 = lean_ctor_get(x_121, 1); +lean_inc(x_154); +if (lean_is_exclusive(x_121)) { + lean_ctor_release(x_121, 0); + lean_ctor_release(x_121, 1); + x_155 = x_121; } else { - lean_dec_ref(x_115); - x_147 = lean_box(0); + lean_dec_ref(x_121); + x_155 = lean_box(0); } -if (lean_is_scalar(x_147)) { - x_148 = lean_alloc_ctor(1, 2, 0); +if (lean_is_scalar(x_155)) { + x_156 = lean_alloc_ctor(1, 2, 0); } else { - x_148 = x_147; + x_156 = x_155; } -lean_ctor_set(x_148, 0, x_145); -lean_ctor_set(x_148, 1, x_146); -return x_148; +lean_ctor_set(x_156, 0, x_153); +lean_ctor_set(x_156, 1, x_154); +return x_156; } } } else { -lean_object* x_149; lean_object* x_150; lean_object* x_151; lean_object* x_152; -x_149 = lean_ctor_get(x_11, 1); -lean_inc(x_149); +lean_object* x_157; lean_object* x_158; lean_object* x_159; lean_object* x_160; +x_157 = lean_ctor_get(x_11, 1); +lean_inc(x_157); lean_dec(x_11); -x_150 = lean_ctor_get(x_19, 0); -lean_inc(x_150); +x_158 = lean_ctor_get(x_19, 0); +lean_inc(x_158); if (lean_is_exclusive(x_19)) { lean_ctor_release(x_19, 0); - x_151 = x_19; + x_159 = x_19; } else { lean_dec_ref(x_19); - x_151 = lean_box(0); + x_159 = lean_box(0); } lean_inc(x_6); lean_inc(x_5); lean_inc(x_4); lean_inc(x_3); -x_152 = l_Lean_Meta_getBitVecValue_x3f(x_150, x_3, x_4, x_5, x_6, x_7); -if (lean_obj_tag(x_152) == 0) +x_160 = l_Lean_Meta_getBitVecValue_x3f(x_158, x_3, x_4, x_5, x_6, x_7); +if (lean_obj_tag(x_160) == 0) { -lean_object* x_153; -x_153 = lean_ctor_get(x_152, 0); -lean_inc(x_153); -if (lean_obj_tag(x_153) == 0) +lean_object* x_161; +x_161 = lean_ctor_get(x_160, 0); +lean_inc(x_161); +if (lean_obj_tag(x_161) == 0) { -lean_object* x_154; lean_object* x_155; -lean_dec(x_151); -lean_dec(x_149); +lean_object* x_162; lean_object* x_163; +lean_dec(x_159); +lean_dec(x_157); lean_dec(x_27); lean_dec(x_26); lean_dec(x_25); lean_dec(x_24); lean_dec(x_23); lean_free_object(x_1); -x_154 = lean_ctor_get(x_152, 1); -lean_inc(x_154); -lean_dec(x_152); -x_155 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_155, 0, x_10); -lean_ctor_set(x_155, 1, x_2); +x_162 = lean_ctor_get(x_160, 1); +lean_inc(x_162); +lean_dec(x_160); +x_163 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_163, 0, x_10); +lean_ctor_set(x_163, 1, x_2); x_1 = x_21; -x_2 = x_155; -x_7 = x_154; +x_2 = x_163; +x_7 = x_162; goto _start; } else { -lean_object* x_157; lean_object* x_158; lean_object* x_159; lean_object* x_160; lean_object* x_161; lean_object* x_162; lean_object* x_163; lean_object* x_164; lean_object* x_165; lean_object* x_166; lean_object* x_167; lean_object* x_168; lean_object* x_169; lean_object* x_170; lean_object* x_171; lean_object* x_172; lean_object* x_173; lean_object* x_174; lean_object* x_175; lean_object* x_176; lean_object* x_177; lean_object* x_178; lean_object* x_179; lean_object* x_180; lean_object* x_181; lean_object* x_182; +lean_object* x_165; lean_object* x_166; lean_object* x_167; lean_object* x_168; lean_object* x_169; lean_object* x_170; lean_object* x_171; lean_object* x_172; lean_object* x_173; lean_object* x_174; lean_object* x_175; lean_object* x_176; lean_object* x_177; lean_object* x_178; lean_object* x_179; lean_object* x_180; lean_object* x_181; lean_object* x_182; lean_object* x_183; lean_object* x_184; lean_object* x_185; lean_object* x_186; lean_object* x_187; lean_object* x_188; lean_object* x_189; lean_object* x_190; lean_object* x_191; lean_object* x_192; if (lean_is_exclusive(x_10)) { lean_ctor_release(x_10, 0); lean_ctor_release(x_10, 1); @@ -27862,94 +27904,97 @@ if (lean_is_exclusive(x_10)) { lean_ctor_release(x_10, 3); lean_ctor_release(x_10, 4); lean_ctor_release(x_10, 5); - x_157 = x_10; + x_165 = x_10; } else { lean_dec_ref(x_10); - x_157 = lean_box(0); + x_165 = lean_box(0); } -x_158 = lean_ctor_get(x_153, 0); -lean_inc(x_158); -lean_dec(x_153); -x_159 = lean_ctor_get(x_152, 1); -lean_inc(x_159); -lean_dec(x_152); -x_160 = lean_ctor_get(x_158, 0); -lean_inc(x_160); -x_161 = lean_ctor_get(x_158, 1); -lean_inc(x_161); -if (lean_is_exclusive(x_158)) { - lean_ctor_release(x_158, 0); - lean_ctor_release(x_158, 1); - x_162 = x_158; +x_166 = lean_ctor_get(x_161, 0); +lean_inc(x_166); +lean_dec(x_161); +x_167 = lean_ctor_get(x_160, 1); +lean_inc(x_167); +lean_dec(x_160); +x_168 = lean_ctor_get(x_166, 0); +lean_inc(x_168); +x_169 = lean_ctor_get(x_166, 1); +lean_inc(x_169); +if (lean_is_exclusive(x_166)) { + lean_ctor_release(x_166, 0); + lean_ctor_release(x_166, 1); + x_170 = x_166; } else { - lean_dec_ref(x_158); - x_162 = lean_box(0); + lean_dec_ref(x_166); + x_170 = lean_box(0); } -x_163 = lean_box(0); -x_164 = lean_unsigned_to_nat(2u); -x_165 = lean_nat_pow(x_164, x_160); -lean_dec(x_160); -x_166 = l_Lean_mkRawNatLit(x_161); -lean_inc(x_165); -x_167 = l_Lean_mkNatLit(x_165); -x_168 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__10; -x_169 = l_Lean_Expr_app___override(x_168, x_167); -x_170 = lean_unsigned_to_nat(1u); -x_171 = lean_nat_sub(x_165, x_170); -lean_dec(x_165); -x_172 = l_Lean_mkNatLit(x_171); -x_173 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; -lean_inc(x_166); -x_174 = l_Lean_mkAppB(x_173, x_172, x_166); -x_175 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; -x_176 = l_Lean_mkApp3(x_175, x_169, x_166, x_174); -if (lean_is_scalar(x_151)) { - x_177 = lean_alloc_ctor(3, 1, 0); +x_171 = lean_box(0); +x_172 = lean_unsigned_to_nat(2u); +x_173 = lean_nat_pow(x_172, x_168); +lean_dec(x_168); +x_174 = l_Lean_mkRawNatLit(x_169); +lean_inc(x_173); +x_175 = l_Lean_mkNatLit(x_173); +x_176 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__10; +lean_inc(x_175); +x_177 = l_Lean_Expr_app___override(x_176, x_175); +x_178 = lean_unsigned_to_nat(1u); +x_179 = lean_nat_sub(x_173, x_178); +lean_dec(x_173); +x_180 = l_Lean_mkNatLit(x_179); +x_181 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__16; +x_182 = l_Lean_Expr_app___override(x_181, x_180); +x_183 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; +lean_inc(x_174); +x_184 = l_Lean_mkApp3(x_183, x_175, x_182, x_174); +x_185 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; +x_186 = l_Lean_mkApp3(x_185, x_177, x_174, x_184); +if (lean_is_scalar(x_159)) { + x_187 = lean_alloc_ctor(3, 1, 0); } else { - x_177 = x_151; + x_187 = x_159; } -lean_ctor_set(x_177, 0, x_176); -x_178 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_178, 0, x_177); -lean_ctor_set(x_178, 1, x_163); -x_179 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; -x_180 = lean_alloc_ctor(2, 4, 0); -lean_ctor_set(x_180, 0, x_179); -lean_ctor_set(x_180, 1, x_163); -lean_ctor_set(x_180, 2, x_163); -lean_ctor_set(x_180, 3, x_178); -lean_ctor_set(x_1, 1, x_149); -lean_ctor_set(x_1, 0, x_180); -if (lean_is_scalar(x_157)) { - x_181 = lean_alloc_ctor(0, 6, 0); +lean_ctor_set(x_187, 0, x_186); +x_188 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_188, 0, x_187); +lean_ctor_set(x_188, 1, x_171); +x_189 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; +x_190 = lean_alloc_ctor(2, 4, 0); +lean_ctor_set(x_190, 0, x_189); +lean_ctor_set(x_190, 1, x_171); +lean_ctor_set(x_190, 2, x_171); +lean_ctor_set(x_190, 3, x_188); +lean_ctor_set(x_1, 1, x_157); +lean_ctor_set(x_1, 0, x_190); +if (lean_is_scalar(x_165)) { + x_191 = lean_alloc_ctor(0, 6, 0); } else { - x_181 = x_157; + x_191 = x_165; } -lean_ctor_set(x_181, 0, x_23); -lean_ctor_set(x_181, 1, x_24); -lean_ctor_set(x_181, 2, x_25); -lean_ctor_set(x_181, 3, x_26); -lean_ctor_set(x_181, 4, x_1); -lean_ctor_set(x_181, 5, x_27); -if (lean_is_scalar(x_162)) { - x_182 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_191, 0, x_23); +lean_ctor_set(x_191, 1, x_24); +lean_ctor_set(x_191, 2, x_25); +lean_ctor_set(x_191, 3, x_26); +lean_ctor_set(x_191, 4, x_1); +lean_ctor_set(x_191, 5, x_27); +if (lean_is_scalar(x_170)) { + x_192 = lean_alloc_ctor(1, 2, 0); } else { - x_182 = x_162; - lean_ctor_set_tag(x_182, 1); + x_192 = x_170; + lean_ctor_set_tag(x_192, 1); } -lean_ctor_set(x_182, 0, x_181); -lean_ctor_set(x_182, 1, x_2); +lean_ctor_set(x_192, 0, x_191); +lean_ctor_set(x_192, 1, x_2); x_1 = x_21; -x_2 = x_182; -x_7 = x_159; +x_2 = x_192; +x_7 = x_167; goto _start; } } else { -lean_object* x_184; lean_object* x_185; lean_object* x_186; lean_object* x_187; -lean_dec(x_151); -lean_dec(x_149); +lean_object* x_194; lean_object* x_195; lean_object* x_196; lean_object* x_197; +lean_dec(x_159); +lean_dec(x_157); lean_dec(x_27); lean_dec(x_26); lean_dec(x_25); @@ -27963,102 +28008,102 @@ lean_dec(x_5); lean_dec(x_4); lean_dec(x_3); lean_dec(x_2); -x_184 = lean_ctor_get(x_152, 0); -lean_inc(x_184); -x_185 = lean_ctor_get(x_152, 1); -lean_inc(x_185); -if (lean_is_exclusive(x_152)) { - lean_ctor_release(x_152, 0); - lean_ctor_release(x_152, 1); - x_186 = x_152; +x_194 = lean_ctor_get(x_160, 0); +lean_inc(x_194); +x_195 = lean_ctor_get(x_160, 1); +lean_inc(x_195); +if (lean_is_exclusive(x_160)) { + lean_ctor_release(x_160, 0); + lean_ctor_release(x_160, 1); + x_196 = x_160; } else { - lean_dec_ref(x_152); - x_186 = lean_box(0); + lean_dec_ref(x_160); + x_196 = lean_box(0); } -if (lean_is_scalar(x_186)) { - x_187 = lean_alloc_ctor(1, 2, 0); +if (lean_is_scalar(x_196)) { + x_197 = lean_alloc_ctor(1, 2, 0); } else { - x_187 = x_186; + x_197 = x_196; } -lean_ctor_set(x_187, 0, x_184); -lean_ctor_set(x_187, 1, x_185); -return x_187; +lean_ctor_set(x_197, 0, x_194); +lean_ctor_set(x_197, 1, x_195); +return x_197; } } } else { -lean_object* x_188; lean_object* x_189; lean_object* x_190; lean_object* x_191; lean_object* x_192; lean_object* x_193; lean_object* x_194; lean_object* x_195; lean_object* x_196; lean_object* x_197; lean_object* x_198; -x_188 = lean_ctor_get(x_1, 1); -lean_inc(x_188); +lean_object* x_198; lean_object* x_199; lean_object* x_200; lean_object* x_201; lean_object* x_202; lean_object* x_203; lean_object* x_204; lean_object* x_205; lean_object* x_206; lean_object* x_207; lean_object* x_208; +x_198 = lean_ctor_get(x_1, 1); +lean_inc(x_198); lean_dec(x_1); -x_189 = lean_ctor_get(x_10, 0); -lean_inc(x_189); -x_190 = lean_ctor_get(x_10, 1); -lean_inc(x_190); -x_191 = lean_ctor_get(x_10, 2); -lean_inc(x_191); -x_192 = lean_ctor_get(x_10, 3); -lean_inc(x_192); -x_193 = lean_ctor_get(x_10, 5); -lean_inc(x_193); -x_194 = lean_ctor_get(x_11, 1); -lean_inc(x_194); +x_199 = lean_ctor_get(x_10, 0); +lean_inc(x_199); +x_200 = lean_ctor_get(x_10, 1); +lean_inc(x_200); +x_201 = lean_ctor_get(x_10, 2); +lean_inc(x_201); +x_202 = lean_ctor_get(x_10, 3); +lean_inc(x_202); +x_203 = lean_ctor_get(x_10, 5); +lean_inc(x_203); +x_204 = lean_ctor_get(x_11, 1); +lean_inc(x_204); if (lean_is_exclusive(x_11)) { lean_ctor_release(x_11, 0); lean_ctor_release(x_11, 1); - x_195 = x_11; + x_205 = x_11; } else { lean_dec_ref(x_11); - x_195 = lean_box(0); + x_205 = lean_box(0); } -x_196 = lean_ctor_get(x_19, 0); -lean_inc(x_196); +x_206 = lean_ctor_get(x_19, 0); +lean_inc(x_206); if (lean_is_exclusive(x_19)) { lean_ctor_release(x_19, 0); - x_197 = x_19; + x_207 = x_19; } else { lean_dec_ref(x_19); - x_197 = lean_box(0); + x_207 = lean_box(0); } lean_inc(x_6); lean_inc(x_5); lean_inc(x_4); lean_inc(x_3); -x_198 = l_Lean_Meta_getBitVecValue_x3f(x_196, x_3, x_4, x_5, x_6, x_7); -if (lean_obj_tag(x_198) == 0) +x_208 = l_Lean_Meta_getBitVecValue_x3f(x_206, x_3, x_4, x_5, x_6, x_7); +if (lean_obj_tag(x_208) == 0) { -lean_object* x_199; -x_199 = lean_ctor_get(x_198, 0); -lean_inc(x_199); -if (lean_obj_tag(x_199) == 0) +lean_object* x_209; +x_209 = lean_ctor_get(x_208, 0); +lean_inc(x_209); +if (lean_obj_tag(x_209) == 0) { -lean_object* x_200; lean_object* x_201; -lean_dec(x_197); -lean_dec(x_194); -lean_dec(x_193); -lean_dec(x_192); -lean_dec(x_191); -lean_dec(x_190); -lean_dec(x_189); -x_200 = lean_ctor_get(x_198, 1); -lean_inc(x_200); -lean_dec(x_198); -if (lean_is_scalar(x_195)) { - x_201 = lean_alloc_ctor(1, 2, 0); +lean_object* x_210; lean_object* x_211; +lean_dec(x_207); +lean_dec(x_204); +lean_dec(x_203); +lean_dec(x_202); +lean_dec(x_201); +lean_dec(x_200); +lean_dec(x_199); +x_210 = lean_ctor_get(x_208, 1); +lean_inc(x_210); +lean_dec(x_208); +if (lean_is_scalar(x_205)) { + x_211 = lean_alloc_ctor(1, 2, 0); } else { - x_201 = x_195; + x_211 = x_205; } -lean_ctor_set(x_201, 0, x_10); -lean_ctor_set(x_201, 1, x_2); -x_1 = x_188; -x_2 = x_201; -x_7 = x_200; +lean_ctor_set(x_211, 0, x_10); +lean_ctor_set(x_211, 1, x_2); +x_1 = x_198; +x_2 = x_211; +x_7 = x_210; goto _start; } else { -lean_object* x_203; lean_object* x_204; lean_object* x_205; lean_object* x_206; lean_object* x_207; lean_object* x_208; lean_object* x_209; lean_object* x_210; lean_object* x_211; lean_object* x_212; lean_object* x_213; lean_object* x_214; lean_object* x_215; lean_object* x_216; lean_object* x_217; lean_object* x_218; lean_object* x_219; lean_object* x_220; lean_object* x_221; lean_object* x_222; lean_object* x_223; lean_object* x_224; lean_object* x_225; lean_object* x_226; lean_object* x_227; lean_object* x_228; lean_object* x_229; +lean_object* x_213; lean_object* x_214; lean_object* x_215; lean_object* x_216; lean_object* x_217; lean_object* x_218; lean_object* x_219; lean_object* x_220; lean_object* x_221; lean_object* x_222; lean_object* x_223; lean_object* x_224; lean_object* x_225; lean_object* x_226; lean_object* x_227; lean_object* x_228; lean_object* x_229; lean_object* x_230; lean_object* x_231; lean_object* x_232; lean_object* x_233; lean_object* x_234; lean_object* x_235; lean_object* x_236; lean_object* x_237; lean_object* x_238; lean_object* x_239; lean_object* x_240; lean_object* x_241; if (lean_is_exclusive(x_10)) { lean_ctor_release(x_10, 0); lean_ctor_release(x_10, 1); @@ -28066,168 +28111,171 @@ if (lean_is_exclusive(x_10)) { lean_ctor_release(x_10, 3); lean_ctor_release(x_10, 4); lean_ctor_release(x_10, 5); - x_203 = x_10; + x_213 = x_10; } else { lean_dec_ref(x_10); - x_203 = lean_box(0); -} -x_204 = lean_ctor_get(x_199, 0); -lean_inc(x_204); -lean_dec(x_199); -x_205 = lean_ctor_get(x_198, 1); -lean_inc(x_205); -lean_dec(x_198); -x_206 = lean_ctor_get(x_204, 0); -lean_inc(x_206); -x_207 = lean_ctor_get(x_204, 1); -lean_inc(x_207); -if (lean_is_exclusive(x_204)) { - lean_ctor_release(x_204, 0); - lean_ctor_release(x_204, 1); - x_208 = x_204; -} else { - lean_dec_ref(x_204); - x_208 = lean_box(0); -} -x_209 = lean_box(0); -x_210 = lean_unsigned_to_nat(2u); -x_211 = lean_nat_pow(x_210, x_206); -lean_dec(x_206); -x_212 = l_Lean_mkRawNatLit(x_207); -lean_inc(x_211); -x_213 = l_Lean_mkNatLit(x_211); -x_214 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__10; -x_215 = l_Lean_Expr_app___override(x_214, x_213); -x_216 = lean_unsigned_to_nat(1u); -x_217 = lean_nat_sub(x_211, x_216); -lean_dec(x_211); -x_218 = l_Lean_mkNatLit(x_217); -x_219 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; -lean_inc(x_212); -x_220 = l_Lean_mkAppB(x_219, x_218, x_212); -x_221 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; -x_222 = l_Lean_mkApp3(x_221, x_215, x_212, x_220); -if (lean_is_scalar(x_197)) { - x_223 = lean_alloc_ctor(3, 1, 0); -} else { - x_223 = x_197; + x_213 = lean_box(0); } -lean_ctor_set(x_223, 0, x_222); -if (lean_is_scalar(x_195)) { - x_224 = lean_alloc_ctor(1, 2, 0); +x_214 = lean_ctor_get(x_209, 0); +lean_inc(x_214); +lean_dec(x_209); +x_215 = lean_ctor_get(x_208, 1); +lean_inc(x_215); +lean_dec(x_208); +x_216 = lean_ctor_get(x_214, 0); +lean_inc(x_216); +x_217 = lean_ctor_get(x_214, 1); +lean_inc(x_217); +if (lean_is_exclusive(x_214)) { + lean_ctor_release(x_214, 0); + lean_ctor_release(x_214, 1); + x_218 = x_214; +} else { + lean_dec_ref(x_214); + x_218 = lean_box(0); +} +x_219 = lean_box(0); +x_220 = lean_unsigned_to_nat(2u); +x_221 = lean_nat_pow(x_220, x_216); +lean_dec(x_216); +x_222 = l_Lean_mkRawNatLit(x_217); +lean_inc(x_221); +x_223 = l_Lean_mkNatLit(x_221); +x_224 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__10; +lean_inc(x_223); +x_225 = l_Lean_Expr_app___override(x_224, x_223); +x_226 = lean_unsigned_to_nat(1u); +x_227 = lean_nat_sub(x_221, x_226); +lean_dec(x_221); +x_228 = l_Lean_mkNatLit(x_227); +x_229 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__16; +x_230 = l_Lean_Expr_app___override(x_229, x_228); +x_231 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13; +lean_inc(x_222); +x_232 = l_Lean_mkApp3(x_231, x_223, x_230, x_222); +x_233 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__8; +x_234 = l_Lean_mkApp3(x_233, x_225, x_222, x_232); +if (lean_is_scalar(x_207)) { + x_235 = lean_alloc_ctor(3, 1, 0); } else { - x_224 = x_195; + x_235 = x_207; } -lean_ctor_set(x_224, 0, x_223); -lean_ctor_set(x_224, 1, x_209); -x_225 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; -x_226 = lean_alloc_ctor(2, 4, 0); -lean_ctor_set(x_226, 0, x_225); -lean_ctor_set(x_226, 1, x_209); -lean_ctor_set(x_226, 2, x_209); -lean_ctor_set(x_226, 3, x_224); -x_227 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_227, 0, x_226); -lean_ctor_set(x_227, 1, x_194); -if (lean_is_scalar(x_203)) { - x_228 = lean_alloc_ctor(0, 6, 0); +lean_ctor_set(x_235, 0, x_234); +if (lean_is_scalar(x_205)) { + x_236 = lean_alloc_ctor(1, 2, 0); +} else { + x_236 = x_205; +} +lean_ctor_set(x_236, 0, x_235); +lean_ctor_set(x_236, 1, x_219); +x_237 = l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__3; +x_238 = lean_alloc_ctor(2, 4, 0); +lean_ctor_set(x_238, 0, x_237); +lean_ctor_set(x_238, 1, x_219); +lean_ctor_set(x_238, 2, x_219); +lean_ctor_set(x_238, 3, x_236); +x_239 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_239, 0, x_238); +lean_ctor_set(x_239, 1, x_204); +if (lean_is_scalar(x_213)) { + x_240 = lean_alloc_ctor(0, 6, 0); } else { - x_228 = x_203; + x_240 = x_213; } -lean_ctor_set(x_228, 0, x_189); -lean_ctor_set(x_228, 1, x_190); -lean_ctor_set(x_228, 2, x_191); -lean_ctor_set(x_228, 3, x_192); -lean_ctor_set(x_228, 4, x_227); -lean_ctor_set(x_228, 5, x_193); -if (lean_is_scalar(x_208)) { - x_229 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_240, 0, x_199); +lean_ctor_set(x_240, 1, x_200); +lean_ctor_set(x_240, 2, x_201); +lean_ctor_set(x_240, 3, x_202); +lean_ctor_set(x_240, 4, x_239); +lean_ctor_set(x_240, 5, x_203); +if (lean_is_scalar(x_218)) { + x_241 = lean_alloc_ctor(1, 2, 0); } else { - x_229 = x_208; - lean_ctor_set_tag(x_229, 1); + x_241 = x_218; + lean_ctor_set_tag(x_241, 1); } -lean_ctor_set(x_229, 0, x_228); -lean_ctor_set(x_229, 1, x_2); -x_1 = x_188; -x_2 = x_229; -x_7 = x_205; +lean_ctor_set(x_241, 0, x_240); +lean_ctor_set(x_241, 1, x_2); +x_1 = x_198; +x_2 = x_241; +x_7 = x_215; goto _start; } } else { -lean_object* x_231; lean_object* x_232; lean_object* x_233; lean_object* x_234; -lean_dec(x_197); -lean_dec(x_195); -lean_dec(x_194); -lean_dec(x_193); -lean_dec(x_192); -lean_dec(x_191); -lean_dec(x_190); -lean_dec(x_189); -lean_dec(x_188); +lean_object* x_243; lean_object* x_244; lean_object* x_245; lean_object* x_246; +lean_dec(x_207); +lean_dec(x_205); +lean_dec(x_204); +lean_dec(x_203); +lean_dec(x_202); +lean_dec(x_201); +lean_dec(x_200); +lean_dec(x_199); +lean_dec(x_198); lean_dec(x_10); lean_dec(x_6); lean_dec(x_5); lean_dec(x_4); lean_dec(x_3); lean_dec(x_2); -x_231 = lean_ctor_get(x_198, 0); -lean_inc(x_231); -x_232 = lean_ctor_get(x_198, 1); -lean_inc(x_232); -if (lean_is_exclusive(x_198)) { - lean_ctor_release(x_198, 0); - lean_ctor_release(x_198, 1); - x_233 = x_198; +x_243 = lean_ctor_get(x_208, 0); +lean_inc(x_243); +x_244 = lean_ctor_get(x_208, 1); +lean_inc(x_244); +if (lean_is_exclusive(x_208)) { + lean_ctor_release(x_208, 0); + lean_ctor_release(x_208, 1); + x_245 = x_208; } else { - lean_dec_ref(x_198); - x_233 = lean_box(0); + lean_dec_ref(x_208); + x_245 = lean_box(0); } -if (lean_is_scalar(x_233)) { - x_234 = lean_alloc_ctor(1, 2, 0); +if (lean_is_scalar(x_245)) { + x_246 = lean_alloc_ctor(1, 2, 0); } else { - x_234 = x_233; + x_246 = x_245; } -lean_ctor_set(x_234, 0, x_231); -lean_ctor_set(x_234, 1, x_232); -return x_234; +lean_ctor_set(x_246, 0, x_243); +lean_ctor_set(x_246, 1, x_244); +return x_246; } } } else { -uint8_t x_235; +uint8_t x_247; lean_dec(x_19); -x_235 = !lean_is_exclusive(x_11); -if (x_235 == 0) +x_247 = !lean_is_exclusive(x_11); +if (x_247 == 0) { -lean_object* x_236; lean_object* x_237; lean_object* x_238; -x_236 = lean_ctor_get(x_11, 1); -lean_dec(x_236); -x_237 = lean_ctor_get(x_11, 0); -lean_dec(x_237); -x_238 = lean_ctor_get(x_1, 1); -lean_inc(x_238); +lean_object* x_248; lean_object* x_249; lean_object* x_250; +x_248 = lean_ctor_get(x_11, 1); +lean_dec(x_248); +x_249 = lean_ctor_get(x_11, 0); +lean_dec(x_249); +x_250 = lean_ctor_get(x_1, 1); +lean_inc(x_250); lean_dec(x_1); lean_ctor_set(x_11, 1, x_2); lean_ctor_set(x_11, 0, x_10); -x_1 = x_238; +x_1 = x_250; x_2 = x_11; goto _start; } else { -lean_object* x_240; lean_object* x_241; +lean_object* x_252; lean_object* x_253; lean_dec(x_11); -x_240 = lean_ctor_get(x_1, 1); -lean_inc(x_240); +x_252 = lean_ctor_get(x_1, 1); +lean_inc(x_252); lean_dec(x_1); -x_241 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_241, 0, x_10); -lean_ctor_set(x_241, 1, x_2); -x_1 = x_240; -x_2 = x_241; +x_253 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_253, 0, x_10); +lean_ctor_set(x_253, 1, x_2); +x_1 = x_252; +x_2 = x_253; goto _start; } } @@ -48015,6 +48063,12 @@ l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandB lean_mark_persistent(l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__12); l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13 = _init_l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13(); lean_mark_persistent(l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__13); +l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__14 = _init_l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__14(); +lean_mark_persistent(l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__14); +l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__15 = _init_l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__15(); +lean_mark_persistent(l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__15); +l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__16 = _init_l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__16(); +lean_mark_persistent(l_List_mapM_loop___at___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_expandBitVecValuePattern___spec__1___closed__16); l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_traceStep___closed__1 = _init_l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_traceStep___closed__1(); lean_mark_persistent(l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_traceStep___closed__1); l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_traceStep___closed__2 = _init_l___private_Lean_Meta_Match_Match_0__Lean_Meta_Match_traceStep___closed__2(); diff --git a/stage0/stdlib/Lean/Meta/Tactic/Simp/BuiltinSimprocs/BitVec.c b/stage0/stdlib/Lean/Meta/Tactic/Simp/BuiltinSimprocs/BitVec.c index 4db5594b4b1e..6a90d98f7e50 100644 --- a/stage0/stdlib/Lean/Meta/Tactic/Simp/BuiltinSimprocs/BitVec.c +++ b/stage0/stdlib/Lean/Meta/Tactic/Simp/BuiltinSimprocs/BitVec.c @@ -478,6 +478,7 @@ static lean_object* l___regBuiltin_BitVec_reduceReplicate_declare__1____x40_Lean static lean_object* l_BitVec_reduceHShiftRight___closed__3; LEAN_EXPORT lean_object* l___regBuiltin_BitVec_reduceULT_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_4241_(lean_object*); LEAN_EXPORT lean_object* l___regBuiltin_BitVec_reduceSShiftRight_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_2215_(lean_object*); +static lean_object* l_BitVec_reduceBitVecToFin___lambda__1___closed__11; LEAN_EXPORT lean_object* l___regBuiltin_BitVec_reduceGT_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_4177_(lean_object*); LEAN_EXPORT lean_object* l_BitVec_reduceUShiftRight___lambda__1(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); LEAN_EXPORT lean_object* l_BitVec_reduceBin(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); @@ -612,6 +613,7 @@ static lean_object* l_BitVec_reduceGE___closed__2; LEAN_EXPORT lean_object* l_BitVec_reduceUShiftRight(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); lean_object* l_Lean_Expr_appFnCleanup(lean_object*, lean_object*); static lean_object* l___regBuiltin_BitVec_reduceMul_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_1791____closed__12; +static lean_object* l_BitVec_reduceBitVecToFin___lambda__1___closed__13; static lean_object* l___regBuiltin_BitVec_reduceCast_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_3065____closed__6; lean_object* l_BitVec_srem(lean_object*, lean_object*, lean_object*); LEAN_EXPORT lean_object* l___regBuiltin_BitVec_reduceSDiv_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_2077_(lean_object*); @@ -790,6 +792,7 @@ LEAN_EXPORT lean_object* l_BitVec_reduceBitVecToFin___lambda__1(lean_object*, le LEAN_EXPORT lean_object* l___regBuiltin_BitVec_reduceSRem_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_2050_(lean_object*); static lean_object* l___regBuiltin_BitVec_reduceNeg_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_1532____closed__12; static lean_object* l___regBuiltin_BitVec_reduceAdd_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_1751____closed__8; +static lean_object* l_BitVec_reduceBitVecToFin___lambda__1___closed__10; static lean_object* l___regBuiltin_BitVec_reduceULT_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_4237____closed__7; static lean_object* l___regBuiltin_BitVec_reduceZeroExtend_x27_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_4632____closed__2; static lean_object* l___regBuiltin_BitVec_reduceNe_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_4050____closed__6; @@ -829,6 +832,7 @@ static lean_object* l___regBuiltin_BitVec_reduceSMTUDiv_declare__1____x40_Lean_M static lean_object* l___regBuiltin_BitVec_reduceXOr_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_1711____closed__3; static lean_object* l___regBuiltin_BitVec_reduceNeg_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_1532____closed__1; static lean_object* l___regBuiltin_BitVec_reduceMul_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_1791____closed__7; +static lean_object* l_BitVec_reduceBitVecToFin___lambda__1___closed__12; LEAN_EXPORT lean_object* l_BitVec_reduceBitVecToFin(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); LEAN_EXPORT lean_object* l_BitVec_reduceOfNat___lambda__1___boxed(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); static lean_object* l___regBuiltin_BitVec_reduceUMod_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_BitVec___hyg_1940____closed__1; @@ -28076,6 +28080,42 @@ x_3 = l_Lean_Expr_const___override(x_2, x_1); return x_3; } } +static lean_object* _init_l_BitVec_reduceBitVecToFin___lambda__1___closed__10() { +_start: +{ +lean_object* x_1; +x_1 = lean_mk_string_unchecked("Nat", 3, 3); +return x_1; +} +} +static lean_object* _init_l_BitVec_reduceBitVecToFin___lambda__1___closed__11() { +_start: +{ +lean_object* x_1; +x_1 = lean_mk_string_unchecked("instNeZeroSucc", 14, 14); +return x_1; +} +} +static lean_object* _init_l_BitVec_reduceBitVecToFin___lambda__1___closed__12() { +_start: +{ +lean_object* x_1; lean_object* x_2; lean_object* x_3; +x_1 = l_BitVec_reduceBitVecToFin___lambda__1___closed__10; +x_2 = l_BitVec_reduceBitVecToFin___lambda__1___closed__11; +x_3 = l_Lean_Name_mkStr2(x_1, x_2); +return x_3; +} +} +static lean_object* _init_l_BitVec_reduceBitVecToFin___lambda__1___closed__13() { +_start: +{ +lean_object* x_1; lean_object* x_2; lean_object* x_3; +x_1 = lean_box(0); +x_2 = l_BitVec_reduceBitVecToFin___lambda__1___closed__12; +x_3 = l_Lean_Expr_const___override(x_2, x_1); +return x_3; +} +} LEAN_EXPORT lean_object* l_BitVec_reduceBitVecToFin___lambda__1(lean_object* x_1, lean_object* x_2, lean_object* x_3, lean_object* x_4, lean_object* x_5, lean_object* x_6, lean_object* x_7, lean_object* x_8, lean_object* x_9) { _start: { @@ -28122,7 +28162,7 @@ uint8_t x_19; x_19 = !lean_is_exclusive(x_10); if (x_19 == 0) { -lean_object* x_20; lean_object* x_21; lean_object* x_22; lean_object* x_23; lean_object* x_24; lean_object* x_25; lean_object* x_26; lean_object* x_27; lean_object* x_28; lean_object* x_29; lean_object* x_30; lean_object* x_31; lean_object* x_32; lean_object* x_33; lean_object* x_34; lean_object* x_35; lean_object* x_36; +lean_object* x_20; lean_object* x_21; lean_object* x_22; lean_object* x_23; lean_object* x_24; lean_object* x_25; lean_object* x_26; lean_object* x_27; lean_object* x_28; lean_object* x_29; lean_object* x_30; lean_object* x_31; lean_object* x_32; lean_object* x_33; lean_object* x_34; lean_object* x_35; lean_object* x_36; lean_object* x_37; lean_object* x_38; x_20 = lean_ctor_get(x_11, 0); x_21 = lean_ctor_get(x_10, 0); lean_dec(x_21); @@ -28138,128 +28178,137 @@ x_26 = l_Lean_mkRawNatLit(x_23); lean_inc(x_25); x_27 = l_Lean_mkNatLit(x_25); x_28 = l_BitVec_reduceBitVecToFin___lambda__1___closed__6; +lean_inc(x_27); x_29 = l_Lean_Expr_app___override(x_28, x_27); x_30 = lean_unsigned_to_nat(1u); x_31 = lean_nat_sub(x_25, x_30); lean_dec(x_25); x_32 = l_Lean_mkNatLit(x_31); -x_33 = l_BitVec_reduceBitVecToFin___lambda__1___closed__9; +x_33 = l_BitVec_reduceBitVecToFin___lambda__1___closed__13; +x_34 = l_Lean_Expr_app___override(x_33, x_32); +x_35 = l_BitVec_reduceBitVecToFin___lambda__1___closed__9; lean_inc(x_26); -x_34 = l_Lean_mkAppB(x_33, x_32, x_26); -x_35 = l_BitVec_reduceBitVecToFin___lambda__1___closed__3; -x_36 = l_Lean_mkApp3(x_35, x_29, x_26, x_34); +x_36 = l_Lean_mkApp3(x_35, x_27, x_34, x_26); +x_37 = l_BitVec_reduceBitVecToFin___lambda__1___closed__3; +x_38 = l_Lean_mkApp3(x_37, x_29, x_26, x_36); lean_ctor_set_tag(x_11, 0); -lean_ctor_set(x_11, 0, x_36); +lean_ctor_set(x_11, 0, x_38); return x_10; } else { -lean_object* x_37; lean_object* x_38; lean_object* x_39; lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; lean_object* x_54; -x_37 = lean_ctor_get(x_11, 0); -x_38 = lean_ctor_get(x_10, 1); -lean_inc(x_38); -lean_dec(x_10); -x_39 = lean_ctor_get(x_37, 0); -lean_inc(x_39); -x_40 = lean_ctor_get(x_37, 1); +lean_object* x_39; lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; lean_object* x_54; lean_object* x_55; lean_object* x_56; lean_object* x_57; lean_object* x_58; +x_39 = lean_ctor_get(x_11, 0); +x_40 = lean_ctor_get(x_10, 1); lean_inc(x_40); -lean_dec(x_37); -x_41 = lean_unsigned_to_nat(2u); -x_42 = lean_nat_pow(x_41, x_39); -lean_dec(x_39); -x_43 = l_Lean_mkRawNatLit(x_40); +lean_dec(x_10); +x_41 = lean_ctor_get(x_39, 0); +lean_inc(x_41); +x_42 = lean_ctor_get(x_39, 1); lean_inc(x_42); -x_44 = l_Lean_mkNatLit(x_42); -x_45 = l_BitVec_reduceBitVecToFin___lambda__1___closed__6; -x_46 = l_Lean_Expr_app___override(x_45, x_44); -x_47 = lean_unsigned_to_nat(1u); -x_48 = lean_nat_sub(x_42, x_47); -lean_dec(x_42); -x_49 = l_Lean_mkNatLit(x_48); -x_50 = l_BitVec_reduceBitVecToFin___lambda__1___closed__9; -lean_inc(x_43); -x_51 = l_Lean_mkAppB(x_50, x_49, x_43); -x_52 = l_BitVec_reduceBitVecToFin___lambda__1___closed__3; -x_53 = l_Lean_mkApp3(x_52, x_46, x_43, x_51); +lean_dec(x_39); +x_43 = lean_unsigned_to_nat(2u); +x_44 = lean_nat_pow(x_43, x_41); +lean_dec(x_41); +x_45 = l_Lean_mkRawNatLit(x_42); +lean_inc(x_44); +x_46 = l_Lean_mkNatLit(x_44); +x_47 = l_BitVec_reduceBitVecToFin___lambda__1___closed__6; +lean_inc(x_46); +x_48 = l_Lean_Expr_app___override(x_47, x_46); +x_49 = lean_unsigned_to_nat(1u); +x_50 = lean_nat_sub(x_44, x_49); +lean_dec(x_44); +x_51 = l_Lean_mkNatLit(x_50); +x_52 = l_BitVec_reduceBitVecToFin___lambda__1___closed__13; +x_53 = l_Lean_Expr_app___override(x_52, x_51); +x_54 = l_BitVec_reduceBitVecToFin___lambda__1___closed__9; +lean_inc(x_45); +x_55 = l_Lean_mkApp3(x_54, x_46, x_53, x_45); +x_56 = l_BitVec_reduceBitVecToFin___lambda__1___closed__3; +x_57 = l_Lean_mkApp3(x_56, x_48, x_45, x_55); lean_ctor_set_tag(x_11, 0); -lean_ctor_set(x_11, 0, x_53); -x_54 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_54, 0, x_11); -lean_ctor_set(x_54, 1, x_38); -return x_54; +lean_ctor_set(x_11, 0, x_57); +x_58 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_58, 0, x_11); +lean_ctor_set(x_58, 1, x_40); +return x_58; } } else { -lean_object* x_55; lean_object* x_56; lean_object* x_57; lean_object* x_58; lean_object* x_59; lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; lean_object* x_74; -x_55 = lean_ctor_get(x_11, 0); -lean_inc(x_55); +lean_object* x_59; lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; lean_object* x_78; lean_object* x_79; lean_object* x_80; +x_59 = lean_ctor_get(x_11, 0); +lean_inc(x_59); lean_dec(x_11); -x_56 = lean_ctor_get(x_10, 1); -lean_inc(x_56); +x_60 = lean_ctor_get(x_10, 1); +lean_inc(x_60); if (lean_is_exclusive(x_10)) { lean_ctor_release(x_10, 0); lean_ctor_release(x_10, 1); - x_57 = x_10; + x_61 = x_10; } else { lean_dec_ref(x_10); - x_57 = lean_box(0); + x_61 = lean_box(0); } -x_58 = lean_ctor_get(x_55, 0); -lean_inc(x_58); -x_59 = lean_ctor_get(x_55, 1); -lean_inc(x_59); -lean_dec(x_55); -x_60 = lean_unsigned_to_nat(2u); -x_61 = lean_nat_pow(x_60, x_58); -lean_dec(x_58); -x_62 = l_Lean_mkRawNatLit(x_59); -lean_inc(x_61); -x_63 = l_Lean_mkNatLit(x_61); -x_64 = l_BitVec_reduceBitVecToFin___lambda__1___closed__6; -x_65 = l_Lean_Expr_app___override(x_64, x_63); -x_66 = lean_unsigned_to_nat(1u); -x_67 = lean_nat_sub(x_61, x_66); -lean_dec(x_61); -x_68 = l_Lean_mkNatLit(x_67); -x_69 = l_BitVec_reduceBitVecToFin___lambda__1___closed__9; +x_62 = lean_ctor_get(x_59, 0); lean_inc(x_62); -x_70 = l_Lean_mkAppB(x_69, x_68, x_62); -x_71 = l_BitVec_reduceBitVecToFin___lambda__1___closed__3; -x_72 = l_Lean_mkApp3(x_71, x_65, x_62, x_70); -x_73 = lean_alloc_ctor(0, 1, 0); -lean_ctor_set(x_73, 0, x_72); -if (lean_is_scalar(x_57)) { - x_74 = lean_alloc_ctor(0, 2, 0); +x_63 = lean_ctor_get(x_59, 1); +lean_inc(x_63); +lean_dec(x_59); +x_64 = lean_unsigned_to_nat(2u); +x_65 = lean_nat_pow(x_64, x_62); +lean_dec(x_62); +x_66 = l_Lean_mkRawNatLit(x_63); +lean_inc(x_65); +x_67 = l_Lean_mkNatLit(x_65); +x_68 = l_BitVec_reduceBitVecToFin___lambda__1___closed__6; +lean_inc(x_67); +x_69 = l_Lean_Expr_app___override(x_68, x_67); +x_70 = lean_unsigned_to_nat(1u); +x_71 = lean_nat_sub(x_65, x_70); +lean_dec(x_65); +x_72 = l_Lean_mkNatLit(x_71); +x_73 = l_BitVec_reduceBitVecToFin___lambda__1___closed__13; +x_74 = l_Lean_Expr_app___override(x_73, x_72); +x_75 = l_BitVec_reduceBitVecToFin___lambda__1___closed__9; +lean_inc(x_66); +x_76 = l_Lean_mkApp3(x_75, x_67, x_74, x_66); +x_77 = l_BitVec_reduceBitVecToFin___lambda__1___closed__3; +x_78 = l_Lean_mkApp3(x_77, x_69, x_66, x_76); +x_79 = lean_alloc_ctor(0, 1, 0); +lean_ctor_set(x_79, 0, x_78); +if (lean_is_scalar(x_61)) { + x_80 = lean_alloc_ctor(0, 2, 0); } else { - x_74 = x_57; + x_80 = x_61; } -lean_ctor_set(x_74, 0, x_73); -lean_ctor_set(x_74, 1, x_56); -return x_74; +lean_ctor_set(x_80, 0, x_79); +lean_ctor_set(x_80, 1, x_60); +return x_80; } } } else { -uint8_t x_75; -x_75 = !lean_is_exclusive(x_10); -if (x_75 == 0) +uint8_t x_81; +x_81 = !lean_is_exclusive(x_10); +if (x_81 == 0) { return x_10; } else { -lean_object* x_76; lean_object* x_77; lean_object* x_78; -x_76 = lean_ctor_get(x_10, 0); -x_77 = lean_ctor_get(x_10, 1); -lean_inc(x_77); -lean_inc(x_76); +lean_object* x_82; lean_object* x_83; lean_object* x_84; +x_82 = lean_ctor_get(x_10, 0); +x_83 = lean_ctor_get(x_10, 1); +lean_inc(x_83); +lean_inc(x_82); lean_dec(x_10); -x_78 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_78, 0, x_76); -lean_ctor_set(x_78, 1, x_77); -return x_78; +x_84 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_84, 0, x_82); +lean_ctor_set(x_84, 1, x_83); +return x_84; } } } @@ -31684,6 +31733,14 @@ l_BitVec_reduceBitVecToFin___lambda__1___closed__8 = _init_l_BitVec_reduceBitVec lean_mark_persistent(l_BitVec_reduceBitVecToFin___lambda__1___closed__8); l_BitVec_reduceBitVecToFin___lambda__1___closed__9 = _init_l_BitVec_reduceBitVecToFin___lambda__1___closed__9(); lean_mark_persistent(l_BitVec_reduceBitVecToFin___lambda__1___closed__9); +l_BitVec_reduceBitVecToFin___lambda__1___closed__10 = _init_l_BitVec_reduceBitVecToFin___lambda__1___closed__10(); +lean_mark_persistent(l_BitVec_reduceBitVecToFin___lambda__1___closed__10); +l_BitVec_reduceBitVecToFin___lambda__1___closed__11 = _init_l_BitVec_reduceBitVecToFin___lambda__1___closed__11(); +lean_mark_persistent(l_BitVec_reduceBitVecToFin___lambda__1___closed__11); +l_BitVec_reduceBitVecToFin___lambda__1___closed__12 = _init_l_BitVec_reduceBitVecToFin___lambda__1___closed__12(); +lean_mark_persistent(l_BitVec_reduceBitVecToFin___lambda__1___closed__12); +l_BitVec_reduceBitVecToFin___lambda__1___closed__13 = _init_l_BitVec_reduceBitVecToFin___lambda__1___closed__13(); +lean_mark_persistent(l_BitVec_reduceBitVecToFin___lambda__1___closed__13); l_BitVec_reduceBitVecToFin___closed__1 = _init_l_BitVec_reduceBitVecToFin___closed__1(); lean_mark_persistent(l_BitVec_reduceBitVecToFin___closed__1); l_BitVec_reduceBitVecToFin___closed__2 = _init_l_BitVec_reduceBitVecToFin___closed__2(); diff --git a/stage0/stdlib/Lean/Meta/Tactic/Simp/BuiltinSimprocs/Fin.c b/stage0/stdlib/Lean/Meta/Tactic/Simp/BuiltinSimprocs/Fin.c index 8ebe585e47bd..f1435e414612 100644 --- a/stage0/stdlib/Lean/Meta/Tactic/Simp/BuiltinSimprocs/Fin.c +++ b/stage0/stdlib/Lean/Meta/Tactic/Simp/BuiltinSimprocs/Fin.c @@ -52,7 +52,6 @@ LEAN_EXPORT lean_object* l_Fin_reduceGT___boxed(lean_object*, lean_object*, lean LEAN_EXPORT lean_object* l_Fin_reduceBinPred(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); LEAN_EXPORT lean_object* l_Fin_reduceMod(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); static lean_object* l___private_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin_0__Fin_reprValue____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_172____closed__7; -lean_object* l_Lean_mkAppB(lean_object*, lean_object*, lean_object*); static lean_object* l_Fin_reduceBin___lambda__1___closed__5; static lean_object* l_Fin_reduceDiv___closed__2; LEAN_EXPORT lean_object* l_Fin_reduceDiv___lambda__1___boxed(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); @@ -95,12 +94,14 @@ static lean_object* l___regBuiltin_Fin_reduceSub_declare__1____x40_Lean_Meta_Tac static lean_object* l___regBuiltin_Fin_reduceNe_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_1198____closed__9; static lean_object* l___regBuiltin_Fin_reduceFinMk_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_1891____closed__7; LEAN_EXPORT lean_object* l___regBuiltin_Fin_reduceMul_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_843_(lean_object*); +static lean_object* l_Fin_reduceBin___lambda__1___closed__17; static lean_object* l___regBuiltin_Fin_reduceLE_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_1039____closed__4; static lean_object* l___regBuiltin_Fin_reduceMul_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_839____closed__4; static lean_object* l___regBuiltin_Fin_reduceBEq_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_1238____closed__1; static lean_object* l___regBuiltin_Fin_reduceAdd_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_799____closed__1; static lean_object* l___regBuiltin_Fin_reduceAdd_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_799____closed__7; static lean_object* l___regBuiltin_Fin_reduceAdd_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_799____closed__13; +static lean_object* l_Fin_reduceBin___lambda__1___closed__15; lean_object* l_Lean_Meta_Simp_evalPropStep(lean_object*, uint8_t, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); LEAN_EXPORT lean_object* l_Fin_reduceBEq___lambda__1___boxed(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); LEAN_EXPORT uint8_t l___private_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin_0__Fin_decEqValue____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_27_(lean_object*, lean_object*); @@ -115,6 +116,7 @@ static lean_object* l___regBuiltin_Fin_reduceBEq_declare__1____x40_Lean_Meta_Tac lean_object* l_Lean_Meta_evalNat(lean_object*, lean_object*, lean_object*, lean_object*, lean_object*, lean_object*); static lean_object* l___regBuiltin_Fin_reduceMul_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_839____closed__10; static lean_object* l_Fin_reduceMul___closed__3; +static lean_object* l_Fin_reduceBin___lambda__1___closed__16; LEAN_EXPORT lean_object* l___regBuiltin_Fin_reduceEq_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_1161_(lean_object*); LEAN_EXPORT lean_object* l___regBuiltin_Fin_reduceGT_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_1079_(lean_object*); static lean_object* l___regBuiltin_Fin_reduceBNe_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_1277____closed__3; @@ -130,6 +132,7 @@ lean_object* l_Lean_Expr_appArg_x21(lean_object*); static lean_object* l___regBuiltin_Fin_reduceEq_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_1159____closed__8; static lean_object* l___regBuiltin_Fin_reduceDiv_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_919____closed__1; static lean_object* l___regBuiltin_Fin_reduceMul_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_839____closed__1; +static lean_object* l_Fin_reduceBin___lambda__1___closed__14; static lean_object* l___regBuiltin_Fin_reduceDiv_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_919____closed__10; static lean_object* l___regBuiltin_Fin_reduceMul_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_839____closed__5; static lean_object* l___regBuiltin_Fin_reduceLT_declare__1____x40_Lean_Meta_Tactic_Simp_BuiltinSimprocs_Fin___hyg_999____closed__7; @@ -1068,6 +1071,42 @@ x_3 = l_Lean_Expr_const___override(x_2, x_1); return x_3; } } +static lean_object* _init_l_Fin_reduceBin___lambda__1___closed__14() { +_start: +{ +lean_object* x_1; +x_1 = lean_mk_string_unchecked("Nat", 3, 3); +return x_1; +} +} +static lean_object* _init_l_Fin_reduceBin___lambda__1___closed__15() { +_start: +{ +lean_object* x_1; +x_1 = lean_mk_string_unchecked("instNeZeroSucc", 14, 14); +return x_1; +} +} +static lean_object* _init_l_Fin_reduceBin___lambda__1___closed__16() { +_start: +{ +lean_object* x_1; lean_object* x_2; lean_object* x_3; +x_1 = l_Fin_reduceBin___lambda__1___closed__14; +x_2 = l_Fin_reduceBin___lambda__1___closed__15; +x_3 = l_Lean_Name_mkStr2(x_1, x_2); +return x_3; +} +} +static lean_object* _init_l_Fin_reduceBin___lambda__1___closed__17() { +_start: +{ +lean_object* x_1; lean_object* x_2; lean_object* x_3; +x_1 = lean_box(0); +x_2 = l_Fin_reduceBin___lambda__1___closed__16; +x_3 = l_Lean_Expr_const___override(x_2, x_1); +return x_3; +} +} LEAN_EXPORT lean_object* l_Fin_reduceBin___lambda__1(lean_object* x_1, lean_object* x_2, lean_object* x_3, lean_object* x_4, lean_object* x_5, lean_object* x_6, lean_object* x_7, lean_object* x_8, lean_object* x_9, lean_object* x_10, lean_object* x_11) { _start: { @@ -1194,7 +1233,7 @@ return x_25; } else { -lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; lean_object* x_54; +lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; lean_object* x_54; lean_object* x_55; lean_object* x_56; x_41 = lean_ctor_get(x_23, 1); lean_inc(x_41); lean_dec(x_23); @@ -1207,202 +1246,211 @@ x_44 = l_Lean_mkRawNatLit(x_43); lean_inc(x_37); x_45 = l_Lean_mkNatLit(x_37); x_46 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_45); x_47 = l_Lean_Expr_app___override(x_46, x_45); x_48 = lean_unsigned_to_nat(1u); x_49 = lean_nat_sub(x_37, x_48); lean_dec(x_37); x_50 = l_Lean_mkNatLit(x_49); -x_51 = l_Fin_reduceBin___lambda__1___closed__13; +x_51 = l_Fin_reduceBin___lambda__1___closed__17; +x_52 = l_Lean_Expr_app___override(x_51, x_50); +x_53 = l_Fin_reduceBin___lambda__1___closed__13; lean_inc(x_44); -x_52 = l_Lean_mkAppB(x_51, x_50, x_44); -x_53 = l_Fin_reduceBin___lambda__1___closed__7; -x_54 = l_Lean_mkApp3(x_53, x_47, x_44, x_52); +x_54 = l_Lean_mkApp3(x_53, x_45, x_52, x_44); +x_55 = l_Fin_reduceBin___lambda__1___closed__7; +x_56 = l_Lean_mkApp3(x_55, x_47, x_44, x_54); lean_ctor_set_tag(x_26, 0); -lean_ctor_set(x_26, 0, x_54); +lean_ctor_set(x_26, 0, x_56); return x_25; } } else { -lean_object* x_55; lean_object* x_56; lean_object* x_57; uint8_t x_58; -x_55 = lean_ctor_get(x_26, 0); -lean_inc(x_55); -lean_dec(x_26); -x_56 = lean_ctor_get(x_23, 0); -lean_inc(x_56); -x_57 = lean_ctor_get(x_55, 0); +lean_object* x_57; lean_object* x_58; lean_object* x_59; uint8_t x_60; +x_57 = lean_ctor_get(x_26, 0); lean_inc(x_57); -x_58 = lean_nat_dec_eq(x_56, x_57); -lean_dec(x_57); -if (x_58 == 0) +lean_dec(x_26); +x_58 = lean_ctor_get(x_23, 0); +lean_inc(x_58); +x_59 = lean_ctor_get(x_57, 0); +lean_inc(x_59); +x_60 = lean_nat_dec_eq(x_58, x_59); +lean_dec(x_59); +if (x_60 == 0) { -lean_object* x_59; -lean_dec(x_56); -lean_dec(x_55); +lean_object* x_61; +lean_dec(x_58); +lean_dec(x_57); lean_dec(x_23); lean_dec(x_2); -x_59 = l_Fin_reduceBin___lambda__1___closed__1; -lean_ctor_set(x_25, 0, x_59); +x_61 = l_Fin_reduceBin___lambda__1___closed__1; +lean_ctor_set(x_25, 0, x_61); return x_25; } else { -lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; lean_object* x_74; -x_60 = lean_ctor_get(x_23, 1); -lean_inc(x_60); +lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; lean_object* x_78; +x_62 = lean_ctor_get(x_23, 1); +lean_inc(x_62); lean_dec(x_23); -x_61 = lean_ctor_get(x_55, 1); -lean_inc(x_61); -lean_dec(x_55); -lean_inc(x_56); -x_62 = lean_apply_3(x_2, x_56, x_60, x_61); -x_63 = l_Lean_mkRawNatLit(x_62); -lean_inc(x_56); -x_64 = l_Lean_mkNatLit(x_56); -x_65 = l_Fin_reduceBin___lambda__1___closed__10; -x_66 = l_Lean_Expr_app___override(x_65, x_64); -x_67 = lean_unsigned_to_nat(1u); -x_68 = lean_nat_sub(x_56, x_67); -lean_dec(x_56); -x_69 = l_Lean_mkNatLit(x_68); -x_70 = l_Fin_reduceBin___lambda__1___closed__13; +x_63 = lean_ctor_get(x_57, 1); lean_inc(x_63); -x_71 = l_Lean_mkAppB(x_70, x_69, x_63); -x_72 = l_Fin_reduceBin___lambda__1___closed__7; -x_73 = l_Lean_mkApp3(x_72, x_66, x_63, x_71); -x_74 = lean_alloc_ctor(0, 1, 0); -lean_ctor_set(x_74, 0, x_73); -lean_ctor_set(x_25, 0, x_74); +lean_dec(x_57); +lean_inc(x_58); +x_64 = lean_apply_3(x_2, x_58, x_62, x_63); +x_65 = l_Lean_mkRawNatLit(x_64); +lean_inc(x_58); +x_66 = l_Lean_mkNatLit(x_58); +x_67 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_66); +x_68 = l_Lean_Expr_app___override(x_67, x_66); +x_69 = lean_unsigned_to_nat(1u); +x_70 = lean_nat_sub(x_58, x_69); +lean_dec(x_58); +x_71 = l_Lean_mkNatLit(x_70); +x_72 = l_Fin_reduceBin___lambda__1___closed__17; +x_73 = l_Lean_Expr_app___override(x_72, x_71); +x_74 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_65); +x_75 = l_Lean_mkApp3(x_74, x_66, x_73, x_65); +x_76 = l_Fin_reduceBin___lambda__1___closed__7; +x_77 = l_Lean_mkApp3(x_76, x_68, x_65, x_75); +x_78 = lean_alloc_ctor(0, 1, 0); +lean_ctor_set(x_78, 0, x_77); +lean_ctor_set(x_25, 0, x_78); return x_25; } } } else { -lean_object* x_75; lean_object* x_76; lean_object* x_77; lean_object* x_78; lean_object* x_79; uint8_t x_80; -x_75 = lean_ctor_get(x_25, 1); -lean_inc(x_75); +lean_object* x_79; lean_object* x_80; lean_object* x_81; lean_object* x_82; lean_object* x_83; uint8_t x_84; +x_79 = lean_ctor_get(x_25, 1); +lean_inc(x_79); lean_dec(x_25); -x_76 = lean_ctor_get(x_26, 0); -lean_inc(x_76); +x_80 = lean_ctor_get(x_26, 0); +lean_inc(x_80); if (lean_is_exclusive(x_26)) { lean_ctor_release(x_26, 0); - x_77 = x_26; + x_81 = x_26; } else { lean_dec_ref(x_26); - x_77 = lean_box(0); + x_81 = lean_box(0); } -x_78 = lean_ctor_get(x_23, 0); -lean_inc(x_78); -x_79 = lean_ctor_get(x_76, 0); -lean_inc(x_79); -x_80 = lean_nat_dec_eq(x_78, x_79); -lean_dec(x_79); -if (x_80 == 0) +x_82 = lean_ctor_get(x_23, 0); +lean_inc(x_82); +x_83 = lean_ctor_get(x_80, 0); +lean_inc(x_83); +x_84 = lean_nat_dec_eq(x_82, x_83); +lean_dec(x_83); +if (x_84 == 0) { -lean_object* x_81; lean_object* x_82; -lean_dec(x_78); -lean_dec(x_77); -lean_dec(x_76); +lean_object* x_85; lean_object* x_86; +lean_dec(x_82); +lean_dec(x_81); +lean_dec(x_80); lean_dec(x_23); lean_dec(x_2); -x_81 = l_Fin_reduceBin___lambda__1___closed__1; -x_82 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_82, 0, x_81); -lean_ctor_set(x_82, 1, x_75); -return x_82; +x_85 = l_Fin_reduceBin___lambda__1___closed__1; +x_86 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_86, 0, x_85); +lean_ctor_set(x_86, 1, x_79); +return x_86; } else { -lean_object* x_83; lean_object* x_84; lean_object* x_85; lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; lean_object* x_98; -x_83 = lean_ctor_get(x_23, 1); -lean_inc(x_83); +lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; lean_object* x_98; lean_object* x_99; lean_object* x_100; lean_object* x_101; lean_object* x_102; lean_object* x_103; lean_object* x_104; +x_87 = lean_ctor_get(x_23, 1); +lean_inc(x_87); lean_dec(x_23); -x_84 = lean_ctor_get(x_76, 1); -lean_inc(x_84); -lean_dec(x_76); -lean_inc(x_78); -x_85 = lean_apply_3(x_2, x_78, x_83, x_84); -x_86 = l_Lean_mkRawNatLit(x_85); -lean_inc(x_78); -x_87 = l_Lean_mkNatLit(x_78); -x_88 = l_Fin_reduceBin___lambda__1___closed__10; -x_89 = l_Lean_Expr_app___override(x_88, x_87); -x_90 = lean_unsigned_to_nat(1u); -x_91 = lean_nat_sub(x_78, x_90); -lean_dec(x_78); -x_92 = l_Lean_mkNatLit(x_91); -x_93 = l_Fin_reduceBin___lambda__1___closed__13; -lean_inc(x_86); -x_94 = l_Lean_mkAppB(x_93, x_92, x_86); -x_95 = l_Fin_reduceBin___lambda__1___closed__7; -x_96 = l_Lean_mkApp3(x_95, x_89, x_86, x_94); -if (lean_is_scalar(x_77)) { - x_97 = lean_alloc_ctor(0, 1, 0); +x_88 = lean_ctor_get(x_80, 1); +lean_inc(x_88); +lean_dec(x_80); +lean_inc(x_82); +x_89 = lean_apply_3(x_2, x_82, x_87, x_88); +x_90 = l_Lean_mkRawNatLit(x_89); +lean_inc(x_82); +x_91 = l_Lean_mkNatLit(x_82); +x_92 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_91); +x_93 = l_Lean_Expr_app___override(x_92, x_91); +x_94 = lean_unsigned_to_nat(1u); +x_95 = lean_nat_sub(x_82, x_94); +lean_dec(x_82); +x_96 = l_Lean_mkNatLit(x_95); +x_97 = l_Fin_reduceBin___lambda__1___closed__17; +x_98 = l_Lean_Expr_app___override(x_97, x_96); +x_99 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_90); +x_100 = l_Lean_mkApp3(x_99, x_91, x_98, x_90); +x_101 = l_Fin_reduceBin___lambda__1___closed__7; +x_102 = l_Lean_mkApp3(x_101, x_93, x_90, x_100); +if (lean_is_scalar(x_81)) { + x_103 = lean_alloc_ctor(0, 1, 0); } else { - x_97 = x_77; - lean_ctor_set_tag(x_97, 0); + x_103 = x_81; + lean_ctor_set_tag(x_103, 0); } -lean_ctor_set(x_97, 0, x_96); -x_98 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_98, 0, x_97); -lean_ctor_set(x_98, 1, x_75); -return x_98; +lean_ctor_set(x_103, 0, x_102); +x_104 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_104, 0, x_103); +lean_ctor_set(x_104, 1, x_79); +return x_104; } } } } else { -uint8_t x_99; +uint8_t x_105; lean_dec(x_23); lean_dec(x_2); -x_99 = !lean_is_exclusive(x_25); -if (x_99 == 0) +x_105 = !lean_is_exclusive(x_25); +if (x_105 == 0) { return x_25; } else { -lean_object* x_100; lean_object* x_101; lean_object* x_102; -x_100 = lean_ctor_get(x_25, 0); -x_101 = lean_ctor_get(x_25, 1); -lean_inc(x_101); -lean_inc(x_100); +lean_object* x_106; lean_object* x_107; lean_object* x_108; +x_106 = lean_ctor_get(x_25, 0); +x_107 = lean_ctor_get(x_25, 1); +lean_inc(x_107); +lean_inc(x_106); lean_dec(x_25); -x_102 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_102, 0, x_100); -lean_ctor_set(x_102, 1, x_101); -return x_102; +x_108 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_108, 0, x_106); +lean_ctor_set(x_108, 1, x_107); +return x_108; } } } } else { -uint8_t x_103; +uint8_t x_109; lean_dec(x_10); lean_dec(x_9); lean_dec(x_8); lean_dec(x_7); lean_dec(x_2); -x_103 = !lean_is_exclusive(x_14); -if (x_103 == 0) +x_109 = !lean_is_exclusive(x_14); +if (x_109 == 0) { return x_14; } else { -lean_object* x_104; lean_object* x_105; lean_object* x_106; -x_104 = lean_ctor_get(x_14, 0); -x_105 = lean_ctor_get(x_14, 1); -lean_inc(x_105); -lean_inc(x_104); +lean_object* x_110; lean_object* x_111; lean_object* x_112; +x_110 = lean_ctor_get(x_14, 0); +x_111 = lean_ctor_get(x_14, 1); +lean_inc(x_111); +lean_inc(x_110); lean_dec(x_14); -x_106 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_106, 0, x_104); -lean_ctor_set(x_106, 1, x_105); -return x_106; +x_112 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_112, 0, x_110); +lean_ctor_set(x_112, 1, x_111); +return x_112; } } } @@ -2192,7 +2240,7 @@ return x_24; } else { -lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; +lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; lean_object* x_54; lean_object* x_55; x_40 = lean_ctor_get(x_22, 1); lean_inc(x_40); lean_dec(x_22); @@ -2206,200 +2254,209 @@ x_43 = l_Lean_mkRawNatLit(x_42); lean_inc(x_36); x_44 = l_Lean_mkNatLit(x_36); x_45 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_44); x_46 = l_Lean_Expr_app___override(x_45, x_44); x_47 = lean_unsigned_to_nat(1u); x_48 = lean_nat_sub(x_36, x_47); lean_dec(x_36); x_49 = l_Lean_mkNatLit(x_48); -x_50 = l_Fin_reduceBin___lambda__1___closed__13; +x_50 = l_Fin_reduceBin___lambda__1___closed__17; +x_51 = l_Lean_Expr_app___override(x_50, x_49); +x_52 = l_Fin_reduceBin___lambda__1___closed__13; lean_inc(x_43); -x_51 = l_Lean_mkAppB(x_50, x_49, x_43); -x_52 = l_Fin_reduceBin___lambda__1___closed__7; -x_53 = l_Lean_mkApp3(x_52, x_46, x_43, x_51); +x_53 = l_Lean_mkApp3(x_52, x_44, x_51, x_43); +x_54 = l_Fin_reduceBin___lambda__1___closed__7; +x_55 = l_Lean_mkApp3(x_54, x_46, x_43, x_53); lean_ctor_set_tag(x_25, 0); -lean_ctor_set(x_25, 0, x_53); +lean_ctor_set(x_25, 0, x_55); return x_24; } } else { -lean_object* x_54; lean_object* x_55; lean_object* x_56; uint8_t x_57; -x_54 = lean_ctor_get(x_25, 0); -lean_inc(x_54); -lean_dec(x_25); -x_55 = lean_ctor_get(x_22, 0); -lean_inc(x_55); -x_56 = lean_ctor_get(x_54, 0); +lean_object* x_56; lean_object* x_57; lean_object* x_58; uint8_t x_59; +x_56 = lean_ctor_get(x_25, 0); lean_inc(x_56); -x_57 = lean_nat_dec_eq(x_55, x_56); -lean_dec(x_56); -if (x_57 == 0) +lean_dec(x_25); +x_57 = lean_ctor_get(x_22, 0); +lean_inc(x_57); +x_58 = lean_ctor_get(x_56, 0); +lean_inc(x_58); +x_59 = lean_nat_dec_eq(x_57, x_58); +lean_dec(x_58); +if (x_59 == 0) { -lean_object* x_58; -lean_dec(x_55); -lean_dec(x_54); +lean_object* x_60; +lean_dec(x_57); +lean_dec(x_56); lean_dec(x_22); -x_58 = l_Fin_reduceBin___lambda__1___closed__1; -lean_ctor_set(x_24, 0, x_58); +x_60 = l_Fin_reduceBin___lambda__1___closed__1; +lean_ctor_set(x_24, 0, x_60); return x_24; } else { -lean_object* x_59; lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; -x_59 = lean_ctor_get(x_22, 1); -lean_inc(x_59); +lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; +x_61 = lean_ctor_get(x_22, 1); +lean_inc(x_61); lean_dec(x_22); -x_60 = lean_ctor_get(x_54, 1); -lean_inc(x_60); -lean_dec(x_54); -x_61 = l_Fin_add(x_55, x_59, x_60); -lean_dec(x_60); -lean_dec(x_59); -x_62 = l_Lean_mkRawNatLit(x_61); -lean_inc(x_55); -x_63 = l_Lean_mkNatLit(x_55); -x_64 = l_Fin_reduceBin___lambda__1___closed__10; -x_65 = l_Lean_Expr_app___override(x_64, x_63); -x_66 = lean_unsigned_to_nat(1u); -x_67 = lean_nat_sub(x_55, x_66); -lean_dec(x_55); -x_68 = l_Lean_mkNatLit(x_67); -x_69 = l_Fin_reduceBin___lambda__1___closed__13; +x_62 = lean_ctor_get(x_56, 1); lean_inc(x_62); -x_70 = l_Lean_mkAppB(x_69, x_68, x_62); -x_71 = l_Fin_reduceBin___lambda__1___closed__7; -x_72 = l_Lean_mkApp3(x_71, x_65, x_62, x_70); -x_73 = lean_alloc_ctor(0, 1, 0); -lean_ctor_set(x_73, 0, x_72); -lean_ctor_set(x_24, 0, x_73); +lean_dec(x_56); +x_63 = l_Fin_add(x_57, x_61, x_62); +lean_dec(x_62); +lean_dec(x_61); +x_64 = l_Lean_mkRawNatLit(x_63); +lean_inc(x_57); +x_65 = l_Lean_mkNatLit(x_57); +x_66 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_65); +x_67 = l_Lean_Expr_app___override(x_66, x_65); +x_68 = lean_unsigned_to_nat(1u); +x_69 = lean_nat_sub(x_57, x_68); +lean_dec(x_57); +x_70 = l_Lean_mkNatLit(x_69); +x_71 = l_Fin_reduceBin___lambda__1___closed__17; +x_72 = l_Lean_Expr_app___override(x_71, x_70); +x_73 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_64); +x_74 = l_Lean_mkApp3(x_73, x_65, x_72, x_64); +x_75 = l_Fin_reduceBin___lambda__1___closed__7; +x_76 = l_Lean_mkApp3(x_75, x_67, x_64, x_74); +x_77 = lean_alloc_ctor(0, 1, 0); +lean_ctor_set(x_77, 0, x_76); +lean_ctor_set(x_24, 0, x_77); return x_24; } } } else { -lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; lean_object* x_78; uint8_t x_79; -x_74 = lean_ctor_get(x_24, 1); -lean_inc(x_74); +lean_object* x_78; lean_object* x_79; lean_object* x_80; lean_object* x_81; lean_object* x_82; uint8_t x_83; +x_78 = lean_ctor_get(x_24, 1); +lean_inc(x_78); lean_dec(x_24); -x_75 = lean_ctor_get(x_25, 0); -lean_inc(x_75); +x_79 = lean_ctor_get(x_25, 0); +lean_inc(x_79); if (lean_is_exclusive(x_25)) { lean_ctor_release(x_25, 0); - x_76 = x_25; + x_80 = x_25; } else { lean_dec_ref(x_25); - x_76 = lean_box(0); + x_80 = lean_box(0); } -x_77 = lean_ctor_get(x_22, 0); -lean_inc(x_77); -x_78 = lean_ctor_get(x_75, 0); -lean_inc(x_78); -x_79 = lean_nat_dec_eq(x_77, x_78); -lean_dec(x_78); -if (x_79 == 0) -{ -lean_object* x_80; lean_object* x_81; -lean_dec(x_77); -lean_dec(x_76); -lean_dec(x_75); +x_81 = lean_ctor_get(x_22, 0); +lean_inc(x_81); +x_82 = lean_ctor_get(x_79, 0); +lean_inc(x_82); +x_83 = lean_nat_dec_eq(x_81, x_82); +lean_dec(x_82); +if (x_83 == 0) +{ +lean_object* x_84; lean_object* x_85; +lean_dec(x_81); +lean_dec(x_80); +lean_dec(x_79); lean_dec(x_22); -x_80 = l_Fin_reduceBin___lambda__1___closed__1; -x_81 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_81, 0, x_80); -lean_ctor_set(x_81, 1, x_74); -return x_81; +x_84 = l_Fin_reduceBin___lambda__1___closed__1; +x_85 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_85, 0, x_84); +lean_ctor_set(x_85, 1, x_78); +return x_85; } else { -lean_object* x_82; lean_object* x_83; lean_object* x_84; lean_object* x_85; lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; -x_82 = lean_ctor_get(x_22, 1); -lean_inc(x_82); +lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; lean_object* x_98; lean_object* x_99; lean_object* x_100; lean_object* x_101; lean_object* x_102; lean_object* x_103; +x_86 = lean_ctor_get(x_22, 1); +lean_inc(x_86); lean_dec(x_22); -x_83 = lean_ctor_get(x_75, 1); -lean_inc(x_83); -lean_dec(x_75); -x_84 = l_Fin_add(x_77, x_82, x_83); -lean_dec(x_83); -lean_dec(x_82); -x_85 = l_Lean_mkRawNatLit(x_84); -lean_inc(x_77); -x_86 = l_Lean_mkNatLit(x_77); -x_87 = l_Fin_reduceBin___lambda__1___closed__10; -x_88 = l_Lean_Expr_app___override(x_87, x_86); -x_89 = lean_unsigned_to_nat(1u); -x_90 = lean_nat_sub(x_77, x_89); -lean_dec(x_77); -x_91 = l_Lean_mkNatLit(x_90); -x_92 = l_Fin_reduceBin___lambda__1___closed__13; -lean_inc(x_85); -x_93 = l_Lean_mkAppB(x_92, x_91, x_85); -x_94 = l_Fin_reduceBin___lambda__1___closed__7; -x_95 = l_Lean_mkApp3(x_94, x_88, x_85, x_93); -if (lean_is_scalar(x_76)) { - x_96 = lean_alloc_ctor(0, 1, 0); +x_87 = lean_ctor_get(x_79, 1); +lean_inc(x_87); +lean_dec(x_79); +x_88 = l_Fin_add(x_81, x_86, x_87); +lean_dec(x_87); +lean_dec(x_86); +x_89 = l_Lean_mkRawNatLit(x_88); +lean_inc(x_81); +x_90 = l_Lean_mkNatLit(x_81); +x_91 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_90); +x_92 = l_Lean_Expr_app___override(x_91, x_90); +x_93 = lean_unsigned_to_nat(1u); +x_94 = lean_nat_sub(x_81, x_93); +lean_dec(x_81); +x_95 = l_Lean_mkNatLit(x_94); +x_96 = l_Fin_reduceBin___lambda__1___closed__17; +x_97 = l_Lean_Expr_app___override(x_96, x_95); +x_98 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_89); +x_99 = l_Lean_mkApp3(x_98, x_90, x_97, x_89); +x_100 = l_Fin_reduceBin___lambda__1___closed__7; +x_101 = l_Lean_mkApp3(x_100, x_92, x_89, x_99); +if (lean_is_scalar(x_80)) { + x_102 = lean_alloc_ctor(0, 1, 0); } else { - x_96 = x_76; - lean_ctor_set_tag(x_96, 0); + x_102 = x_80; + lean_ctor_set_tag(x_102, 0); } -lean_ctor_set(x_96, 0, x_95); -x_97 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_97, 0, x_96); -lean_ctor_set(x_97, 1, x_74); -return x_97; +lean_ctor_set(x_102, 0, x_101); +x_103 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_103, 0, x_102); +lean_ctor_set(x_103, 1, x_78); +return x_103; } } } } else { -uint8_t x_98; +uint8_t x_104; lean_dec(x_22); -x_98 = !lean_is_exclusive(x_24); -if (x_98 == 0) +x_104 = !lean_is_exclusive(x_24); +if (x_104 == 0) { return x_24; } else { -lean_object* x_99; lean_object* x_100; lean_object* x_101; -x_99 = lean_ctor_get(x_24, 0); -x_100 = lean_ctor_get(x_24, 1); -lean_inc(x_100); -lean_inc(x_99); +lean_object* x_105; lean_object* x_106; lean_object* x_107; +x_105 = lean_ctor_get(x_24, 0); +x_106 = lean_ctor_get(x_24, 1); +lean_inc(x_106); +lean_inc(x_105); lean_dec(x_24); -x_101 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_101, 0, x_99); -lean_ctor_set(x_101, 1, x_100); -return x_101; +x_107 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_107, 0, x_105); +lean_ctor_set(x_107, 1, x_106); +return x_107; } } } } else { -uint8_t x_102; +uint8_t x_108; lean_dec(x_9); lean_dec(x_8); lean_dec(x_7); lean_dec(x_6); -x_102 = !lean_is_exclusive(x_13); -if (x_102 == 0) +x_108 = !lean_is_exclusive(x_13); +if (x_108 == 0) { return x_13; } else { -lean_object* x_103; lean_object* x_104; lean_object* x_105; -x_103 = lean_ctor_get(x_13, 0); -x_104 = lean_ctor_get(x_13, 1); -lean_inc(x_104); -lean_inc(x_103); +lean_object* x_109; lean_object* x_110; lean_object* x_111; +x_109 = lean_ctor_get(x_13, 0); +x_110 = lean_ctor_get(x_13, 1); +lean_inc(x_110); +lean_inc(x_109); lean_dec(x_13); -x_105 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_105, 0, x_103); -lean_ctor_set(x_105, 1, x_104); -return x_105; +x_111 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_111, 0, x_109); +lean_ctor_set(x_111, 1, x_110); +return x_111; } } } @@ -2827,7 +2884,7 @@ return x_24; } else { -lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; +lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; lean_object* x_54; lean_object* x_55; x_40 = lean_ctor_get(x_22, 1); lean_inc(x_40); lean_dec(x_22); @@ -2841,200 +2898,209 @@ x_43 = l_Lean_mkRawNatLit(x_42); lean_inc(x_36); x_44 = l_Lean_mkNatLit(x_36); x_45 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_44); x_46 = l_Lean_Expr_app___override(x_45, x_44); x_47 = lean_unsigned_to_nat(1u); x_48 = lean_nat_sub(x_36, x_47); lean_dec(x_36); x_49 = l_Lean_mkNatLit(x_48); -x_50 = l_Fin_reduceBin___lambda__1___closed__13; +x_50 = l_Fin_reduceBin___lambda__1___closed__17; +x_51 = l_Lean_Expr_app___override(x_50, x_49); +x_52 = l_Fin_reduceBin___lambda__1___closed__13; lean_inc(x_43); -x_51 = l_Lean_mkAppB(x_50, x_49, x_43); -x_52 = l_Fin_reduceBin___lambda__1___closed__7; -x_53 = l_Lean_mkApp3(x_52, x_46, x_43, x_51); +x_53 = l_Lean_mkApp3(x_52, x_44, x_51, x_43); +x_54 = l_Fin_reduceBin___lambda__1___closed__7; +x_55 = l_Lean_mkApp3(x_54, x_46, x_43, x_53); lean_ctor_set_tag(x_25, 0); -lean_ctor_set(x_25, 0, x_53); +lean_ctor_set(x_25, 0, x_55); return x_24; } } else { -lean_object* x_54; lean_object* x_55; lean_object* x_56; uint8_t x_57; -x_54 = lean_ctor_get(x_25, 0); -lean_inc(x_54); -lean_dec(x_25); -x_55 = lean_ctor_get(x_22, 0); -lean_inc(x_55); -x_56 = lean_ctor_get(x_54, 0); +lean_object* x_56; lean_object* x_57; lean_object* x_58; uint8_t x_59; +x_56 = lean_ctor_get(x_25, 0); lean_inc(x_56); -x_57 = lean_nat_dec_eq(x_55, x_56); -lean_dec(x_56); -if (x_57 == 0) +lean_dec(x_25); +x_57 = lean_ctor_get(x_22, 0); +lean_inc(x_57); +x_58 = lean_ctor_get(x_56, 0); +lean_inc(x_58); +x_59 = lean_nat_dec_eq(x_57, x_58); +lean_dec(x_58); +if (x_59 == 0) { -lean_object* x_58; -lean_dec(x_55); -lean_dec(x_54); +lean_object* x_60; +lean_dec(x_57); +lean_dec(x_56); lean_dec(x_22); -x_58 = l_Fin_reduceBin___lambda__1___closed__1; -lean_ctor_set(x_24, 0, x_58); +x_60 = l_Fin_reduceBin___lambda__1___closed__1; +lean_ctor_set(x_24, 0, x_60); return x_24; } else { -lean_object* x_59; lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; -x_59 = lean_ctor_get(x_22, 1); -lean_inc(x_59); +lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; +x_61 = lean_ctor_get(x_22, 1); +lean_inc(x_61); lean_dec(x_22); -x_60 = lean_ctor_get(x_54, 1); -lean_inc(x_60); -lean_dec(x_54); -x_61 = l_Fin_mul(x_55, x_59, x_60); -lean_dec(x_60); -lean_dec(x_59); -x_62 = l_Lean_mkRawNatLit(x_61); -lean_inc(x_55); -x_63 = l_Lean_mkNatLit(x_55); -x_64 = l_Fin_reduceBin___lambda__1___closed__10; -x_65 = l_Lean_Expr_app___override(x_64, x_63); -x_66 = lean_unsigned_to_nat(1u); -x_67 = lean_nat_sub(x_55, x_66); -lean_dec(x_55); -x_68 = l_Lean_mkNatLit(x_67); -x_69 = l_Fin_reduceBin___lambda__1___closed__13; +x_62 = lean_ctor_get(x_56, 1); lean_inc(x_62); -x_70 = l_Lean_mkAppB(x_69, x_68, x_62); -x_71 = l_Fin_reduceBin___lambda__1___closed__7; -x_72 = l_Lean_mkApp3(x_71, x_65, x_62, x_70); -x_73 = lean_alloc_ctor(0, 1, 0); -lean_ctor_set(x_73, 0, x_72); -lean_ctor_set(x_24, 0, x_73); +lean_dec(x_56); +x_63 = l_Fin_mul(x_57, x_61, x_62); +lean_dec(x_62); +lean_dec(x_61); +x_64 = l_Lean_mkRawNatLit(x_63); +lean_inc(x_57); +x_65 = l_Lean_mkNatLit(x_57); +x_66 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_65); +x_67 = l_Lean_Expr_app___override(x_66, x_65); +x_68 = lean_unsigned_to_nat(1u); +x_69 = lean_nat_sub(x_57, x_68); +lean_dec(x_57); +x_70 = l_Lean_mkNatLit(x_69); +x_71 = l_Fin_reduceBin___lambda__1___closed__17; +x_72 = l_Lean_Expr_app___override(x_71, x_70); +x_73 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_64); +x_74 = l_Lean_mkApp3(x_73, x_65, x_72, x_64); +x_75 = l_Fin_reduceBin___lambda__1___closed__7; +x_76 = l_Lean_mkApp3(x_75, x_67, x_64, x_74); +x_77 = lean_alloc_ctor(0, 1, 0); +lean_ctor_set(x_77, 0, x_76); +lean_ctor_set(x_24, 0, x_77); return x_24; } } } else { -lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; lean_object* x_78; uint8_t x_79; -x_74 = lean_ctor_get(x_24, 1); -lean_inc(x_74); +lean_object* x_78; lean_object* x_79; lean_object* x_80; lean_object* x_81; lean_object* x_82; uint8_t x_83; +x_78 = lean_ctor_get(x_24, 1); +lean_inc(x_78); lean_dec(x_24); -x_75 = lean_ctor_get(x_25, 0); -lean_inc(x_75); +x_79 = lean_ctor_get(x_25, 0); +lean_inc(x_79); if (lean_is_exclusive(x_25)) { lean_ctor_release(x_25, 0); - x_76 = x_25; + x_80 = x_25; } else { lean_dec_ref(x_25); - x_76 = lean_box(0); + x_80 = lean_box(0); } -x_77 = lean_ctor_get(x_22, 0); -lean_inc(x_77); -x_78 = lean_ctor_get(x_75, 0); -lean_inc(x_78); -x_79 = lean_nat_dec_eq(x_77, x_78); -lean_dec(x_78); -if (x_79 == 0) -{ -lean_object* x_80; lean_object* x_81; -lean_dec(x_77); -lean_dec(x_76); -lean_dec(x_75); +x_81 = lean_ctor_get(x_22, 0); +lean_inc(x_81); +x_82 = lean_ctor_get(x_79, 0); +lean_inc(x_82); +x_83 = lean_nat_dec_eq(x_81, x_82); +lean_dec(x_82); +if (x_83 == 0) +{ +lean_object* x_84; lean_object* x_85; +lean_dec(x_81); +lean_dec(x_80); +lean_dec(x_79); lean_dec(x_22); -x_80 = l_Fin_reduceBin___lambda__1___closed__1; -x_81 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_81, 0, x_80); -lean_ctor_set(x_81, 1, x_74); -return x_81; +x_84 = l_Fin_reduceBin___lambda__1___closed__1; +x_85 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_85, 0, x_84); +lean_ctor_set(x_85, 1, x_78); +return x_85; } else { -lean_object* x_82; lean_object* x_83; lean_object* x_84; lean_object* x_85; lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; -x_82 = lean_ctor_get(x_22, 1); -lean_inc(x_82); +lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; lean_object* x_98; lean_object* x_99; lean_object* x_100; lean_object* x_101; lean_object* x_102; lean_object* x_103; +x_86 = lean_ctor_get(x_22, 1); +lean_inc(x_86); lean_dec(x_22); -x_83 = lean_ctor_get(x_75, 1); -lean_inc(x_83); -lean_dec(x_75); -x_84 = l_Fin_mul(x_77, x_82, x_83); -lean_dec(x_83); -lean_dec(x_82); -x_85 = l_Lean_mkRawNatLit(x_84); -lean_inc(x_77); -x_86 = l_Lean_mkNatLit(x_77); -x_87 = l_Fin_reduceBin___lambda__1___closed__10; -x_88 = l_Lean_Expr_app___override(x_87, x_86); -x_89 = lean_unsigned_to_nat(1u); -x_90 = lean_nat_sub(x_77, x_89); -lean_dec(x_77); -x_91 = l_Lean_mkNatLit(x_90); -x_92 = l_Fin_reduceBin___lambda__1___closed__13; -lean_inc(x_85); -x_93 = l_Lean_mkAppB(x_92, x_91, x_85); -x_94 = l_Fin_reduceBin___lambda__1___closed__7; -x_95 = l_Lean_mkApp3(x_94, x_88, x_85, x_93); -if (lean_is_scalar(x_76)) { - x_96 = lean_alloc_ctor(0, 1, 0); +x_87 = lean_ctor_get(x_79, 1); +lean_inc(x_87); +lean_dec(x_79); +x_88 = l_Fin_mul(x_81, x_86, x_87); +lean_dec(x_87); +lean_dec(x_86); +x_89 = l_Lean_mkRawNatLit(x_88); +lean_inc(x_81); +x_90 = l_Lean_mkNatLit(x_81); +x_91 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_90); +x_92 = l_Lean_Expr_app___override(x_91, x_90); +x_93 = lean_unsigned_to_nat(1u); +x_94 = lean_nat_sub(x_81, x_93); +lean_dec(x_81); +x_95 = l_Lean_mkNatLit(x_94); +x_96 = l_Fin_reduceBin___lambda__1___closed__17; +x_97 = l_Lean_Expr_app___override(x_96, x_95); +x_98 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_89); +x_99 = l_Lean_mkApp3(x_98, x_90, x_97, x_89); +x_100 = l_Fin_reduceBin___lambda__1___closed__7; +x_101 = l_Lean_mkApp3(x_100, x_92, x_89, x_99); +if (lean_is_scalar(x_80)) { + x_102 = lean_alloc_ctor(0, 1, 0); } else { - x_96 = x_76; - lean_ctor_set_tag(x_96, 0); + x_102 = x_80; + lean_ctor_set_tag(x_102, 0); } -lean_ctor_set(x_96, 0, x_95); -x_97 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_97, 0, x_96); -lean_ctor_set(x_97, 1, x_74); -return x_97; +lean_ctor_set(x_102, 0, x_101); +x_103 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_103, 0, x_102); +lean_ctor_set(x_103, 1, x_78); +return x_103; } } } } else { -uint8_t x_98; +uint8_t x_104; lean_dec(x_22); -x_98 = !lean_is_exclusive(x_24); -if (x_98 == 0) +x_104 = !lean_is_exclusive(x_24); +if (x_104 == 0) { return x_24; } else { -lean_object* x_99; lean_object* x_100; lean_object* x_101; -x_99 = lean_ctor_get(x_24, 0); -x_100 = lean_ctor_get(x_24, 1); -lean_inc(x_100); -lean_inc(x_99); +lean_object* x_105; lean_object* x_106; lean_object* x_107; +x_105 = lean_ctor_get(x_24, 0); +x_106 = lean_ctor_get(x_24, 1); +lean_inc(x_106); +lean_inc(x_105); lean_dec(x_24); -x_101 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_101, 0, x_99); -lean_ctor_set(x_101, 1, x_100); -return x_101; +x_107 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_107, 0, x_105); +lean_ctor_set(x_107, 1, x_106); +return x_107; } } } } else { -uint8_t x_102; +uint8_t x_108; lean_dec(x_9); lean_dec(x_8); lean_dec(x_7); lean_dec(x_6); -x_102 = !lean_is_exclusive(x_13); -if (x_102 == 0) +x_108 = !lean_is_exclusive(x_13); +if (x_108 == 0) { return x_13; } else { -lean_object* x_103; lean_object* x_104; lean_object* x_105; -x_103 = lean_ctor_get(x_13, 0); -x_104 = lean_ctor_get(x_13, 1); -lean_inc(x_104); -lean_inc(x_103); +lean_object* x_109; lean_object* x_110; lean_object* x_111; +x_109 = lean_ctor_get(x_13, 0); +x_110 = lean_ctor_get(x_13, 1); +lean_inc(x_110); +lean_inc(x_109); lean_dec(x_13); -x_105 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_105, 0, x_103); -lean_ctor_set(x_105, 1, x_104); -return x_105; +x_111 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_111, 0, x_109); +lean_ctor_set(x_111, 1, x_110); +return x_111; } } } @@ -3425,7 +3491,7 @@ return x_24; } else { -lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; +lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; lean_object* x_54; lean_object* x_55; x_40 = lean_ctor_get(x_22, 1); lean_inc(x_40); lean_dec(x_22); @@ -3439,200 +3505,209 @@ x_43 = l_Lean_mkRawNatLit(x_42); lean_inc(x_36); x_44 = l_Lean_mkNatLit(x_36); x_45 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_44); x_46 = l_Lean_Expr_app___override(x_45, x_44); x_47 = lean_unsigned_to_nat(1u); x_48 = lean_nat_sub(x_36, x_47); lean_dec(x_36); x_49 = l_Lean_mkNatLit(x_48); -x_50 = l_Fin_reduceBin___lambda__1___closed__13; +x_50 = l_Fin_reduceBin___lambda__1___closed__17; +x_51 = l_Lean_Expr_app___override(x_50, x_49); +x_52 = l_Fin_reduceBin___lambda__1___closed__13; lean_inc(x_43); -x_51 = l_Lean_mkAppB(x_50, x_49, x_43); -x_52 = l_Fin_reduceBin___lambda__1___closed__7; -x_53 = l_Lean_mkApp3(x_52, x_46, x_43, x_51); +x_53 = l_Lean_mkApp3(x_52, x_44, x_51, x_43); +x_54 = l_Fin_reduceBin___lambda__1___closed__7; +x_55 = l_Lean_mkApp3(x_54, x_46, x_43, x_53); lean_ctor_set_tag(x_25, 0); -lean_ctor_set(x_25, 0, x_53); +lean_ctor_set(x_25, 0, x_55); return x_24; } } else { -lean_object* x_54; lean_object* x_55; lean_object* x_56; uint8_t x_57; -x_54 = lean_ctor_get(x_25, 0); -lean_inc(x_54); -lean_dec(x_25); -x_55 = lean_ctor_get(x_22, 0); -lean_inc(x_55); -x_56 = lean_ctor_get(x_54, 0); +lean_object* x_56; lean_object* x_57; lean_object* x_58; uint8_t x_59; +x_56 = lean_ctor_get(x_25, 0); lean_inc(x_56); -x_57 = lean_nat_dec_eq(x_55, x_56); -lean_dec(x_56); -if (x_57 == 0) +lean_dec(x_25); +x_57 = lean_ctor_get(x_22, 0); +lean_inc(x_57); +x_58 = lean_ctor_get(x_56, 0); +lean_inc(x_58); +x_59 = lean_nat_dec_eq(x_57, x_58); +lean_dec(x_58); +if (x_59 == 0) { -lean_object* x_58; -lean_dec(x_55); -lean_dec(x_54); +lean_object* x_60; +lean_dec(x_57); +lean_dec(x_56); lean_dec(x_22); -x_58 = l_Fin_reduceBin___lambda__1___closed__1; -lean_ctor_set(x_24, 0, x_58); +x_60 = l_Fin_reduceBin___lambda__1___closed__1; +lean_ctor_set(x_24, 0, x_60); return x_24; } else { -lean_object* x_59; lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; -x_59 = lean_ctor_get(x_22, 1); -lean_inc(x_59); +lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; +x_61 = lean_ctor_get(x_22, 1); +lean_inc(x_61); lean_dec(x_22); -x_60 = lean_ctor_get(x_54, 1); -lean_inc(x_60); -lean_dec(x_54); -x_61 = l_Fin_sub(x_55, x_59, x_60); -lean_dec(x_60); -lean_dec(x_59); -x_62 = l_Lean_mkRawNatLit(x_61); -lean_inc(x_55); -x_63 = l_Lean_mkNatLit(x_55); -x_64 = l_Fin_reduceBin___lambda__1___closed__10; -x_65 = l_Lean_Expr_app___override(x_64, x_63); -x_66 = lean_unsigned_to_nat(1u); -x_67 = lean_nat_sub(x_55, x_66); -lean_dec(x_55); -x_68 = l_Lean_mkNatLit(x_67); -x_69 = l_Fin_reduceBin___lambda__1___closed__13; +x_62 = lean_ctor_get(x_56, 1); lean_inc(x_62); -x_70 = l_Lean_mkAppB(x_69, x_68, x_62); -x_71 = l_Fin_reduceBin___lambda__1___closed__7; -x_72 = l_Lean_mkApp3(x_71, x_65, x_62, x_70); -x_73 = lean_alloc_ctor(0, 1, 0); -lean_ctor_set(x_73, 0, x_72); -lean_ctor_set(x_24, 0, x_73); +lean_dec(x_56); +x_63 = l_Fin_sub(x_57, x_61, x_62); +lean_dec(x_62); +lean_dec(x_61); +x_64 = l_Lean_mkRawNatLit(x_63); +lean_inc(x_57); +x_65 = l_Lean_mkNatLit(x_57); +x_66 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_65); +x_67 = l_Lean_Expr_app___override(x_66, x_65); +x_68 = lean_unsigned_to_nat(1u); +x_69 = lean_nat_sub(x_57, x_68); +lean_dec(x_57); +x_70 = l_Lean_mkNatLit(x_69); +x_71 = l_Fin_reduceBin___lambda__1___closed__17; +x_72 = l_Lean_Expr_app___override(x_71, x_70); +x_73 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_64); +x_74 = l_Lean_mkApp3(x_73, x_65, x_72, x_64); +x_75 = l_Fin_reduceBin___lambda__1___closed__7; +x_76 = l_Lean_mkApp3(x_75, x_67, x_64, x_74); +x_77 = lean_alloc_ctor(0, 1, 0); +lean_ctor_set(x_77, 0, x_76); +lean_ctor_set(x_24, 0, x_77); return x_24; } } } else { -lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; lean_object* x_78; uint8_t x_79; -x_74 = lean_ctor_get(x_24, 1); -lean_inc(x_74); +lean_object* x_78; lean_object* x_79; lean_object* x_80; lean_object* x_81; lean_object* x_82; uint8_t x_83; +x_78 = lean_ctor_get(x_24, 1); +lean_inc(x_78); lean_dec(x_24); -x_75 = lean_ctor_get(x_25, 0); -lean_inc(x_75); +x_79 = lean_ctor_get(x_25, 0); +lean_inc(x_79); if (lean_is_exclusive(x_25)) { lean_ctor_release(x_25, 0); - x_76 = x_25; + x_80 = x_25; } else { lean_dec_ref(x_25); - x_76 = lean_box(0); + x_80 = lean_box(0); } -x_77 = lean_ctor_get(x_22, 0); -lean_inc(x_77); -x_78 = lean_ctor_get(x_75, 0); -lean_inc(x_78); -x_79 = lean_nat_dec_eq(x_77, x_78); -lean_dec(x_78); -if (x_79 == 0) -{ -lean_object* x_80; lean_object* x_81; -lean_dec(x_77); -lean_dec(x_76); -lean_dec(x_75); +x_81 = lean_ctor_get(x_22, 0); +lean_inc(x_81); +x_82 = lean_ctor_get(x_79, 0); +lean_inc(x_82); +x_83 = lean_nat_dec_eq(x_81, x_82); +lean_dec(x_82); +if (x_83 == 0) +{ +lean_object* x_84; lean_object* x_85; +lean_dec(x_81); +lean_dec(x_80); +lean_dec(x_79); lean_dec(x_22); -x_80 = l_Fin_reduceBin___lambda__1___closed__1; -x_81 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_81, 0, x_80); -lean_ctor_set(x_81, 1, x_74); -return x_81; +x_84 = l_Fin_reduceBin___lambda__1___closed__1; +x_85 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_85, 0, x_84); +lean_ctor_set(x_85, 1, x_78); +return x_85; } else { -lean_object* x_82; lean_object* x_83; lean_object* x_84; lean_object* x_85; lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; -x_82 = lean_ctor_get(x_22, 1); -lean_inc(x_82); +lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; lean_object* x_98; lean_object* x_99; lean_object* x_100; lean_object* x_101; lean_object* x_102; lean_object* x_103; +x_86 = lean_ctor_get(x_22, 1); +lean_inc(x_86); lean_dec(x_22); -x_83 = lean_ctor_get(x_75, 1); -lean_inc(x_83); -lean_dec(x_75); -x_84 = l_Fin_sub(x_77, x_82, x_83); -lean_dec(x_83); -lean_dec(x_82); -x_85 = l_Lean_mkRawNatLit(x_84); -lean_inc(x_77); -x_86 = l_Lean_mkNatLit(x_77); -x_87 = l_Fin_reduceBin___lambda__1___closed__10; -x_88 = l_Lean_Expr_app___override(x_87, x_86); -x_89 = lean_unsigned_to_nat(1u); -x_90 = lean_nat_sub(x_77, x_89); -lean_dec(x_77); -x_91 = l_Lean_mkNatLit(x_90); -x_92 = l_Fin_reduceBin___lambda__1___closed__13; -lean_inc(x_85); -x_93 = l_Lean_mkAppB(x_92, x_91, x_85); -x_94 = l_Fin_reduceBin___lambda__1___closed__7; -x_95 = l_Lean_mkApp3(x_94, x_88, x_85, x_93); -if (lean_is_scalar(x_76)) { - x_96 = lean_alloc_ctor(0, 1, 0); +x_87 = lean_ctor_get(x_79, 1); +lean_inc(x_87); +lean_dec(x_79); +x_88 = l_Fin_sub(x_81, x_86, x_87); +lean_dec(x_87); +lean_dec(x_86); +x_89 = l_Lean_mkRawNatLit(x_88); +lean_inc(x_81); +x_90 = l_Lean_mkNatLit(x_81); +x_91 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_90); +x_92 = l_Lean_Expr_app___override(x_91, x_90); +x_93 = lean_unsigned_to_nat(1u); +x_94 = lean_nat_sub(x_81, x_93); +lean_dec(x_81); +x_95 = l_Lean_mkNatLit(x_94); +x_96 = l_Fin_reduceBin___lambda__1___closed__17; +x_97 = l_Lean_Expr_app___override(x_96, x_95); +x_98 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_89); +x_99 = l_Lean_mkApp3(x_98, x_90, x_97, x_89); +x_100 = l_Fin_reduceBin___lambda__1___closed__7; +x_101 = l_Lean_mkApp3(x_100, x_92, x_89, x_99); +if (lean_is_scalar(x_80)) { + x_102 = lean_alloc_ctor(0, 1, 0); } else { - x_96 = x_76; - lean_ctor_set_tag(x_96, 0); + x_102 = x_80; + lean_ctor_set_tag(x_102, 0); } -lean_ctor_set(x_96, 0, x_95); -x_97 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_97, 0, x_96); -lean_ctor_set(x_97, 1, x_74); -return x_97; +lean_ctor_set(x_102, 0, x_101); +x_103 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_103, 0, x_102); +lean_ctor_set(x_103, 1, x_78); +return x_103; } } } } else { -uint8_t x_98; +uint8_t x_104; lean_dec(x_22); -x_98 = !lean_is_exclusive(x_24); -if (x_98 == 0) +x_104 = !lean_is_exclusive(x_24); +if (x_104 == 0) { return x_24; } else { -lean_object* x_99; lean_object* x_100; lean_object* x_101; -x_99 = lean_ctor_get(x_24, 0); -x_100 = lean_ctor_get(x_24, 1); -lean_inc(x_100); -lean_inc(x_99); +lean_object* x_105; lean_object* x_106; lean_object* x_107; +x_105 = lean_ctor_get(x_24, 0); +x_106 = lean_ctor_get(x_24, 1); +lean_inc(x_106); +lean_inc(x_105); lean_dec(x_24); -x_101 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_101, 0, x_99); -lean_ctor_set(x_101, 1, x_100); -return x_101; +x_107 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_107, 0, x_105); +lean_ctor_set(x_107, 1, x_106); +return x_107; } } } } else { -uint8_t x_102; +uint8_t x_108; lean_dec(x_9); lean_dec(x_8); lean_dec(x_7); lean_dec(x_6); -x_102 = !lean_is_exclusive(x_13); -if (x_102 == 0) +x_108 = !lean_is_exclusive(x_13); +if (x_108 == 0) { return x_13; } else { -lean_object* x_103; lean_object* x_104; lean_object* x_105; -x_103 = lean_ctor_get(x_13, 0); -x_104 = lean_ctor_get(x_13, 1); -lean_inc(x_104); -lean_inc(x_103); +lean_object* x_109; lean_object* x_110; lean_object* x_111; +x_109 = lean_ctor_get(x_13, 0); +x_110 = lean_ctor_get(x_13, 1); +lean_inc(x_110); +lean_inc(x_109); lean_dec(x_13); -x_105 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_105, 0, x_103); -lean_ctor_set(x_105, 1, x_104); -return x_105; +x_111 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_111, 0, x_109); +lean_ctor_set(x_111, 1, x_110); +return x_111; } } } @@ -4023,7 +4098,7 @@ return x_24; } else { -lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; +lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; lean_object* x_54; lean_object* x_55; x_40 = lean_ctor_get(x_22, 1); lean_inc(x_40); lean_dec(x_22); @@ -4037,200 +4112,209 @@ x_43 = l_Lean_mkRawNatLit(x_42); lean_inc(x_36); x_44 = l_Lean_mkNatLit(x_36); x_45 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_44); x_46 = l_Lean_Expr_app___override(x_45, x_44); x_47 = lean_unsigned_to_nat(1u); x_48 = lean_nat_sub(x_36, x_47); lean_dec(x_36); x_49 = l_Lean_mkNatLit(x_48); -x_50 = l_Fin_reduceBin___lambda__1___closed__13; +x_50 = l_Fin_reduceBin___lambda__1___closed__17; +x_51 = l_Lean_Expr_app___override(x_50, x_49); +x_52 = l_Fin_reduceBin___lambda__1___closed__13; lean_inc(x_43); -x_51 = l_Lean_mkAppB(x_50, x_49, x_43); -x_52 = l_Fin_reduceBin___lambda__1___closed__7; -x_53 = l_Lean_mkApp3(x_52, x_46, x_43, x_51); +x_53 = l_Lean_mkApp3(x_52, x_44, x_51, x_43); +x_54 = l_Fin_reduceBin___lambda__1___closed__7; +x_55 = l_Lean_mkApp3(x_54, x_46, x_43, x_53); lean_ctor_set_tag(x_25, 0); -lean_ctor_set(x_25, 0, x_53); +lean_ctor_set(x_25, 0, x_55); return x_24; } } else { -lean_object* x_54; lean_object* x_55; lean_object* x_56; uint8_t x_57; -x_54 = lean_ctor_get(x_25, 0); -lean_inc(x_54); -lean_dec(x_25); -x_55 = lean_ctor_get(x_22, 0); -lean_inc(x_55); -x_56 = lean_ctor_get(x_54, 0); +lean_object* x_56; lean_object* x_57; lean_object* x_58; uint8_t x_59; +x_56 = lean_ctor_get(x_25, 0); lean_inc(x_56); -x_57 = lean_nat_dec_eq(x_55, x_56); -lean_dec(x_56); -if (x_57 == 0) +lean_dec(x_25); +x_57 = lean_ctor_get(x_22, 0); +lean_inc(x_57); +x_58 = lean_ctor_get(x_56, 0); +lean_inc(x_58); +x_59 = lean_nat_dec_eq(x_57, x_58); +lean_dec(x_58); +if (x_59 == 0) { -lean_object* x_58; -lean_dec(x_55); -lean_dec(x_54); +lean_object* x_60; +lean_dec(x_57); +lean_dec(x_56); lean_dec(x_22); -x_58 = l_Fin_reduceBin___lambda__1___closed__1; -lean_ctor_set(x_24, 0, x_58); +x_60 = l_Fin_reduceBin___lambda__1___closed__1; +lean_ctor_set(x_24, 0, x_60); return x_24; } else { -lean_object* x_59; lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; -x_59 = lean_ctor_get(x_22, 1); -lean_inc(x_59); +lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; +x_61 = lean_ctor_get(x_22, 1); +lean_inc(x_61); lean_dec(x_22); -x_60 = lean_ctor_get(x_54, 1); -lean_inc(x_60); -lean_dec(x_54); -x_61 = lean_nat_div(x_59, x_60); -lean_dec(x_60); -lean_dec(x_59); -x_62 = l_Lean_mkRawNatLit(x_61); -lean_inc(x_55); -x_63 = l_Lean_mkNatLit(x_55); -x_64 = l_Fin_reduceBin___lambda__1___closed__10; -x_65 = l_Lean_Expr_app___override(x_64, x_63); -x_66 = lean_unsigned_to_nat(1u); -x_67 = lean_nat_sub(x_55, x_66); -lean_dec(x_55); -x_68 = l_Lean_mkNatLit(x_67); -x_69 = l_Fin_reduceBin___lambda__1___closed__13; +x_62 = lean_ctor_get(x_56, 1); lean_inc(x_62); -x_70 = l_Lean_mkAppB(x_69, x_68, x_62); -x_71 = l_Fin_reduceBin___lambda__1___closed__7; -x_72 = l_Lean_mkApp3(x_71, x_65, x_62, x_70); -x_73 = lean_alloc_ctor(0, 1, 0); -lean_ctor_set(x_73, 0, x_72); -lean_ctor_set(x_24, 0, x_73); +lean_dec(x_56); +x_63 = lean_nat_div(x_61, x_62); +lean_dec(x_62); +lean_dec(x_61); +x_64 = l_Lean_mkRawNatLit(x_63); +lean_inc(x_57); +x_65 = l_Lean_mkNatLit(x_57); +x_66 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_65); +x_67 = l_Lean_Expr_app___override(x_66, x_65); +x_68 = lean_unsigned_to_nat(1u); +x_69 = lean_nat_sub(x_57, x_68); +lean_dec(x_57); +x_70 = l_Lean_mkNatLit(x_69); +x_71 = l_Fin_reduceBin___lambda__1___closed__17; +x_72 = l_Lean_Expr_app___override(x_71, x_70); +x_73 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_64); +x_74 = l_Lean_mkApp3(x_73, x_65, x_72, x_64); +x_75 = l_Fin_reduceBin___lambda__1___closed__7; +x_76 = l_Lean_mkApp3(x_75, x_67, x_64, x_74); +x_77 = lean_alloc_ctor(0, 1, 0); +lean_ctor_set(x_77, 0, x_76); +lean_ctor_set(x_24, 0, x_77); return x_24; } } } else { -lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; lean_object* x_78; uint8_t x_79; -x_74 = lean_ctor_get(x_24, 1); -lean_inc(x_74); +lean_object* x_78; lean_object* x_79; lean_object* x_80; lean_object* x_81; lean_object* x_82; uint8_t x_83; +x_78 = lean_ctor_get(x_24, 1); +lean_inc(x_78); lean_dec(x_24); -x_75 = lean_ctor_get(x_25, 0); -lean_inc(x_75); +x_79 = lean_ctor_get(x_25, 0); +lean_inc(x_79); if (lean_is_exclusive(x_25)) { lean_ctor_release(x_25, 0); - x_76 = x_25; + x_80 = x_25; } else { lean_dec_ref(x_25); - x_76 = lean_box(0); + x_80 = lean_box(0); } -x_77 = lean_ctor_get(x_22, 0); -lean_inc(x_77); -x_78 = lean_ctor_get(x_75, 0); -lean_inc(x_78); -x_79 = lean_nat_dec_eq(x_77, x_78); -lean_dec(x_78); -if (x_79 == 0) -{ -lean_object* x_80; lean_object* x_81; -lean_dec(x_77); -lean_dec(x_76); -lean_dec(x_75); +x_81 = lean_ctor_get(x_22, 0); +lean_inc(x_81); +x_82 = lean_ctor_get(x_79, 0); +lean_inc(x_82); +x_83 = lean_nat_dec_eq(x_81, x_82); +lean_dec(x_82); +if (x_83 == 0) +{ +lean_object* x_84; lean_object* x_85; +lean_dec(x_81); +lean_dec(x_80); +lean_dec(x_79); lean_dec(x_22); -x_80 = l_Fin_reduceBin___lambda__1___closed__1; -x_81 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_81, 0, x_80); -lean_ctor_set(x_81, 1, x_74); -return x_81; +x_84 = l_Fin_reduceBin___lambda__1___closed__1; +x_85 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_85, 0, x_84); +lean_ctor_set(x_85, 1, x_78); +return x_85; } else { -lean_object* x_82; lean_object* x_83; lean_object* x_84; lean_object* x_85; lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; -x_82 = lean_ctor_get(x_22, 1); -lean_inc(x_82); +lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; lean_object* x_98; lean_object* x_99; lean_object* x_100; lean_object* x_101; lean_object* x_102; lean_object* x_103; +x_86 = lean_ctor_get(x_22, 1); +lean_inc(x_86); lean_dec(x_22); -x_83 = lean_ctor_get(x_75, 1); -lean_inc(x_83); -lean_dec(x_75); -x_84 = lean_nat_div(x_82, x_83); -lean_dec(x_83); -lean_dec(x_82); -x_85 = l_Lean_mkRawNatLit(x_84); -lean_inc(x_77); -x_86 = l_Lean_mkNatLit(x_77); -x_87 = l_Fin_reduceBin___lambda__1___closed__10; -x_88 = l_Lean_Expr_app___override(x_87, x_86); -x_89 = lean_unsigned_to_nat(1u); -x_90 = lean_nat_sub(x_77, x_89); -lean_dec(x_77); -x_91 = l_Lean_mkNatLit(x_90); -x_92 = l_Fin_reduceBin___lambda__1___closed__13; -lean_inc(x_85); -x_93 = l_Lean_mkAppB(x_92, x_91, x_85); -x_94 = l_Fin_reduceBin___lambda__1___closed__7; -x_95 = l_Lean_mkApp3(x_94, x_88, x_85, x_93); -if (lean_is_scalar(x_76)) { - x_96 = lean_alloc_ctor(0, 1, 0); +x_87 = lean_ctor_get(x_79, 1); +lean_inc(x_87); +lean_dec(x_79); +x_88 = lean_nat_div(x_86, x_87); +lean_dec(x_87); +lean_dec(x_86); +x_89 = l_Lean_mkRawNatLit(x_88); +lean_inc(x_81); +x_90 = l_Lean_mkNatLit(x_81); +x_91 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_90); +x_92 = l_Lean_Expr_app___override(x_91, x_90); +x_93 = lean_unsigned_to_nat(1u); +x_94 = lean_nat_sub(x_81, x_93); +lean_dec(x_81); +x_95 = l_Lean_mkNatLit(x_94); +x_96 = l_Fin_reduceBin___lambda__1___closed__17; +x_97 = l_Lean_Expr_app___override(x_96, x_95); +x_98 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_89); +x_99 = l_Lean_mkApp3(x_98, x_90, x_97, x_89); +x_100 = l_Fin_reduceBin___lambda__1___closed__7; +x_101 = l_Lean_mkApp3(x_100, x_92, x_89, x_99); +if (lean_is_scalar(x_80)) { + x_102 = lean_alloc_ctor(0, 1, 0); } else { - x_96 = x_76; - lean_ctor_set_tag(x_96, 0); + x_102 = x_80; + lean_ctor_set_tag(x_102, 0); } -lean_ctor_set(x_96, 0, x_95); -x_97 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_97, 0, x_96); -lean_ctor_set(x_97, 1, x_74); -return x_97; +lean_ctor_set(x_102, 0, x_101); +x_103 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_103, 0, x_102); +lean_ctor_set(x_103, 1, x_78); +return x_103; } } } } else { -uint8_t x_98; +uint8_t x_104; lean_dec(x_22); -x_98 = !lean_is_exclusive(x_24); -if (x_98 == 0) +x_104 = !lean_is_exclusive(x_24); +if (x_104 == 0) { return x_24; } else { -lean_object* x_99; lean_object* x_100; lean_object* x_101; -x_99 = lean_ctor_get(x_24, 0); -x_100 = lean_ctor_get(x_24, 1); -lean_inc(x_100); -lean_inc(x_99); +lean_object* x_105; lean_object* x_106; lean_object* x_107; +x_105 = lean_ctor_get(x_24, 0); +x_106 = lean_ctor_get(x_24, 1); +lean_inc(x_106); +lean_inc(x_105); lean_dec(x_24); -x_101 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_101, 0, x_99); -lean_ctor_set(x_101, 1, x_100); -return x_101; +x_107 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_107, 0, x_105); +lean_ctor_set(x_107, 1, x_106); +return x_107; } } } } else { -uint8_t x_102; +uint8_t x_108; lean_dec(x_9); lean_dec(x_8); lean_dec(x_7); lean_dec(x_6); -x_102 = !lean_is_exclusive(x_13); -if (x_102 == 0) +x_108 = !lean_is_exclusive(x_13); +if (x_108 == 0) { return x_13; } else { -lean_object* x_103; lean_object* x_104; lean_object* x_105; -x_103 = lean_ctor_get(x_13, 0); -x_104 = lean_ctor_get(x_13, 1); -lean_inc(x_104); -lean_inc(x_103); +lean_object* x_109; lean_object* x_110; lean_object* x_111; +x_109 = lean_ctor_get(x_13, 0); +x_110 = lean_ctor_get(x_13, 1); +lean_inc(x_110); +lean_inc(x_109); lean_dec(x_13); -x_105 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_105, 0, x_103); -lean_ctor_set(x_105, 1, x_104); -return x_105; +x_111 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_111, 0, x_109); +lean_ctor_set(x_111, 1, x_110); +return x_111; } } } @@ -4621,7 +4705,7 @@ return x_24; } else { -lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; +lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; lean_object* x_50; lean_object* x_51; lean_object* x_52; lean_object* x_53; lean_object* x_54; lean_object* x_55; x_40 = lean_ctor_get(x_22, 1); lean_inc(x_40); lean_dec(x_22); @@ -4635,200 +4719,209 @@ x_43 = l_Lean_mkRawNatLit(x_42); lean_inc(x_36); x_44 = l_Lean_mkNatLit(x_36); x_45 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_44); x_46 = l_Lean_Expr_app___override(x_45, x_44); x_47 = lean_unsigned_to_nat(1u); x_48 = lean_nat_sub(x_36, x_47); lean_dec(x_36); x_49 = l_Lean_mkNatLit(x_48); -x_50 = l_Fin_reduceBin___lambda__1___closed__13; +x_50 = l_Fin_reduceBin___lambda__1___closed__17; +x_51 = l_Lean_Expr_app___override(x_50, x_49); +x_52 = l_Fin_reduceBin___lambda__1___closed__13; lean_inc(x_43); -x_51 = l_Lean_mkAppB(x_50, x_49, x_43); -x_52 = l_Fin_reduceBin___lambda__1___closed__7; -x_53 = l_Lean_mkApp3(x_52, x_46, x_43, x_51); +x_53 = l_Lean_mkApp3(x_52, x_44, x_51, x_43); +x_54 = l_Fin_reduceBin___lambda__1___closed__7; +x_55 = l_Lean_mkApp3(x_54, x_46, x_43, x_53); lean_ctor_set_tag(x_25, 0); -lean_ctor_set(x_25, 0, x_53); +lean_ctor_set(x_25, 0, x_55); return x_24; } } else { -lean_object* x_54; lean_object* x_55; lean_object* x_56; uint8_t x_57; -x_54 = lean_ctor_get(x_25, 0); -lean_inc(x_54); -lean_dec(x_25); -x_55 = lean_ctor_get(x_22, 0); -lean_inc(x_55); -x_56 = lean_ctor_get(x_54, 0); +lean_object* x_56; lean_object* x_57; lean_object* x_58; uint8_t x_59; +x_56 = lean_ctor_get(x_25, 0); lean_inc(x_56); -x_57 = lean_nat_dec_eq(x_55, x_56); -lean_dec(x_56); -if (x_57 == 0) +lean_dec(x_25); +x_57 = lean_ctor_get(x_22, 0); +lean_inc(x_57); +x_58 = lean_ctor_get(x_56, 0); +lean_inc(x_58); +x_59 = lean_nat_dec_eq(x_57, x_58); +lean_dec(x_58); +if (x_59 == 0) { -lean_object* x_58; -lean_dec(x_55); -lean_dec(x_54); +lean_object* x_60; +lean_dec(x_57); +lean_dec(x_56); lean_dec(x_22); -x_58 = l_Fin_reduceBin___lambda__1___closed__1; -lean_ctor_set(x_24, 0, x_58); +x_60 = l_Fin_reduceBin___lambda__1___closed__1; +lean_ctor_set(x_24, 0, x_60); return x_24; } else { -lean_object* x_59; lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; -x_59 = lean_ctor_get(x_22, 1); -lean_inc(x_59); +lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; lean_object* x_73; lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; +x_61 = lean_ctor_get(x_22, 1); +lean_inc(x_61); lean_dec(x_22); -x_60 = lean_ctor_get(x_54, 1); -lean_inc(x_60); -lean_dec(x_54); -x_61 = lean_nat_mod(x_59, x_60); -lean_dec(x_60); -lean_dec(x_59); -x_62 = l_Lean_mkRawNatLit(x_61); -lean_inc(x_55); -x_63 = l_Lean_mkNatLit(x_55); -x_64 = l_Fin_reduceBin___lambda__1___closed__10; -x_65 = l_Lean_Expr_app___override(x_64, x_63); -x_66 = lean_unsigned_to_nat(1u); -x_67 = lean_nat_sub(x_55, x_66); -lean_dec(x_55); -x_68 = l_Lean_mkNatLit(x_67); -x_69 = l_Fin_reduceBin___lambda__1___closed__13; +x_62 = lean_ctor_get(x_56, 1); lean_inc(x_62); -x_70 = l_Lean_mkAppB(x_69, x_68, x_62); -x_71 = l_Fin_reduceBin___lambda__1___closed__7; -x_72 = l_Lean_mkApp3(x_71, x_65, x_62, x_70); -x_73 = lean_alloc_ctor(0, 1, 0); -lean_ctor_set(x_73, 0, x_72); -lean_ctor_set(x_24, 0, x_73); +lean_dec(x_56); +x_63 = lean_nat_mod(x_61, x_62); +lean_dec(x_62); +lean_dec(x_61); +x_64 = l_Lean_mkRawNatLit(x_63); +lean_inc(x_57); +x_65 = l_Lean_mkNatLit(x_57); +x_66 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_65); +x_67 = l_Lean_Expr_app___override(x_66, x_65); +x_68 = lean_unsigned_to_nat(1u); +x_69 = lean_nat_sub(x_57, x_68); +lean_dec(x_57); +x_70 = l_Lean_mkNatLit(x_69); +x_71 = l_Fin_reduceBin___lambda__1___closed__17; +x_72 = l_Lean_Expr_app___override(x_71, x_70); +x_73 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_64); +x_74 = l_Lean_mkApp3(x_73, x_65, x_72, x_64); +x_75 = l_Fin_reduceBin___lambda__1___closed__7; +x_76 = l_Lean_mkApp3(x_75, x_67, x_64, x_74); +x_77 = lean_alloc_ctor(0, 1, 0); +lean_ctor_set(x_77, 0, x_76); +lean_ctor_set(x_24, 0, x_77); return x_24; } } } else { -lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; lean_object* x_78; uint8_t x_79; -x_74 = lean_ctor_get(x_24, 1); -lean_inc(x_74); +lean_object* x_78; lean_object* x_79; lean_object* x_80; lean_object* x_81; lean_object* x_82; uint8_t x_83; +x_78 = lean_ctor_get(x_24, 1); +lean_inc(x_78); lean_dec(x_24); -x_75 = lean_ctor_get(x_25, 0); -lean_inc(x_75); +x_79 = lean_ctor_get(x_25, 0); +lean_inc(x_79); if (lean_is_exclusive(x_25)) { lean_ctor_release(x_25, 0); - x_76 = x_25; + x_80 = x_25; } else { lean_dec_ref(x_25); - x_76 = lean_box(0); + x_80 = lean_box(0); } -x_77 = lean_ctor_get(x_22, 0); -lean_inc(x_77); -x_78 = lean_ctor_get(x_75, 0); -lean_inc(x_78); -x_79 = lean_nat_dec_eq(x_77, x_78); -lean_dec(x_78); -if (x_79 == 0) -{ -lean_object* x_80; lean_object* x_81; -lean_dec(x_77); -lean_dec(x_76); -lean_dec(x_75); +x_81 = lean_ctor_get(x_22, 0); +lean_inc(x_81); +x_82 = lean_ctor_get(x_79, 0); +lean_inc(x_82); +x_83 = lean_nat_dec_eq(x_81, x_82); +lean_dec(x_82); +if (x_83 == 0) +{ +lean_object* x_84; lean_object* x_85; +lean_dec(x_81); +lean_dec(x_80); +lean_dec(x_79); lean_dec(x_22); -x_80 = l_Fin_reduceBin___lambda__1___closed__1; -x_81 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_81, 0, x_80); -lean_ctor_set(x_81, 1, x_74); -return x_81; +x_84 = l_Fin_reduceBin___lambda__1___closed__1; +x_85 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_85, 0, x_84); +lean_ctor_set(x_85, 1, x_78); +return x_85; } else { -lean_object* x_82; lean_object* x_83; lean_object* x_84; lean_object* x_85; lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; -x_82 = lean_ctor_get(x_22, 1); -lean_inc(x_82); +lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; lean_object* x_92; lean_object* x_93; lean_object* x_94; lean_object* x_95; lean_object* x_96; lean_object* x_97; lean_object* x_98; lean_object* x_99; lean_object* x_100; lean_object* x_101; lean_object* x_102; lean_object* x_103; +x_86 = lean_ctor_get(x_22, 1); +lean_inc(x_86); lean_dec(x_22); -x_83 = lean_ctor_get(x_75, 1); -lean_inc(x_83); -lean_dec(x_75); -x_84 = lean_nat_mod(x_82, x_83); -lean_dec(x_83); -lean_dec(x_82); -x_85 = l_Lean_mkRawNatLit(x_84); -lean_inc(x_77); -x_86 = l_Lean_mkNatLit(x_77); -x_87 = l_Fin_reduceBin___lambda__1___closed__10; -x_88 = l_Lean_Expr_app___override(x_87, x_86); -x_89 = lean_unsigned_to_nat(1u); -x_90 = lean_nat_sub(x_77, x_89); -lean_dec(x_77); -x_91 = l_Lean_mkNatLit(x_90); -x_92 = l_Fin_reduceBin___lambda__1___closed__13; -lean_inc(x_85); -x_93 = l_Lean_mkAppB(x_92, x_91, x_85); -x_94 = l_Fin_reduceBin___lambda__1___closed__7; -x_95 = l_Lean_mkApp3(x_94, x_88, x_85, x_93); -if (lean_is_scalar(x_76)) { - x_96 = lean_alloc_ctor(0, 1, 0); +x_87 = lean_ctor_get(x_79, 1); +lean_inc(x_87); +lean_dec(x_79); +x_88 = lean_nat_mod(x_86, x_87); +lean_dec(x_87); +lean_dec(x_86); +x_89 = l_Lean_mkRawNatLit(x_88); +lean_inc(x_81); +x_90 = l_Lean_mkNatLit(x_81); +x_91 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_90); +x_92 = l_Lean_Expr_app___override(x_91, x_90); +x_93 = lean_unsigned_to_nat(1u); +x_94 = lean_nat_sub(x_81, x_93); +lean_dec(x_81); +x_95 = l_Lean_mkNatLit(x_94); +x_96 = l_Fin_reduceBin___lambda__1___closed__17; +x_97 = l_Lean_Expr_app___override(x_96, x_95); +x_98 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_89); +x_99 = l_Lean_mkApp3(x_98, x_90, x_97, x_89); +x_100 = l_Fin_reduceBin___lambda__1___closed__7; +x_101 = l_Lean_mkApp3(x_100, x_92, x_89, x_99); +if (lean_is_scalar(x_80)) { + x_102 = lean_alloc_ctor(0, 1, 0); } else { - x_96 = x_76; - lean_ctor_set_tag(x_96, 0); + x_102 = x_80; + lean_ctor_set_tag(x_102, 0); } -lean_ctor_set(x_96, 0, x_95); -x_97 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_97, 0, x_96); -lean_ctor_set(x_97, 1, x_74); -return x_97; +lean_ctor_set(x_102, 0, x_101); +x_103 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_103, 0, x_102); +lean_ctor_set(x_103, 1, x_78); +return x_103; } } } } else { -uint8_t x_98; +uint8_t x_104; lean_dec(x_22); -x_98 = !lean_is_exclusive(x_24); -if (x_98 == 0) +x_104 = !lean_is_exclusive(x_24); +if (x_104 == 0) { return x_24; } else { -lean_object* x_99; lean_object* x_100; lean_object* x_101; -x_99 = lean_ctor_get(x_24, 0); -x_100 = lean_ctor_get(x_24, 1); -lean_inc(x_100); -lean_inc(x_99); +lean_object* x_105; lean_object* x_106; lean_object* x_107; +x_105 = lean_ctor_get(x_24, 0); +x_106 = lean_ctor_get(x_24, 1); +lean_inc(x_106); +lean_inc(x_105); lean_dec(x_24); -x_101 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_101, 0, x_99); -lean_ctor_set(x_101, 1, x_100); -return x_101; +x_107 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_107, 0, x_105); +lean_ctor_set(x_107, 1, x_106); +return x_107; } } } } else { -uint8_t x_102; +uint8_t x_108; lean_dec(x_9); lean_dec(x_8); lean_dec(x_7); lean_dec(x_6); -x_102 = !lean_is_exclusive(x_13); -if (x_102 == 0) +x_108 = !lean_is_exclusive(x_13); +if (x_108 == 0) { return x_13; } else { -lean_object* x_103; lean_object* x_104; lean_object* x_105; -x_103 = lean_ctor_get(x_13, 0); -x_104 = lean_ctor_get(x_13, 1); -lean_inc(x_104); -lean_inc(x_103); +lean_object* x_109; lean_object* x_110; lean_object* x_111; +x_109 = lean_ctor_get(x_13, 0); +x_110 = lean_ctor_get(x_13, 1); +lean_inc(x_110); +lean_inc(x_109); lean_dec(x_13); -x_105 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_105, 0, x_103); -lean_ctor_set(x_105, 1, x_104); -return x_105; +x_111 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_111, 0, x_109); +lean_ctor_set(x_111, 1, x_110); +return x_111; } } } @@ -8258,27 +8351,30 @@ return x_6; LEAN_EXPORT lean_object* l_Fin_isValue___lambda__1(lean_object* x_1, lean_object* x_2, lean_object* x_3, lean_object* x_4, lean_object* x_5, lean_object* x_6, lean_object* x_7, lean_object* x_8, lean_object* x_9, lean_object* x_10, lean_object* x_11) { _start: { -lean_object* x_12; lean_object* x_13; lean_object* x_14; lean_object* x_15; lean_object* x_16; lean_object* x_17; lean_object* x_18; lean_object* x_19; lean_object* x_20; lean_object* x_21; lean_object* x_22; lean_object* x_23; lean_object* x_24; +lean_object* x_12; lean_object* x_13; lean_object* x_14; lean_object* x_15; lean_object* x_16; lean_object* x_17; lean_object* x_18; lean_object* x_19; lean_object* x_20; lean_object* x_21; lean_object* x_22; lean_object* x_23; lean_object* x_24; lean_object* x_25; lean_object* x_26; x_12 = l_Lean_mkRawNatLit(x_1); lean_inc(x_2); x_13 = l_Lean_mkNatLit(x_2); x_14 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_13); x_15 = l_Lean_Expr_app___override(x_14, x_13); x_16 = lean_unsigned_to_nat(1u); x_17 = lean_nat_sub(x_2, x_16); lean_dec(x_2); x_18 = l_Lean_mkNatLit(x_17); -x_19 = l_Fin_reduceBin___lambda__1___closed__13; +x_19 = l_Fin_reduceBin___lambda__1___closed__17; +x_20 = l_Lean_Expr_app___override(x_19, x_18); +x_21 = l_Fin_reduceBin___lambda__1___closed__13; lean_inc(x_12); -x_20 = l_Lean_mkAppB(x_19, x_18, x_12); -x_21 = l_Fin_reduceBin___lambda__1___closed__7; -x_22 = l_Lean_mkApp3(x_21, x_15, x_12, x_20); -x_23 = lean_alloc_ctor(0, 1, 0); -lean_ctor_set(x_23, 0, x_22); -x_24 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_24, 0, x_23); -lean_ctor_set(x_24, 1, x_11); -return x_24; +x_22 = l_Lean_mkApp3(x_21, x_13, x_20, x_12); +x_23 = l_Fin_reduceBin___lambda__1___closed__7; +x_24 = l_Lean_mkApp3(x_23, x_15, x_12, x_22); +x_25 = lean_alloc_ctor(0, 1, 0); +lean_ctor_set(x_25, 0, x_24); +x_26 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_26, 0, x_25); +lean_ctor_set(x_26, 1, x_11); +return x_26; } } LEAN_EXPORT lean_object* l_Fin_isValue___lambda__2(lean_object* x_1, lean_object* x_2, lean_object* x_3, lean_object* x_4, lean_object* x_5, lean_object* x_6, lean_object* x_7, lean_object* x_8, lean_object* x_9, lean_object* x_10) { @@ -8973,183 +9069,192 @@ return x_21; } else { -lean_object* x_36; lean_object* x_37; lean_object* x_38; lean_object* x_39; lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; +lean_object* x_36; lean_object* x_37; lean_object* x_38; lean_object* x_39; lean_object* x_40; lean_object* x_41; lean_object* x_42; lean_object* x_43; lean_object* x_44; lean_object* x_45; lean_object* x_46; lean_object* x_47; lean_object* x_48; lean_object* x_49; x_36 = lean_nat_mod(x_32, x_20); lean_dec(x_32); x_37 = l_Lean_mkRawNatLit(x_36); lean_inc(x_20); x_38 = l_Lean_mkNatLit(x_20); x_39 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_38); x_40 = l_Lean_Expr_app___override(x_39, x_38); x_41 = lean_unsigned_to_nat(1u); x_42 = lean_nat_sub(x_20, x_41); lean_dec(x_20); x_43 = l_Lean_mkNatLit(x_42); -x_44 = l_Fin_reduceBin___lambda__1___closed__13; +x_44 = l_Fin_reduceBin___lambda__1___closed__17; +x_45 = l_Lean_Expr_app___override(x_44, x_43); +x_46 = l_Fin_reduceBin___lambda__1___closed__13; lean_inc(x_37); -x_45 = l_Lean_mkAppB(x_44, x_43, x_37); -x_46 = l_Fin_reduceBin___lambda__1___closed__7; -x_47 = l_Lean_mkApp3(x_46, x_40, x_37, x_45); +x_47 = l_Lean_mkApp3(x_46, x_38, x_45, x_37); +x_48 = l_Fin_reduceBin___lambda__1___closed__7; +x_49 = l_Lean_mkApp3(x_48, x_40, x_37, x_47); lean_ctor_set_tag(x_22, 0); -lean_ctor_set(x_22, 0, x_47); +lean_ctor_set(x_22, 0, x_49); return x_21; } } else { -lean_object* x_48; lean_object* x_49; uint8_t x_50; -x_48 = lean_ctor_get(x_22, 0); -lean_inc(x_48); +lean_object* x_50; lean_object* x_51; uint8_t x_52; +x_50 = lean_ctor_get(x_22, 0); +lean_inc(x_50); lean_dec(x_22); -x_49 = lean_unsigned_to_nat(0u); -x_50 = lean_nat_dec_lt(x_49, x_20); -if (x_50 == 0) +x_51 = lean_unsigned_to_nat(0u); +x_52 = lean_nat_dec_lt(x_51, x_20); +if (x_52 == 0) { -lean_object* x_51; -lean_dec(x_48); +lean_object* x_53; +lean_dec(x_50); lean_dec(x_20); -x_51 = l_Fin_reduceBin___lambda__1___closed__1; -lean_ctor_set(x_21, 0, x_51); +x_53 = l_Fin_reduceBin___lambda__1___closed__1; +lean_ctor_set(x_21, 0, x_53); return x_21; } else { -lean_object* x_52; lean_object* x_53; lean_object* x_54; lean_object* x_55; lean_object* x_56; lean_object* x_57; lean_object* x_58; lean_object* x_59; lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; -x_52 = lean_nat_mod(x_48, x_20); -lean_dec(x_48); -x_53 = l_Lean_mkRawNatLit(x_52); +lean_object* x_54; lean_object* x_55; lean_object* x_56; lean_object* x_57; lean_object* x_58; lean_object* x_59; lean_object* x_60; lean_object* x_61; lean_object* x_62; lean_object* x_63; lean_object* x_64; lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; +x_54 = lean_nat_mod(x_50, x_20); +lean_dec(x_50); +x_55 = l_Lean_mkRawNatLit(x_54); lean_inc(x_20); -x_54 = l_Lean_mkNatLit(x_20); -x_55 = l_Fin_reduceBin___lambda__1___closed__10; -x_56 = l_Lean_Expr_app___override(x_55, x_54); -x_57 = lean_unsigned_to_nat(1u); -x_58 = lean_nat_sub(x_20, x_57); +x_56 = l_Lean_mkNatLit(x_20); +x_57 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_56); +x_58 = l_Lean_Expr_app___override(x_57, x_56); +x_59 = lean_unsigned_to_nat(1u); +x_60 = lean_nat_sub(x_20, x_59); lean_dec(x_20); -x_59 = l_Lean_mkNatLit(x_58); -x_60 = l_Fin_reduceBin___lambda__1___closed__13; -lean_inc(x_53); -x_61 = l_Lean_mkAppB(x_60, x_59, x_53); -x_62 = l_Fin_reduceBin___lambda__1___closed__7; -x_63 = l_Lean_mkApp3(x_62, x_56, x_53, x_61); -x_64 = lean_alloc_ctor(0, 1, 0); -lean_ctor_set(x_64, 0, x_63); -lean_ctor_set(x_21, 0, x_64); +x_61 = l_Lean_mkNatLit(x_60); +x_62 = l_Fin_reduceBin___lambda__1___closed__17; +x_63 = l_Lean_Expr_app___override(x_62, x_61); +x_64 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_55); +x_65 = l_Lean_mkApp3(x_64, x_56, x_63, x_55); +x_66 = l_Fin_reduceBin___lambda__1___closed__7; +x_67 = l_Lean_mkApp3(x_66, x_58, x_55, x_65); +x_68 = lean_alloc_ctor(0, 1, 0); +lean_ctor_set(x_68, 0, x_67); +lean_ctor_set(x_21, 0, x_68); return x_21; } } } else { -lean_object* x_65; lean_object* x_66; lean_object* x_67; lean_object* x_68; uint8_t x_69; -x_65 = lean_ctor_get(x_21, 1); -lean_inc(x_65); +lean_object* x_69; lean_object* x_70; lean_object* x_71; lean_object* x_72; uint8_t x_73; +x_69 = lean_ctor_get(x_21, 1); +lean_inc(x_69); lean_dec(x_21); -x_66 = lean_ctor_get(x_22, 0); -lean_inc(x_66); +x_70 = lean_ctor_get(x_22, 0); +lean_inc(x_70); if (lean_is_exclusive(x_22)) { lean_ctor_release(x_22, 0); - x_67 = x_22; + x_71 = x_22; } else { lean_dec_ref(x_22); - x_67 = lean_box(0); + x_71 = lean_box(0); } -x_68 = lean_unsigned_to_nat(0u); -x_69 = lean_nat_dec_lt(x_68, x_20); -if (x_69 == 0) +x_72 = lean_unsigned_to_nat(0u); +x_73 = lean_nat_dec_lt(x_72, x_20); +if (x_73 == 0) { -lean_object* x_70; lean_object* x_71; -lean_dec(x_67); -lean_dec(x_66); +lean_object* x_74; lean_object* x_75; +lean_dec(x_71); +lean_dec(x_70); lean_dec(x_20); -x_70 = l_Fin_reduceBin___lambda__1___closed__1; -x_71 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_71, 0, x_70); -lean_ctor_set(x_71, 1, x_65); -return x_71; +x_74 = l_Fin_reduceBin___lambda__1___closed__1; +x_75 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_75, 0, x_74); +lean_ctor_set(x_75, 1, x_69); +return x_75; } else { -lean_object* x_72; lean_object* x_73; lean_object* x_74; lean_object* x_75; lean_object* x_76; lean_object* x_77; lean_object* x_78; lean_object* x_79; lean_object* x_80; lean_object* x_81; lean_object* x_82; lean_object* x_83; lean_object* x_84; lean_object* x_85; -x_72 = lean_nat_mod(x_66, x_20); -lean_dec(x_66); -x_73 = l_Lean_mkRawNatLit(x_72); +lean_object* x_76; lean_object* x_77; lean_object* x_78; lean_object* x_79; lean_object* x_80; lean_object* x_81; lean_object* x_82; lean_object* x_83; lean_object* x_84; lean_object* x_85; lean_object* x_86; lean_object* x_87; lean_object* x_88; lean_object* x_89; lean_object* x_90; lean_object* x_91; +x_76 = lean_nat_mod(x_70, x_20); +lean_dec(x_70); +x_77 = l_Lean_mkRawNatLit(x_76); lean_inc(x_20); -x_74 = l_Lean_mkNatLit(x_20); -x_75 = l_Fin_reduceBin___lambda__1___closed__10; -x_76 = l_Lean_Expr_app___override(x_75, x_74); -x_77 = lean_unsigned_to_nat(1u); -x_78 = lean_nat_sub(x_20, x_77); +x_78 = l_Lean_mkNatLit(x_20); +x_79 = l_Fin_reduceBin___lambda__1___closed__10; +lean_inc(x_78); +x_80 = l_Lean_Expr_app___override(x_79, x_78); +x_81 = lean_unsigned_to_nat(1u); +x_82 = lean_nat_sub(x_20, x_81); lean_dec(x_20); -x_79 = l_Lean_mkNatLit(x_78); -x_80 = l_Fin_reduceBin___lambda__1___closed__13; -lean_inc(x_73); -x_81 = l_Lean_mkAppB(x_80, x_79, x_73); -x_82 = l_Fin_reduceBin___lambda__1___closed__7; -x_83 = l_Lean_mkApp3(x_82, x_76, x_73, x_81); -if (lean_is_scalar(x_67)) { - x_84 = lean_alloc_ctor(0, 1, 0); +x_83 = l_Lean_mkNatLit(x_82); +x_84 = l_Fin_reduceBin___lambda__1___closed__17; +x_85 = l_Lean_Expr_app___override(x_84, x_83); +x_86 = l_Fin_reduceBin___lambda__1___closed__13; +lean_inc(x_77); +x_87 = l_Lean_mkApp3(x_86, x_78, x_85, x_77); +x_88 = l_Fin_reduceBin___lambda__1___closed__7; +x_89 = l_Lean_mkApp3(x_88, x_80, x_77, x_87); +if (lean_is_scalar(x_71)) { + x_90 = lean_alloc_ctor(0, 1, 0); } else { - x_84 = x_67; - lean_ctor_set_tag(x_84, 0); + x_90 = x_71; + lean_ctor_set_tag(x_90, 0); } -lean_ctor_set(x_84, 0, x_83); -x_85 = lean_alloc_ctor(0, 2, 0); -lean_ctor_set(x_85, 0, x_84); -lean_ctor_set(x_85, 1, x_65); -return x_85; +lean_ctor_set(x_90, 0, x_89); +x_91 = lean_alloc_ctor(0, 2, 0); +lean_ctor_set(x_91, 0, x_90); +lean_ctor_set(x_91, 1, x_69); +return x_91; } } } } else { -uint8_t x_86; +uint8_t x_92; lean_dec(x_20); -x_86 = !lean_is_exclusive(x_21); -if (x_86 == 0) +x_92 = !lean_is_exclusive(x_21); +if (x_92 == 0) { return x_21; } else { -lean_object* x_87; lean_object* x_88; lean_object* x_89; -x_87 = lean_ctor_get(x_21, 0); -x_88 = lean_ctor_get(x_21, 1); -lean_inc(x_88); -lean_inc(x_87); +lean_object* x_93; lean_object* x_94; lean_object* x_95; +x_93 = lean_ctor_get(x_21, 0); +x_94 = lean_ctor_get(x_21, 1); +lean_inc(x_94); +lean_inc(x_93); lean_dec(x_21); -x_89 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_89, 0, x_87); -lean_ctor_set(x_89, 1, x_88); -return x_89; +x_95 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_95, 0, x_93); +lean_ctor_set(x_95, 1, x_94); +return x_95; } } } } else { -uint8_t x_90; +uint8_t x_96; lean_dec(x_9); lean_dec(x_8); lean_dec(x_7); lean_dec(x_6); -x_90 = !lean_is_exclusive(x_11); -if (x_90 == 0) +x_96 = !lean_is_exclusive(x_11); +if (x_96 == 0) { return x_11; } else { -lean_object* x_91; lean_object* x_92; lean_object* x_93; -x_91 = lean_ctor_get(x_11, 0); -x_92 = lean_ctor_get(x_11, 1); -lean_inc(x_92); -lean_inc(x_91); +lean_object* x_97; lean_object* x_98; lean_object* x_99; +x_97 = lean_ctor_get(x_11, 0); +x_98 = lean_ctor_get(x_11, 1); +lean_inc(x_98); +lean_inc(x_97); lean_dec(x_11); -x_93 = lean_alloc_ctor(1, 2, 0); -lean_ctor_set(x_93, 0, x_91); -lean_ctor_set(x_93, 1, x_92); -return x_93; +x_99 = lean_alloc_ctor(1, 2, 0); +lean_ctor_set(x_99, 0, x_97); +lean_ctor_set(x_99, 1, x_98); +return x_99; } } } @@ -9485,6 +9590,14 @@ l_Fin_reduceBin___lambda__1___closed__12 = _init_l_Fin_reduceBin___lambda__1___c lean_mark_persistent(l_Fin_reduceBin___lambda__1___closed__12); l_Fin_reduceBin___lambda__1___closed__13 = _init_l_Fin_reduceBin___lambda__1___closed__13(); lean_mark_persistent(l_Fin_reduceBin___lambda__1___closed__13); +l_Fin_reduceBin___lambda__1___closed__14 = _init_l_Fin_reduceBin___lambda__1___closed__14(); +lean_mark_persistent(l_Fin_reduceBin___lambda__1___closed__14); +l_Fin_reduceBin___lambda__1___closed__15 = _init_l_Fin_reduceBin___lambda__1___closed__15(); +lean_mark_persistent(l_Fin_reduceBin___lambda__1___closed__15); +l_Fin_reduceBin___lambda__1___closed__16 = _init_l_Fin_reduceBin___lambda__1___closed__16(); +lean_mark_persistent(l_Fin_reduceBin___lambda__1___closed__16); +l_Fin_reduceBin___lambda__1___closed__17 = _init_l_Fin_reduceBin___lambda__1___closed__17(); +lean_mark_persistent(l_Fin_reduceBin___lambda__1___closed__17); l_Fin_reduceBinPred___lambda__1___closed__1 = _init_l_Fin_reduceBinPred___lambda__1___closed__1(); lean_mark_persistent(l_Fin_reduceBinPred___lambda__1___closed__1); l_Fin_reduceBoolPred___lambda__1___closed__1 = _init_l_Fin_reduceBoolPred___lambda__1___closed__1(); diff --git a/stage0/stdlib/Lean/ToExpr.c b/stage0/stdlib/Lean/ToExpr.c index 6bfd583513ca..7d27b80250e0 100644 --- a/stage0/stdlib/Lean/ToExpr.c +++ b/stage0/stdlib/Lean/ToExpr.c @@ -65,6 +65,7 @@ static lean_object* l_Lean_instToExprUInt8___closed__2; static lean_object* l_Lean_instToExprProd___rarg___lambda__1___closed__3; static lean_object* l_Lean_instToExprList___rarg___closed__7; static lean_object* l___private_Lean_ToExpr_0__Lean_Name_toExprAux_go___closed__7; +static lean_object* l_Lean_instToExprFin___lambda__1___closed__6; lean_object* l_Lean_Expr_lit___override(lean_object*); static lean_object* l_Lean_instToExprUInt16___closed__3; static lean_object* l_Lean_instToExprOption___rarg___lambda__1___closed__2; @@ -219,6 +220,7 @@ static lean_object* l_Lean_instToExprInt___lambda__1___closed__8; LEAN_EXPORT lean_object* l_Lean_instToExprLiteral___lambda__1(lean_object*); LEAN_EXPORT lean_object* l___private_Lean_ToExpr_0__Lean_List_toExprAux___at_Lean_instToExprPreresolved___spec__1(lean_object*, lean_object*, lean_object*); static lean_object* l_Lean_instToExprNat___closed__2; +static lean_object* l_Lean_instToExprFin___lambda__1___closed__7; LEAN_EXPORT lean_object* l___private_Lean_ToExpr_0__Lean_Name_toExprAux_mkStr(lean_object*, lean_object*, lean_object*); static lean_object* l_Lean_Expr_toCtorIfLit___closed__1; static lean_object* l_Lean_instToExprUInt32___closed__1; @@ -271,6 +273,7 @@ static lean_object* l_Lean_instToExprLiteral___closed__4; static lean_object* l_Lean_instToExprInt___closed__3; static lean_object* l_Lean_instToExprOption___rarg___closed__2; static lean_object* l_Lean_instToExprFin___lambda__1___closed__3; +static lean_object* l_Lean_instToExprFin___lambda__1___closed__8; static lean_object* l_Lean_instToExprFilePath___lambda__1___closed__3; LEAN_EXPORT lean_object* l_Lean_instToExprUnit___lambda__1___boxed(lean_object*); static lean_object* l_Lean_instToExprNat___closed__5; @@ -715,25 +718,56 @@ x_3 = l_Lean_Expr_const___override(x_2, x_1); return x_3; } } +static lean_object* _init_l_Lean_instToExprFin___lambda__1___closed__6() { +_start: +{ +lean_object* x_1; +x_1 = lean_mk_string_unchecked("instNeZeroSucc", 14, 14); +return x_1; +} +} +static lean_object* _init_l_Lean_instToExprFin___lambda__1___closed__7() { +_start: +{ +lean_object* x_1; lean_object* x_2; lean_object* x_3; +x_1 = l_Lean_instToExprNat___closed__1; +x_2 = l_Lean_instToExprFin___lambda__1___closed__6; +x_3 = l_Lean_Name_mkStr2(x_1, x_2); +return x_3; +} +} +static lean_object* _init_l_Lean_instToExprFin___lambda__1___closed__8() { +_start: +{ +lean_object* x_1; lean_object* x_2; lean_object* x_3; +x_1 = lean_box(0); +x_2 = l_Lean_instToExprFin___lambda__1___closed__7; +x_3 = l_Lean_Expr_const___override(x_2, x_1); +return x_3; +} +} LEAN_EXPORT lean_object* l_Lean_instToExprFin___lambda__1(lean_object* x_1, lean_object* x_2) { _start: { -lean_object* x_3; lean_object* x_4; lean_object* x_5; lean_object* x_6; lean_object* x_7; lean_object* x_8; lean_object* x_9; lean_object* x_10; lean_object* x_11; lean_object* x_12; lean_object* x_13; +lean_object* x_3; lean_object* x_4; lean_object* x_5; lean_object* x_6; lean_object* x_7; lean_object* x_8; lean_object* x_9; lean_object* x_10; lean_object* x_11; lean_object* x_12; lean_object* x_13; lean_object* x_14; lean_object* x_15; x_3 = l_Lean_mkRawNatLit(x_2); lean_inc(x_1); x_4 = l_Lean_mkNatLit(x_1); x_5 = l_Lean_instToExprFin___lambda__1___closed__3; +lean_inc(x_4); x_6 = l_Lean_Expr_app___override(x_5, x_4); x_7 = lean_unsigned_to_nat(1u); x_8 = lean_nat_sub(x_1, x_7); lean_dec(x_1); x_9 = l_Lean_mkNatLit(x_8); -x_10 = l_Lean_instToExprFin___lambda__1___closed__5; +x_10 = l_Lean_instToExprFin___lambda__1___closed__8; +x_11 = l_Lean_Expr_app___override(x_10, x_9); +x_12 = l_Lean_instToExprFin___lambda__1___closed__5; lean_inc(x_3); -x_11 = l_Lean_mkAppB(x_10, x_9, x_3); -x_12 = l_Lean_instToExprInt_mkNat___closed__6; -x_13 = l_Lean_mkApp3(x_12, x_6, x_3, x_11); -return x_13; +x_13 = l_Lean_mkApp3(x_12, x_4, x_11, x_3); +x_14 = l_Lean_instToExprInt_mkNat___closed__6; +x_15 = l_Lean_mkApp3(x_14, x_6, x_3, x_13); +return x_15; } } static lean_object* _init_l_Lean_instToExprFin___closed__1() { @@ -2008,7 +2042,7 @@ static lean_object* _init_l___private_Lean_ToExpr_0__Lean_Name_toExprAux_mkStr__ lean_object* x_1; lean_object* x_2; lean_object* x_3; lean_object* x_4; lean_object* x_5; lean_object* x_6; x_1 = l___private_Lean_ToExpr_0__Lean_Name_toExprAux_mkStr___closed__5; x_2 = l___private_Lean_ToExpr_0__Lean_Name_toExprAux_mkStr___closed__6; -x_3 = lean_unsigned_to_nat(129u); +x_3 = lean_unsigned_to_nat(130u); x_4 = lean_unsigned_to_nat(11u); x_5 = l___private_Lean_ToExpr_0__Lean_Name_toExprAux_mkStr___closed__7; x_6 = l___private_Init_Util_0__mkPanicMessageWithDecl(x_1, x_2, x_3, x_4, x_5); @@ -3523,6 +3557,12 @@ l_Lean_instToExprFin___lambda__1___closed__4 = _init_l_Lean_instToExprFin___lamb lean_mark_persistent(l_Lean_instToExprFin___lambda__1___closed__4); l_Lean_instToExprFin___lambda__1___closed__5 = _init_l_Lean_instToExprFin___lambda__1___closed__5(); lean_mark_persistent(l_Lean_instToExprFin___lambda__1___closed__5); +l_Lean_instToExprFin___lambda__1___closed__6 = _init_l_Lean_instToExprFin___lambda__1___closed__6(); +lean_mark_persistent(l_Lean_instToExprFin___lambda__1___closed__6); +l_Lean_instToExprFin___lambda__1___closed__7 = _init_l_Lean_instToExprFin___lambda__1___closed__7(); +lean_mark_persistent(l_Lean_instToExprFin___lambda__1___closed__7); +l_Lean_instToExprFin___lambda__1___closed__8 = _init_l_Lean_instToExprFin___lambda__1___closed__8(); +lean_mark_persistent(l_Lean_instToExprFin___lambda__1___closed__8); l_Lean_instToExprFin___closed__1 = _init_l_Lean_instToExprFin___closed__1(); lean_mark_persistent(l_Lean_instToExprFin___closed__1); l_Lean_instToExprBitVec___lambda__1___closed__1 = _init_l_Lean_instToExprBitVec___lambda__1___closed__1(); diff --git a/stage0/stdlib/Std/Tactic/BVDecide/Bitblast/BVExpr/Basic.c b/stage0/stdlib/Std/Tactic/BVDecide/Bitblast/BVExpr/Basic.c index c1dafc69eb04..1bf05b63230a 100644 --- a/stage0/stdlib/Std/Tactic/BVDecide/Bitblast/BVExpr/Basic.c +++ b/stage0/stdlib/Std/Tactic/BVDecide/Bitblast/BVExpr/Basic.c @@ -48,7 +48,6 @@ lean_object* l_BitVec_shiftLeft(lean_object*, lean_object*, lean_object*); lean_object* lean_nat_shiftr(lean_object*, lean_object*); static lean_object* l___private_Std_Tactic_BVDecide_Bitblast_BVExpr_Basic_0__Std_Tactic_BVDecide_reprBVBit____x40_Std_Tactic_BVDecide_Bitblast_BVExpr_Basic___hyg_287____closed__1; static lean_object* l___private_Std_Tactic_BVDecide_Bitblast_BVExpr_Basic_0__Std_Tactic_BVDecide_reprBVBit____x40_Std_Tactic_BVDecide_Bitblast_BVExpr_Basic___hyg_287____closed__17; -lean_object* l_Fin_ofNat(lean_object*, lean_object*); LEAN_EXPORT uint64_t l___private_Std_Tactic_BVDecide_Bitblast_BVExpr_Basic_0__Std_Tactic_BVDecide_hashBVBit____x40_Std_Tactic_BVDecide_Bitblast_BVExpr_Basic___hyg_35_(lean_object*); LEAN_EXPORT lean_object* l_Std_Tactic_BVDecide_BVExpr_eval(lean_object*, lean_object*, lean_object*); static lean_object* l_Std_Tactic_BVDecide_BVUnOp_toString___closed__3; @@ -64,7 +63,6 @@ static lean_object* l_Std_Tactic_BVDecide_BVBinPred_toString___closed__1; static lean_object* l_Std_Tactic_BVDecide_BVBinPred_instToString___closed__1; lean_object* l_BitVec_not(lean_object*, lean_object*); lean_object* lean_nat_to_int(lean_object*); -static lean_object* l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__2; LEAN_EXPORT lean_object* l_Std_Tactic_BVDecide_BVUnOp_eval___boxed(lean_object*, lean_object*, lean_object*); LEAN_EXPORT lean_object* l_Std_Tactic_BVDecide_BVExpr_Assignment_getD(lean_object*, lean_object*); static lean_object* l_Std_Tactic_BVDecide_BVBinOp_noConfusion___rarg___closed__1; @@ -624,31 +622,21 @@ return x_12; static lean_object* _init_l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__1() { _start: { -lean_object* x_1; lean_object* x_2; -x_1 = lean_unsigned_to_nat(0u); -x_2 = l_Fin_ofNat(x_1, x_1); -return x_2; -} -} -static lean_object* _init_l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__2() { -_start: -{ -lean_object* x_1; lean_object* x_2; lean_object* x_3; lean_object* x_4; +lean_object* x_1; lean_object* x_2; lean_object* x_3; x_1 = lean_unsigned_to_nat(1u); x_2 = lean_unsigned_to_nat(0u); -x_3 = l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__1; -x_4 = lean_alloc_ctor(0, 3, 0); -lean_ctor_set(x_4, 0, x_1); -lean_ctor_set(x_4, 1, x_2); -lean_ctor_set(x_4, 2, x_3); -return x_4; +x_3 = lean_alloc_ctor(0, 3, 0); +lean_ctor_set(x_3, 0, x_1); +lean_ctor_set(x_3, 1, x_2); +lean_ctor_set(x_3, 2, x_2); +return x_3; } } static lean_object* _init_l_Std_Tactic_BVDecide_instInhabitedBVBit() { _start: { lean_object* x_1; -x_1 = l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__2; +x_1 = l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__1; return x_1; } } @@ -2130,8 +2118,6 @@ l_Std_Tactic_BVDecide_instToStringBVBit___closed__3 = _init_l_Std_Tactic_BVDecid lean_mark_persistent(l_Std_Tactic_BVDecide_instToStringBVBit___closed__3); l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__1 = _init_l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__1(); lean_mark_persistent(l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__1); -l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__2 = _init_l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__2(); -lean_mark_persistent(l_Std_Tactic_BVDecide_instInhabitedBVBit___closed__2); l_Std_Tactic_BVDecide_instInhabitedBVBit = _init_l_Std_Tactic_BVDecide_instInhabitedBVBit(); lean_mark_persistent(l_Std_Tactic_BVDecide_instInhabitedBVBit); l_Std_Tactic_BVDecide_BVBinOp_noConfusion___rarg___closed__1 = _init_l_Std_Tactic_BVDecide_BVBinOp_noConfusion___rarg___closed__1();