/* Copyright (c) 2014 Microsoft Corporation. All rights reserved. Released under Apache 2.0 license as described in the file LICENSE. Author: Leonardo de Moura */ #include #include #include "util/pair.h" #include "frontends/lean/token_table.h" namespace lean { static unsigned g_arrow_prec = 25; static unsigned g_decreasing_prec = 100; static unsigned g_max_prec = 1024; static unsigned g_Max_prec = 1024*1024; static unsigned g_plus_prec = 65; static unsigned g_cup_prec = 60; unsigned get_max_prec() { return g_max_prec; } unsigned get_Max_prec() { return g_Max_prec; } unsigned get_arrow_prec() { return g_arrow_prec; } unsigned get_decreasing_prec() { return g_decreasing_prec; } static token_table update(token_table const & s, char const * token, char const * val, optional expr_prec, optional tac_prec) { lean_assert(expr_prec || tac_prec); token_info info(token, val, 0, 0); if (token_info const * old_info = find(s, token)) { info = info.update_expr_precedence(old_info->expr_precedence()); info = info.update_tactic_precedence(old_info->tactic_precedence()); } if (expr_prec) info = info.update_expr_precedence(*expr_prec); if (tac_prec) info = info.update_tactic_precedence(*tac_prec); return insert(s, token, info); } token_table add_command_token(token_table const & s, char const * token) { return insert(s, token, token_info(token)); } token_table add_command_token(token_table const & s, char const * token, char const * val) { return insert(s, token, token_info(token, val)); } token_table add_token(token_table const & s, char const * token, unsigned prec) { return update(s, token, token, optional(prec), optional()); } token_table add_token(token_table const & s, char const * token, char const * val, unsigned prec) { return update(s, token, val, optional(prec), optional()); } token_table add_tactic_token(token_table const & s, char const * token, unsigned prec) { return update(s, token, token, optional(), optional(prec)); } token_table add_tactic_token(token_table const & s, char const * token, char const * val, unsigned prec) { return update(s, token, val, optional(), optional(prec)); } token_table const * find(token_table const & s, char c) { return s.find(c); } token_info const * value_of(token_table const & s) { return s.value(); } optional get_expr_precedence(token_table const & s, char const * token) { auto it = find(s, token); return it ? optional(it->expr_precedence()) : optional(); } optional get_tactic_precedence(token_table const & s, char const * token) { auto it = find(s, token); return it ? optional(it->tactic_precedence()) : optional(); } bool is_token(token_table const & s, char const * token) { return static_cast(find(s, token)); } void for_each(token_table const & s, std::function const & fn) { s.for_each([&](unsigned num, char const * keys, token_info const & info) { buffer str; str.append(num, keys); str.push_back(0); fn(str.data(), info); }); } static char const * g_lambda_unicode = "\u03BB"; static char const * g_pi_unicode = "\u03A0"; static char const * g_forall_unicode = "\u2200"; static char const * g_arrow_unicode = "\u2192"; static char const * g_cup = "\u2294"; static char const * g_qed_unicode = "∎"; static char const * g_decreasing_unicode = "↓"; void init_token_table(token_table & t) { pair builtin[] = {{"fun", 0}, {"Pi", 0}, {"let", 0}, {"in", 0}, {"at", 0}, {"have", 0}, {"assert", 0}, {"show", 0}, {"obtain", 0}, {"if", 0}, {"then", 0}, {"else", 0}, {"by", 0}, {"by+", 0}, {"hiding", 0}, {"replacing", 0}, {"renaming", 0}, {"from", 0}, {"(", g_max_prec}, {")", 0}, {"{", g_max_prec}, {"}", 0}, {"_", g_max_prec}, {"[", g_max_prec}, {"]", 0}, {"⦃", g_max_prec}, {"⦄", 0}, {".{", 0}, {"Type", g_max_prec}, {"{|", g_max_prec}, {"|}", 0}, {"⊢", 0}, {"⟨", g_max_prec}, {"⟩", 0}, {"^", 0}, {"↑", 0}, {"▸", 0}, {"using", 0}, {"|", 0}, {"!", g_max_prec}, {"?", 0}, {"with", 0}, {"...", 0}, {",", 0}, {".", 0}, {":", 0}, {"::", 0}, {"calc", 0}, {"rewrite", 0}, {"esimp", 0}, {"fold", 0}, {"unfold", 0}, {"generalize", 0}, {"as", 0}, {":=", 0}, {"--", 0}, {"#", 0}, {"(*", 0}, {"/-", 0}, {"begin", g_max_prec}, {"begin+", g_max_prec}, {"proof", g_max_prec}, {"qed", 0}, {"@", g_max_prec}, {"sorry", g_max_prec}, {"+", g_plus_prec}, {g_cup, g_cup_prec}, {"->", g_arrow_prec}, {"?(", g_max_prec}, {"⌞", g_max_prec}, {"⌟", 0}, {"match", 0}, {" aliases[] = {{g_lambda_unicode, "fun"}, {"forall", "Pi"}, {g_forall_unicode, "Pi"}, {g_pi_unicode, "Pi"}, {g_qed_unicode, "qed"}, {nullptr, nullptr}}; pair cmd_aliases[] = {{"lemma", "theorem"}, {"premise", "variable"}, {"premises", "variables"}, {"corollary", "theorem"}, {"hypothesis", "parameter"}, {"conjecture", "parameter"}, {"record", "structure"}, {nullptr, nullptr}}; auto it = builtin; while (it->first) { t = add_token(t, it->first, it->second); it++; } auto it2 = commands; while (*it2) { t = add_command_token(t, *it2); ++it2; } auto it3 = aliases; while (it3->first) { t = add_token(t, it3->first, it3->second, 0); it3++; } t = add_token(t, g_arrow_unicode, "->", get_arrow_prec()); t = add_token(t, g_decreasing_unicode, "first) { t = add_command_token(t, it4->first, it4->second); ++it4; } } static token_table * g_default_token_table = nullptr; token_table mk_default_token_table() { return *g_default_token_table; } void initialize_token_table() { g_default_token_table = new token_table(); init_token_table(*g_default_token_table); } void finalize_token_table() { delete g_default_token_table; } token_table mk_token_table() { return token_table(); } DECL_UDATA(token_table) static int mk_token_table(lua_State * L) { return push_token_table(L, mk_token_table()); } static int mk_default_token_table(lua_State * L) { return push_token_table(L, mk_default_token_table()); } static int add_command_token(lua_State * L) { int nargs = lua_gettop(L); if (nargs == 2) return push_token_table(L, add_command_token(to_token_table(L, 1), lua_tostring(L, 2))); else return push_token_table(L, add_command_token(to_token_table(L, 1), lua_tostring(L, 2), lua_tostring(L, 3))); } static int add_token(lua_State * L) { int nargs = lua_gettop(L); if (nargs == 3) return push_token_table(L, add_token(to_token_table(L, 1), lua_tostring(L, 2), lua_tonumber(L, 3))); else return push_token_table(L, add_token(to_token_table(L, 1), lua_tostring(L, 2), lua_tostring(L, 3), lua_tonumber(L, 4))); } static int merge(lua_State * L) { return push_token_table(L, merge(to_token_table(L, 1), to_token_table(L, 2))); } static int find(lua_State * L) { char k; if (lua_isnumber(L, 2)) { k = lua_tonumber(L, 2); } else { char const * str = lua_tostring(L, 2); if (strlen(str) != 1) throw exception("arg #2 must be a string of length 1"); k = str[0]; } auto it = to_token_table(L, 1).find(k); if (it) return push_token_table(L, *it); else return push_nil(L); } static int value_of(lua_State * L) { auto it = value_of(to_token_table(L, 1)); if (it) { push_boolean(L, it->is_command()); push_name(L, it->value()); push_integer(L, it->expr_precedence()); return 3; } else { push_nil(L); return 1; } } static int for_each(lua_State * L) { token_table const & t = to_token_table(L, 1); luaL_checktype(L, 2, LUA_TFUNCTION); // user-fun for_each(t, [&](char const * k, token_info const & info) { lua_pushvalue(L, 2); lua_pushstring(L, k); lua_pushboolean(L, info.is_command()); push_name(L, info.value()); lua_pushinteger(L, info.expr_precedence()); pcall(L, 4, 0, 0); }); return 0; } static const struct luaL_Reg token_table_m[] = { {"__gc", token_table_gc}, {"add_command_token", safe_function}, {"add_token", safe_function}, {"merge", safe_function}, {"find", safe_function}, {"value_of", safe_function}, {"for_each", safe_function}, {0, 0} }; void open_token_table(lua_State * L) { luaL_newmetatable(L, token_table_mt); lua_pushvalue(L, -1); lua_setfield(L, -2, "__index"); setfuncs(L, token_table_m, 0); SET_GLOBAL_FUN(token_table_pred, "is_token_table"); SET_GLOBAL_FUN(mk_default_token_table, "default_token_table"); SET_GLOBAL_FUN(mk_token_table, "token_table"); } }