X-Git-Url: http://matita.cs.unibo.it/gitweb/?a=blobdiff_plain;f=helm%2Fpapers%2Fmatita%2Fmatita.tex;h=d9a7a525f572972bb90800a547c1d25635a2a146;hb=97c2d258a5c524eb5c4b85208899d80751a2c82f;hp=10a44328010fb13b84dad5f9f15c26c63df36f86;hpb=7b40df1df60fa1dd2bb1058850474c0f66bd01b4;p=helm.git diff --git a/helm/papers/matita/matita.tex b/helm/papers/matita/matita.tex index 10a443280..d9a7a525f 100644 --- a/helm/papers/matita/matita.tex +++ b/helm/papers/matita/matita.tex @@ -37,6 +37,16 @@ \newcommand{\TEXMACRO}[1]{\texttt{\char92 #1}} \newcommand{\UWOBO}{UWOBO} \newcommand{\WHELP}{Whelp} +\newcommand{\DOT}{\ensuremath{\mbox{\textbf{.}}}} +\newcommand{\SEMICOLON}{\ensuremath{\mbox{\textbf{;}}}} +\newcommand{\BRANCH}{\ensuremath{\mbox{\textbf{[}}}} +\newcommand{\SHIFT}{\ensuremath{\mbox{\textbf{\textbar}}}} +\newcommand{\POS}[1]{\ensuremath{#1\mbox{\textbf{:}}}} +\newcommand{\MERGE}{\ensuremath{\mbox{\textbf{]}}}} +\newcommand{\FOCUS}[1]{\ensuremath{\mathtt{focus}~#1}} +\newcommand{\UNFOCUS}{\ensuremath{\mathtt{unfocus}}} +\newcommand{\SKIP}{\MATHTT{skip}} +\newcommand{\TACTIC}[1]{\ensuremath{\mathtt{tactic}~#1}} \definecolor{gray}{gray}{0.85} % 1 -> white; 0 -> black \newcommand{\NT}[1]{\langle\mathit{#1}\rangle} @@ -56,7 +66,7 @@ \newcommand{\ASSIGNEDTO}[1]{\textbf{Assigned to:} #1} \newcommand{\FILE}[1]{\texttt{#1}} % \newcommand{\NOTE}[1]{\ifodd \arabic{page} \else \hspace{-2cm}\fi\ednote{#1}} -\newcommand{\NOTE}[1]{\ednote{x~\hspace{-10cm}y#1}{foo}} +\newcommand{\NOTE}[1]{\ednote{#1}{foo}} \newcommand{\TODO}[1]{\textbf{TODO: #1}} \newsavebox{\tmpxyz} @@ -322,12 +332,99 @@ library as a whole will be logically inconsistent. Logical inconsistency has never been a problem in the daily work of a mathematician. The mathematician simply imposes himself a discipline to restrict himself to consistent subsets of the mathematical knowledge. -However, in doing so he doesn't choose the subset in advance by forgetting -the rest of his knowledge. +However, in doing so he does not choose the subset in advance by forgetting +the rest of his knowledge. On the contrary he may proceed with a sort of +top-down strategy: he may always inspect or use part of his knowledge, but +when he actually does so he should check recursively that inconsistencies are +not exploited. + +Contrarily to the mathematical practice, the usual tendency in the world of +assisted automation is that of building a logical environment (a consistent +subset of the library) in a bottom up way, checking the consistency of a +new axiom or theorem as soon as it is added to the environment. No lemma +or definition outside the environment can be used until it is added to the +library after every notion it depends on. Moreover, very often the logical +environment is the only part of the library that can be inspected, +that we can search lemmas in and that can be exploited by automatic tactics. + +Moving one by one notions from the library to the environment is a costly +operation since it involves re-checking the correctness of the notion. +As a consequence mathematical notions are packages into theories that must +be added to the environment as a whole. However, the consistency problem is +only raised at the level of theories: theories must be imported in a bottom +up way and the system must check that no inconsistency arises. + +The practice of limiting the scope on the library to the logical environment +is contrary to our commitment of being able to fully exploit as much as possible +of the library at any given time. To reconcile consistency and visibility +we have departed from the traditional implementation of an environment +allowing environments to be built on demand in a top-down way. The new +implementation is based on a modified meta-theory that changes the way +convertibility, type checking, unification and refinement judgements. +The modified meta-theory is fully described in \cite{libraryenvironments}. +Here we just remark how a strong commitment on the way the user interacts +with the library has lead to modifications to the logical core of the proof +assistant. This is evidence that breakthroughs in the user interfaces +and in the way the user interacts with the tools and with the library could +be achieved only by means of strong architectural changes. + +\subsubsection{Accessibility} +A large library that is completely in scope needs effective indexing and +searching methods to make the user productive. Libraries of formal results +are particularly critical since they hold a large percentage of technical +lemmas that do not have a significative name and that must be retrieved +using advanced methods based on matching, unification, generalization and +instantiation. + +The efficiency of searching inside the library becomes a critical operation +when automatic tactics exploit the library during the proof search. In this +scenario the tactics must retrieve a set of candidates for backward or +forward reasoning in a few milliseconds. + +The choice of several proof assistants is to use ad-hoc data structures, +such as context trees, to index all the terms currently in scope. These +data structures are expecially designed to quickly retrieve terms up +to matching, instantiation and generalization. All these data structures +try to maximize sharing of identical subterms so that matching can be +reduced to a visit of the tree (or dag) that holds all the maximally shared +terms together. + +Since the terms to be retrieved (or at least their initial prefix) +are stored (actually ``melted'') in the data structure, these data structures +must collect all the terms in a single location. In other words, adopting +such data structures means centralizing the library. + +In the \MOWGLI{} project we have tried to follow an alternative approach +that consists in keeping the library fully distributed and indexing it +by means of spiders that collect metadata and store them in a database. +The challenge is to be able to collect only a smaller as possible number +of metadata that provide enough information to approximate the matching +operation. A matching operation is then performed in two steps. The first +step is a query to the remote search engine that stores the metadata in +order to detect a (hopefully small) complete set of candidates that could +match. Completeness means that no term that matches should be excluded from +the set of candiates. The second step consists in retrieving from the +distributed library all the candidates and attempt the actual matching. + +In the last we years we have progressively improved this technique. +Our achievements can be found in \cite{query1,query2,query3}. + +The technique and tools already developed have been integrated in \MATITA{}, +that is able to contact a remote \WHELP{} search engine \cite{whelp} or that +can be directly linked to the code of the \WHELP. In either case the database +used to store the metadata can be local or remote. + +Our current challenge consists in the exploitation of \WHELP{} inside of +\MATITA. In particular we are developing a set of tactics, for instance +based on paramodulation \cite{paramodulation}, that perform queries to \WHELP{} +to restrict the scope on the library to a set of interesting candidates, +greatly reducing the search space. Moreover, queries to \WHELP{} are performed +during parsing of user provided terms to disambiguate them. + +In Sect.~\ref{sec:metadata} we describe the technique adopted in \MATITA. + +\subsubsection{Library management} -Contrarily to a mathematician, the usual tendency in the world of assisted -automation is that of restricting in advance the part of the library that -will be used later on, checking its consistency by construction. \subsection{ricerca e indicizzazione} \label{sec:metadata} @@ -352,6 +449,54 @@ will be used later on, checking its consistency by construction. \label{sec:disambiguation} \ASSIGNEDTO{zack} + \begin{table} + \caption{\label{tab:termsyn} Concrete syntax of CIC terms: built-in + notation\strut} + \hrule + \[ + \begin{array}{@{}rcll@{}} + \NT{term} & ::= & & \mbox{\bf terms} \\ + & & x & \mbox{(identifier)} \\ + & | & n & \mbox{(number)} \\ + & | & s & \mbox{(symbol)} \\ + & | & \mathrm{URI} & \mbox{(URI)} \\ + & | & \verb+_+ & \mbox{(implicit)}\TODO{sync} \\ + & | & \verb+?+n~[\verb+[+~\{\NT{subst}\}~\verb+]+] & \mbox{(meta)} \\ + & | & \verb+let+~\NT{ptname}~\verb+\def+~\NT{term}~\verb+in+~\NT{term} \\ + & | & \verb+let+~\NT{kind}~\NT{defs}~\verb+in+~\NT{term} \\ + & | & \NT{binder}~\{\NT{ptnames}\}^{+}~\verb+.+~\NT{term} \\ + & | & \NT{term}~\NT{term} & \mbox{(application)} \\ + & | & \verb+Prop+ \mid \verb+Set+ \mid \verb+Type+ \mid \verb+CProp+ & \mbox{(sort)} \\ + & | & \verb+match+~\NT{term}~ & \mbox{(pattern matching)} \\ + & & ~ ~ [\verb+[+~\verb+in+~x~\verb+]+] + ~ [\verb+[+~\verb+return+~\NT{term}~\verb+]+] \\ + & & ~ ~ \verb+with [+~[\NT{rule}~\{\verb+|+~\NT{rule}\}]~\verb+]+ & \\ + & | & \verb+(+~\NT{term}~\verb+:+~\NT{term}~\verb+)+ & \mbox{(cast)} \\ + & | & \verb+(+~\NT{term}~\verb+)+ \\ + \NT{defs} & ::= & & \mbox{\bf mutual definitions} \\ + & & \NT{fun}~\{\verb+and+~\NT{fun}\} \\ + \NT{fun} & ::= & & \mbox{\bf functions} \\ + & & \NT{arg}~\{\NT{ptnames}\}^{+}~[\verb+on+~x]~\verb+\def+~\NT{term} \\ + \NT{binder} & ::= & & \mbox{\bf binders} \\ + & & \verb+\forall+ \mid \verb+\lambda+ \\ + \NT{arg} & ::= & & \mbox{\bf single argument} \\ + & & \verb+_+ \mid x \\ + \NT{ptname} & ::= & & \mbox{\bf possibly typed name} \\ + & & \NT{arg} \\ + & | & \verb+(+~\NT{arg}~\verb+:+~\NT{term}~\verb+)+ \\ + \NT{ptnames} & ::= & & \mbox{\bf bound variables} \\ + & & \NT{arg} \\ + & | & \verb+(+~\NT{arg}~\{\verb+,+~\NT{arg}\}~[\verb+:+~\NT{term}]~\verb+)+ \\ + \NT{kind} & ::= & & \mbox{\bf induction kind} \\ + & & \verb+rec+ \mid \verb+corec+ \\ + \NT{rule} & ::= & & \mbox{\bf rules} \\ + & & x~\{\NT{ptname}\}~\verb+\Rightarrow+~\NT{term} + \end{array} + \] + \hrule + \end{table} + + \subsubsection{Term input} The primary form of user interaction employed by \MATITA{} is textual script @@ -833,18 +978,99 @@ One of the goals of \MATITA{} is to use modern publishing techiques, and adopting a method for restricting tactics application domain that discourages using heavy math notation, would definitively be a bad choice. -\subsection{tatticali} -\ASSIGNEDTO{gares} -\begin{verbatim} -- problemi principali dei PA basati su tattiche - o illeggibilita' dello script - o scarsa strutturazione dello script e mantenibilita' - - ameliorate by tacticals -- intro sui tatticali esistenti -- peculiarita' di matita - - passo passo - - nice handling of sideeffects (see later on metavariables) -\end{verbatim} +\subsection{Tacticals} +\ASSIGNEDTO{gares}\\ +There are mainly two kinds of languages used by proof assistants to recorder +proofs: tactic based and declarative. We will not investigate the philosophy +aroud the choice that many proof assistant made, \MATITA{} included, and we will not compare the two diffrent approaches. We will describe the common issues of the first one and how \MATITA{} tries to solve them. + +First we must highlight the fact that proof scripts made using tactis are +particularly unreadable. This is not a big deal for the user while he is +constructing the proof, but is considerably a problem when he tries to reread +what he did or when he shows his work to someone else. + +Another common issue for tactic based proof scripts is their mantenibility. +Huge libraries have been developed, and backward compatibility is a really time +consuming task. This problem is usually ameliorated with tacticals, that +help in structuring proofs and consequently their maintenance, but have a bad +counterpart in script readability. Since tacticals are executed atomically, +the common practice of executing again a script to review all the proof steps +doesn't work at all. This issue in addition to the really poor feeling that a +list of tactics gives about the proof makes script rereading particularly hard. + +\MATITA{} uses a language of tactics and tacticals, but adopts a peculiar +strategy to make this technique more user friendly without loosing in +mantenibility or expressivity. + +\subsubsection{Tacticals overview} +Before describing the peculiarities of \MATITA{} tacticals we briefly introduce what tacticals are and where they can be useful. + +Tacticals first appered in LCF(cita qualcosa) and can be seen as programming +constructs, like looping, branching, error recovery or sequential composition. +For example $tac_1~.~tac_2$ executes the first tactic and applies the second +only to the first goal opened by $tac_1$. Baranching can be used to specify a +diffrent tactic to apply to each new goal opened by another tactic, for example +$tac_1\verb+;[+~tac_{1.1}~\verb+|+~tac_{1.2}~\verb+|+~\cdots~|~tac_{1.n}~\verb+]+$ +applies respectively $tac_{1.i}$ to the $i$-th goal opened by $tac_1$. Looping +can be used to iterate a tactic until it works: $\verb+repeat+~tac$ applies +$tac$ to the current goal, and again $tac$ to the eventually resulting goals +until all goal are closed or the tactic fails. + +\begin{table} + \caption{\label{tab:tacsyn} Concrete syntax of \MATITA{} tacticals.\strut} +\hrule +\[ +\begin{array}{@{}rcll@{}} + \NT{punctuation} & + ::= & \SEMICOLON \quad|\quad \DOT \quad|\quad \SHIFT \quad|\quad \BRANCH \quad|\quad \MERGE \quad|\quad \POS{\mathrm{NUMBER}~} & \\ + \NT{block\_kind} & + ::= & \verb+focus+ ~|~ \verb+try+ ~|~ \verb+solve+ ~|~ \verb+first+ ~|~ \verb+repeat+ ~|~ \verb+do+~\mathrm{NUMBER} & \\ + \NT{block\_delimiter} & + ::= & \verb+begin+ ~|~ \verb+end+ & \\ + \NT{tactical} & + ::= & \verb+skip+ ~|~ \NT{tactic} ~|~ \NT{block\_delimiter} ~|~ \NT{block\_kind} ~|~ \NT{punctuation} ~|~& \\ +\end{array} +\] +\hrule +\end{table} + +\MATITA{} tacticals syntax is reported in table \ref{tab:tacsyn}. +While one whould expect to find structured constructs like +$\verb+do+~n~\NT{tactic}$ the syntax allows pieces of tacticals to be written. +This is essential for base idea behind matita tacticals: step-by-step execution. + +\subsubsection{\MATITA{} Tinycals} +The low-level tacticals implementation of \MATITA{} allows a step-by-step execution of a tactical, that substantially means that a $\NT{block\_kind}$ is not executed as an atomic operation. This has two major benefits for the user, even being a so simple idea: +\begin{description} +\item[Proof structuring] + is much easyer. Consider for example a proof by induction, and imagine you are using classical tacticals. After applying the + induction principle, with one step tacticals, you have to choose: structure + the proof or not. If you decide for the former you have to branch with + \verb+[+ and write tactics for all the cases separated by \verb+|+ and the + close the tactical with + \verb+]+. You can replace most of the cases by the identity tactic just to + concentrate only on the first goal, but you will have to go one step back and + one further every time you add something inside the tactical. And if you are + boared of doing so, you will finish in giving up structuring the proof and + write a plain list of tactics.\\ + With step-by-step tacticals you can apply the induction principle, and just + open the branching tactical \verb+[+. Then you can interact with the system + reaching a proof of the first case, without having to specify the whole + branching tactical. When you have proved all the induction cases, you close + the branching tactical with \verb+]+ and you are done with a structured proof. +\item[Rereading] + is possible. Going on step by step shows exactly what is going on. + Consider again a proof by induction, that starts applying the induction + principle and suddenly baranches with a \verb+[+. This clearly separates all + the induction cases, but if the square brackets content is executed in one + single step you completely loose the possibility of rereading it. Again, + executing step-by-step is the way you whould like to review the + demonstration. Remember tha understandig the proof from the script is not + easy, and only the execution of tactics (and the resulting transformed goal) + gives you the feeling of what is goning on. +\end{description} + + \subsection{named variable e disambiguazione lazy} \ASSIGNEDTO{csc}