87,046 research outputs found
Interaction platform-orientated perspective in designing novel applications
The lack of HCI offerings in the invention of novel software applications and the bias of design knowledge towards desktop GUI make it difficult for us to design for novel scenarios and applications that leverage emerging computational technologies. These include new media platforms such as mobiles, interactive TV, tabletops and large multi-touch walls on which many of our future applications will operate. We argue that novel application design should come not from user-centred requirements engineering as in developing a conventional application, but from understanding the interaction characteristics of the new platforms. Ensuring general usability for a particular interaction platform without rigorously specifying envisaged usage contexts helps us to design an artifact that does not restrict the possible application contexts and yet is usable enough to help brainstorm its more exact place for future exploitation
A Bi-Directional Refinement Algorithm for the Calculus of (Co)Inductive Constructions
The paper describes the refinement algorithm for the Calculus of
(Co)Inductive Constructions (CIC) implemented in the interactive theorem prover
Matita. The refinement algorithm is in charge of giving a meaning to the terms,
types and proof terms directly written by the user or generated by using
tactics, decision procedures or general automation. The terms are written in an
"external syntax" meant to be user friendly that allows omission of
information, untyped binders and a certain liberal use of user defined
sub-typing. The refiner modifies the terms to obtain related well typed terms
in the internal syntax understood by the kernel of the ITP. In particular, it
acts as a type inference algorithm when all the binders are untyped. The
proposed algorithm is bi-directional: given a term in external syntax and a
type expected for the term, it propagates as much typing information as
possible towards the leaves of the term. Traditional mono-directional
algorithms, instead, proceed in a bottom-up way by inferring the type of a
sub-term and comparing (unifying) it with the type expected by its context only
at the end. We propose some novel bi-directional rules for CIC that are
particularly effective. Among the benefits of bi-directionality we have better
error message reporting and better inference of dependent types. Moreover,
thanks to bi-directionality, the coercion system for sub-typing is more
effective and type inference generates simpler unification problems that are
more likely to be solved by the inherently incomplete higher order unification
algorithms implemented. Finally we introduce in the external syntax the notion
of vector of placeholders that enables to omit at once an arbitrary number of
arguments. Vectors of placeholders allow a trivial implementation of implicit
arguments and greatly simplify the implementation of primitive and simple
tactics
Logic of Negation-Complete Interactive Proofs (Formal Theory of Epistemic Deciders)
We produce a decidable classical normal modal logic of internalised
negation-complete and thus disjunctive non-monotonic interactive proofs (LDiiP)
from an existing logical counterpart of non-monotonic or instant interactive
proofs (LiiP). LDiiP internalises agent-centric proof theories that are
negation-complete (maximal) and consistent (and hence strictly weaker than, for
example, Peano Arithmetic) and enjoy the disjunction property (like
Intuitionistic Logic). In other words, internalised proof theories are
ultrafilters and all internalised proof goals are definite in the sense of
being either provable or disprovable to an agent by means of disjunctive
internalised proofs (thus also called epistemic deciders). Still, LDiiP itself
is classical (monotonic, non-constructive), negation-incomplete, and does not
have the disjunction property. The price to pay for the negation completeness
of our interactive proofs is their non-monotonicity and non-communality (for
singleton agent communities only). As a normal modal logic, LDiiP enjoys a
standard Kripke-semantics, which we justify by invoking the Axiom of Choice on
LiiP's and then construct in terms of a concrete oracle-computable function.
LDiiP's agent-centric internalised notion of proof can also be viewed as a
negation-complete disjunctive explicit refinement of standard KD45-belief, and
yields a disjunctive but negation-incomplete explicit refinement of
S4-provability.Comment: Expanded Introduction. Added Footnote 4. Corrected Corollary 3 and 4.
Continuation of arXiv:1208.184
Logic of Intuitionistic Interactive Proofs (Formal Theory of Perfect Knowledge Transfer)
We produce a decidable super-intuitionistic normal modal logic of
internalised intuitionistic (and thus disjunctive and monotonic) interactive
proofs (LIiP) from an existing classical counterpart of classical monotonic
non-disjunctive interactive proofs (LiP). Intuitionistic interactive proofs
effect a durable epistemic impact in the possibly adversarial communication
medium CM (which is imagined as a distinguished agent), and only in that, that
consists in the permanent induction of the perfect and thus disjunctive
knowledge of their proof goal by means of CM's knowledge of the proof: If CM
knew my proof then CM would persistently and also disjunctively know that my
proof goal is true. So intuitionistic interactive proofs effect a lasting
transfer of disjunctive propositional knowledge (disjunctively knowable facts)
in the communication medium of multi-agent distributed systems via the
transmission of certain individual knowledge (knowable intuitionistic proofs).
Our (necessarily) CM-centred notion of proof is also a disjunctive explicit
refinement of KD45-belief, and yields also such a refinement of standard
S5-knowledge. Monotonicity but not communality is a commonality of LiP, LIiP,
and their internalised notions of proof. As a side-effect, we offer a short
internalised proof of the Disjunction Property of Intuitionistic Logic
(originally proved by Goedel).Comment: continuation of arXiv:1201.3667; extended start of Section 1 and 2.1;
extended paragraph after Fact 1; dropped the N-rule as primitive and proved
it derivable; other, non-intuitionistic family members: arXiv:1208.1842,
arXiv:1208.591
Knowledge revision in systems based on an informed tree search strategy : application to cartographic generalisation
Many real world problems can be expressed as optimisation problems. Solving
this kind of problems means to find, among all possible solutions, the one that
maximises an evaluation function. One approach to solve this kind of problem is
to use an informed search strategy. The principle of this kind of strategy is
to use problem-specific knowledge beyond the definition of the problem itself
to find solutions more efficiently than with an uninformed strategy. This kind
of strategy demands to define problem-specific knowledge (heuristics). The
efficiency and the effectiveness of systems based on it directly depend on the
used knowledge quality. Unfortunately, acquiring and maintaining such knowledge
can be fastidious. The objective of the work presented in this paper is to
propose an automatic knowledge revision approach for systems based on an
informed tree search strategy. Our approach consists in analysing the system
execution logs and revising knowledge based on these logs by modelling the
revision problem as a knowledge space exploration problem. We present an
experiment we carried out in an application domain where informed search
strategies are often used: cartographic generalisation.Comment: Knowledge Revision; Problem Solving; Informed Tree Search Strategy;
Cartographic Generalisation., Paris : France (2008
Expert systems and finite element structural analysis - a review
Finite element analysis of many engineering systems is practised more as an art than as a science . It involves high level expertise (analytical as well as heuristic) regarding problem modelling (e .g. problem specification,13; choosing the appropriate type of elements etc .), optical mesh design for achieving the specified accuracy (e .g . initial mesh selection, adaptive mesh refinement), selection of the appropriate type of analysis and solution13; routines and, finally, diagnosis of the finite element solutions . Very often such expertise is highly dispersed and is not available at a single place with a single expert. The design of an expert system, such that the necessary expertise is available to a novice to perform the same job even in the absence of trained experts, becomes an attractive proposition. 13; In this paper, the areas of finite element structural analysis which require experience and decision-making capabilities are explored . A simple expert system, with a feasible knowledge base for problem modelling, optimal mesh design, type of analysis and solution routines, and diagnosis, is outlined. Several efforts in these directions, reported in the open literature, are also reviewed in this paper
The Calumet Area Hazardous Substance Data Base : A User's Guide with Documentation
HWRIC Project 88/89-055NTIS PB90-25193
- …