1. 12 Oct, 2016 1 commit
  2. 06 Oct, 2016 1 commit
  3. 05 Oct, 2016 1 commit
  4. 09 Sep, 2016 1 commit
    • Robbert Krebbers's avatar
      Support for specialization of P₁ -★ .. -★ Pₙ -★ Q where Q is persistent. · 090aaea3
      Robbert Krebbers authored
      Before this commit, given "HP" : P and "H" : P -★ Q with Q persistent, one
      could write:
      
        iSpecialize ("H" with "#HP")
      
      to eliminate the wand in "H" while keeping the resource "HP". The lemma:
      
        own_valid : own γ x ⊢ ✓ x
      
      was the prototypical example where this pattern (using the #) was used.
      
      However, the pattern was too limited. For example, given "H" : P₁ -★ P₂ -★ Q",
      one could not write iSpecialize ("H" with "#HP₁") because P₂ -★ Q is not
      persistent, even when Q is.
      
      So, instead, this commit introduces the following tactic:
      
        iSpecialize pm_trm as #
      
      which allows one to eliminate implications and wands while being able to use
      all hypotheses to prove the premises, as well as being able to use all
      hypotheses to prove the resulting goal.
      
      In the case of iDestruct, we now check whether all branches of the introduction
      pattern start with an `#` (moving the hypothesis to the persistent context) or
      `%` (moving the hypothesis to the pure Coq context). If this is the case, we
      allow one to use all hypotheses for proving the premises, as well as for proving
      the resulting goal.
      090aaea3
  5. 09 Aug, 2016 1 commit
  6. 08 Aug, 2016 1 commit
  7. 05 Aug, 2016 3 commits
    • Robbert Krebbers's avatar
    • Robbert Krebbers's avatar
      More introduction patterns. · 4d8c4ac8
      Robbert Krebbers authored
      Also make those for introduction and elimination more symmetric:
      
        !%   pure introduction         %        pure elimination
        !#   always introduction       #        always elimination
        !>   later introduction        > pat    timeless later elimination
        !==> view shift introduction   ==> pat  view shift elimination
      4d8c4ac8
    • Robbert Krebbers's avatar
      Iris 3.0: invariants and weakest preconditions encoded in the logic. · 1f589858
      Robbert Krebbers authored
      This commit features:
      
      - A simpler model. The recursive domain equation no longer involves a triple
        containing invariants, physical state and ghost state, but just ghost state.
        Invariants and physical state are encoded using (higher-order) ghost state.
      
      - (Primitive) view shifts are formalized in the logic and all properties about
        it are proven in the logic instead of the model. Instead, the core logic
        features only a notion of raw view shifts which internalizing performing frame
        preserving updates.
      
      - A better behaved notion of mask changing view shifts. In particular, we no
        longer have side-conditions on transitivity of view shifts, and we have a
        rule for introduction of mask changing view shifts |={E1,E2}=> P with
        E2 ⊆ E1 which allows to postpone performing a view shift.
      
      - The weakest precondition connective is formalized in the logic using Banach's
        fixpoint. All properties about the connective are proven in the logic instead
        of directly in the model.
      
      - Adequacy is proven in the logic and uses a primitive form of adequacy for
        uPred that only involves raw views shifts and laters.
      
      Some remarks:
      
      - I have removed binary view shifts. I did not see a way to describe all rules
        of the new mask changing view shifts using those.
      - There is no longer the need for the notion of "frame shifting assertions" and
        these are thus removed. The rules for Hoare triples are thus also stated in
        terms of primitive view shifts.
      
      TODO:
      
      - Maybe rename primitive view shift into something more sensible
      - Figure out a way to deal with closed proofs (see the commented out stuff in
        tests/heap_lang and tests/barrier_client).
      1f589858
  8. 02 Aug, 2016 1 commit
  9. 27 Jul, 2016 2 commits
    • Robbert Krebbers's avatar
      Declare inG arguments of own_* implicit but not maximally inserted. · de9438cb
      Robbert Krebbers authored
      This way type class inference is not invokved when used in tactics
      like iPvs while not having to write an @.
      
      (Idea suggested by Ralf.)
      de9438cb
    • Robbert Krebbers's avatar
      Make type class inference for inG less eager. · a0348d7c
      Robbert Krebbers authored
      This way, it won't pick arbitrary (and possibly wrong!) inG instances
      when multiple ones are available. We achieve this by declaring:
      
        Hint Mode inG - - +
      
      So that type class inference only succeeds when the type of the ghost
      variable does not include any evars.
      
      This required me to make some minor changes throughout the whole
      development making some types explicit.
      a0348d7c
  10. 13 Jul, 2016 1 commit
  11. 03 Jul, 2016 2 commits
  12. 16 Jun, 2016 3 commits
  13. 01 Jun, 2016 1 commit
  14. 31 May, 2016 3 commits
  15. 28 May, 2016 1 commit
    • Robbert Krebbers's avatar
      CMRAs with partial cores. · cfb00b3e
      Robbert Krebbers authored
      Based on an idea and WIP commits of J-H. Jourdan: the core of a CMRA
      A is now a partial function A → option A.
      
      TODO: define sum CMRA
      TODO: remove one shot CMRA and define it in terms of sum
      cfb00b3e
  16. 27 May, 2016 1 commit
  17. 24 May, 2016 1 commit
    • Robbert Krebbers's avatar
      Make specialization patterns for persistent premises more uniform. · 65bfa071
      Robbert Krebbers authored
      Changes:
      - We no longer have a different syntax for specializing a term H : P -★ Q whose
        range P or domain Q is persistent. There is just one syntax, and the system
        automatically determines whether either P or Q is persistent.
      - While specializing a term, always modalities are automatically stripped. This
        gets rid of the specialization pattern !.
      - Make the syntax of specialization patterns more consistent. The syntax for
        generating a goal is [goal_spec] where goal_spec is one of the following:
      
          H1 .. Hn : generate a goal using hypotheses H1 .. Hn
         -H1 .. Hn : generate a goal using all hypotheses but H1 .. Hn
                 # : generate a goal for the premise in which all hypotheses can be
                     used. This is only allowed when specializing H : P -★ Q where
                     either P or Q is persistent.
                 % : generate a goal for a pure premise.
      65bfa071
  18. 07 May, 2016 1 commit
  19. 15 Mar, 2016 1 commit
  20. 11 Mar, 2016 1 commit
  21. 10 Mar, 2016 2 commits
  22. 08 Mar, 2016 1 commit
  23. 07 Mar, 2016 1 commit
    • Ralf Jung's avatar
      Add both non-expansive and contractive functors, and bundle them for the... · 2467bf21
      Ralf Jung authored
      Add both non-expansive and contractive functors, and bundle them for the general Iris instance as well as the global functor construction
      
      This allows us to move the \later in the user-defined functor to any place we want.
      In particular, we can now have "\later (iProp -> iProp)" in the ghost CMRA.
      2467bf21
  24. 06 Mar, 2016 2 commits
  25. 05 Mar, 2016 1 commit
  26. 02 Mar, 2016 1 commit
  27. 01 Mar, 2016 1 commit
  28. 27 Feb, 2016 1 commit
  29. 25 Feb, 2016 2 commits