Compared to conventional interfaces, image guided surgery (IGS) interfaces contain a richer variety and more complex objects and interaction types. The main interactive characteristics emering from systems like this is the interaction focus shared between physical space, where the surgeon interacts with the patient using surgical tools, and with the digital world, where the surgeon interacts with the system. This limitation results in two different interfaces likely inconsistent, thereby the interaction discontinuities do break the natuarl workflow forcing the user to switch between the operation modes. Our work addresses these features by focusing on the model, interaction and ergonomic integrity analysis considering the Augmented Reality paradigm applied to IGS procedures and more specifically applied to the Neurosurgery study case. We followed a methodology according to the model-based approach, including new extensions in order to support interaction technologies and to sensure continuity interaction according to the IGS system requirements. As a result, designers may as soon as possible discover errors in the development process and may perform an efficient interface design coherently integrating constraints favoring continuity instead of discrete interaction with possible inconsistencies.