@inproceedings{Vidal2009TPCG,
author = {F. P. Vidal and M. Garnier and N. Freud and J. M. L\'etang and N. W. John},
title = {Simulation of {X-ray} Attenuation on the {GPU}},
booktitle = {Proceedings of Theory and Practice of Computer Graphics 2009},
year = 2009,
pages = {25-32},
month = jun,
address = {Cardiff, UK},
annotation = {Jun~17--19, 2009},
note = {Winner of Ken Brodlie Prize for Best Paper},
doi = {10.2312/LocalChapterEvents/TPCG/TPCG09/025-032},
abstract = {In this paper, we propose to take advantage of computer graphics hardware
to achieve an accelerated simulation of X-ray transmission imaging,
and we compare results with a fast and robust software-only implementation.
The running times of the GPU and CPU implementations are compared
in different test cases. The results show that the GPU implementation
with full floating point precision is faster by a factor of about
60 to 65 than the CPU implementation, without any significant loss
of accuracy. The increase in performance achieved with GPU calculations
opens up new perspectives. Notably, it paves the way for physically-realistic
simulation of X-ray imaging in interactive time.},
keywords = {Physically based modeling, Raytracing, Physics},
publisher = {Eurographics Association},
pdf = {./pdf/Vidal2009TPCG.pdf}
}
@inproceedings{Sinha2009UKRC,
author = {A. Sinha and S. Johnson and C. Hunt and H. Woolnough and F. P. Vidal and D. Gould},
title = {Preliminary face and content validation of {Imagine-S}: the {CIRSE} \& {BSIR} Experience},
booktitle = {Proceedings of the UK Radiological Congress},
year = 2009,
pages = {2},
month = jun,
address = {Manchester, UK},
annotation = {Jun~8--10, 2009},
abstract = {KEY LEARNING OBJECTIVES:
To determine face and content validity of a physics-based virtual reality (VR)
training simulation of visceral interventional radiology needle puncture procedures.
DESCRIPTION:
Imaging-guided needle puncture procedures use hand-eye coordination to direct needles,
wires and catheters to perform nephrostomy. The visuo-spatial and manipulation skills
required are learnt in a traditional apprenticeship, though Working Time
Directives are reducing the time and case mix available to train. Animal and
fixed models reproduce some training objectives, though they are an imperfect
substitute for the `real patient' experience. ImaGiNe-S is a computer-based
VR training simulation, using variable virtual environments with stereo
3D visual representation and devices to convey feel, realistically mimicking
a percutaneous nephrostomy procedure. With ethical approval, a prospective pilot
study was conducted at two international conferences to assess validity of
Imagine-S. 53 subjects (49 male, 4 female: 30 trainees and 23 subject matter
experts), underwent baseline testing on a simulated percutaneous nephrostomy.
Face and content validation were assessed using a 5-point Likert scale.
Outcomes showed that 41/53 (78\%) participants thought that the design of
Imagine-S was moderately realistic with content validity being rated averagely
for all critical task steps. 44/53 (83\%) participants thought that Imagine-S
is a useful model for training skills for nephrostomy.
CONCLUSION:
Imagine-S may be a useful model for training skills for nephrostomy. With further
development it may allow trainees to develop basic skills of percutaneous renal
collecting system access. Further assessment of face and content validity is needed.}
}
@inproceedings{Villard2008UKRC,
author = {P. F. Villard and P. Littler and V. Gough and F. P. Vidal and C.
Hughes and N. W. John and V. Luboz and F. Bello and Y. Song and R.
Holbrey and A. Bulpitt and D. Mullan and N. Chalmers and D. Kessel
and D. Gould},
title = {Improving the modeling of Medical Imaging data for simulation},
booktitle = {Proceedings of the UK Radiological Congress},
year = 2008,
pages = {61},
month = jun,
address = {Birmingham, UK},
annotation = {Jun~2--4, 2008},
abstract = {PURPOSE-MATERIALS:
To use patient imaging as the basis for developing virtual environments (VE).
BACKGROUND
Interventional radiology basic skills are still taught in an apprenticeship
in patients, though these could be learnt in high fidelity simulations using
VE. Ideally, imaging data sets for simulation of image-guided procedures
would alter dynamically in response to deformation forces such as respiration
and needle insertion. We describe a methodology for deriving such dynamic
volume rendering from patient imaging data.
METHODS
With patient consent, selected, routine imaging (computed tomography,
magnetic resonance, ultrasound) of straightforward and complex anatomy and
pathology was anonymised and uploaded to a repository at Bangor University.
Computer scientists used interactive segmentation processes to label target
anatomy for creation of a surface (triangular) and volume (tetrahedral) mesh.
Computer modeling techniques used a mass spring algorithm to map tissue
deformations such as needle insertion and intrinsic motion (e.g. respiration).
These methods, in conjunction with a haptic device, provide output forces in
real time to mimick the ‘feel’ of a procedure. Feedback from trainees and
practitioners was obtained during preliminary demonstrations.
RESULTS
Data sets were derived from 6 patients and converted into deformable VEs.
Preliminary content validation studies of a framework developed for training
on liver biopsy procedures, demonstrated favourable observations that are
leading to further revisions, including implementation of an immersive VE.
CONCLUSION:
It is possible to develop dynamic volume renderings from static patient data
sets and these are likely to form the basis of future simulations for
IR training of procedural interventions.}
}
@inproceedings{Cosson2004UKRC,
author = {P. Cosson and J. {Yu Cheng} and S. Keswani and G. Debouzy and D.
Deprez and F. Vidal},
title = {Virtual radiographic environments become a reality},
booktitle = {Proceedings of the UK Radiological Congress},
year = 2004,
month = jun,
address = {Manchester, UK},
annotation = {Jun~6--8, 2004}
}
@inproceedings{Cosson2004ALTC,
author = {P. Cosson and G. Debouzy and D. Deprez and F. Vidal and S. Keswani
and J. Warren},
title = {Virtual radiographic environments: what use would they be?},
booktitle = {University of Teesside Annual Learning \& Teaching Conference},
year = 2004,
month = jan,
address = {Middlesbrough, UK},
annotation = {Jan~15, 2004}
}
@inproceedings{Cosson2003ALTC,
author = {P. Cosson and G. Debouzy and D. Deprez and F. Vidal and S. Keswani
and J. Warren},
title = {Virtual radiographic environments},
booktitle = {University of Teesside Annual Learning \& Teaching Conference},
year = 2003,
address = {Middlesbrough, UK}
}
This file was generated by bibtex2html 1.97.
