6
Abstract— The paper presented herein describes the development of an advanced robotized system applied to in vitro implant dentistry research. To the biomedical community this paper shows the possibilities of industrial robots to help the research. Robots are special suitable to the biomedical field specially when integrated with advanced sensors and technologies, which may facilitate both the programming tasks and the data acquisition. Robotics researchers will find in this paper one of the first applications of programming by demonstration with real users with a novel explicit robot programming technique making use of multi-camera vision combined with speech recognition. This programming method targets users who have minimal robot experience but aims at ‘teaching’ the robot to execute a specific task. Keywords – programming-by-demonstration; service-oriented architectures; dentistry I. INTRODUCTION esearch in implant dentistry is significantly constrained by the availability of reliable in vivo data come from clinical cases. The inherent differences between patients and the high variability of the implanting procedure turns collected data more imprecise and difficult to work with. To overcome this limitation we developed a versatile robotic system capable of executing several research procedures mimicking all the process of the implant technique and some of the living functions such as mastication. Robotics in medical research is not a novelty. Robots perform different tasks like automatic dispensing in pharmaceutical labs, mechanical tests of spinal joints [1] or stress/strain analysis in implant dentistry [2]. The choice of the robot as the principal element of the research system was driven by the needs of flexibility and repeatability. Assembled with a force/torque sensor and with appropriate software [3] this system presents high repeatability: less than 0.01 mm in position and less than 0.1 N in force. The global accuracy of the system enables highly reliable Manuscript received January 31, 2012. procedures that are being used, with minimum data, to validate different tests regarding misalignment caused by impression techniques, implants placing and mastication force distribution under numerous scenarios. It should be noticed that the complexity associated with the research procedures promotes the interleaving between the researcher (dentist) and the robot: the robot executes operations when high degree of repeatability is required, like drilling, or highly repetitive tasks like photo-shooting p.e. On the other hand, the human (researcher) executes tasks non-critical for the reliability of the research procedures and also complex manipulation tasks like screwing the prostheses p. e.. Robot programming in this scenario is mainly a two step procedure: In the first step the position of one feature is determined by the researcher and afterwards the robots performs the operation with a set of different parameters. For mastication for example the researcher determines the position of the tool in the teeth and the robot applies the force in that position with predefined directions and forces. Our first experiments shown that the second step can easily be performed independently by the researcher with the help of appropriate software. Nonetheless, for the first step the presence of a robotics expert is mandatory. The alternative is the researcher program himself the robot, which represents a great challenge since it has non expertise in the field, is focused on other demanding tasks (research) and have to move his attention to do extensive reprogramming and adjustment. Although offline programming tools have their important (growing) market, industrial robots are still very dependent on online teach pendant programming. This technology has suffered a great evolution over the past years [4] but the basic concept is still the same, and it is not well adapted to the scenario described above. Therefore, new paradigms for online robot programming are needed, with special focus on explicit programming techniques that can bring the worker or researcher closer to the robot. Recent developments On the use of Robotics in implant dentistry research. Germano Veiga 1* , Francisco Caramelo 2 , Pedro Malaca 3 , Pedro Brito 2,4 and J N Pires 1 , 1 Mechanical Engineering Department, Faculty of Sciences and Technology, University of Coimbra, Portugal 2 Institute of Biomedical Research on Light and Image -Faculty of Medicine, University of Coimbra, Portugal 3 Sarkkis Robotics Lda, Coimbra, Portugal 4 Institute Health Sciences of Porto, Porto, Portugal *Present address: INESC Porto, Porto, Portugal R The Fourth IEEE RAS/EMBS International Conference on Biomedical Robotics and Biomechatronics Roma, Italy. June 24-27, 2012 978-1-4577-1198-5/12/$26.00 ©2012 IEEE 1596

On the Use of Robotics in Implant Dentistry Researchvigir.missouri.edu/~gdesouza/Research/Conference... · impression (methods and materials, see Fig. 1) that avoid s the operator

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: On the Use of Robotics in Implant Dentistry Researchvigir.missouri.edu/~gdesouza/Research/Conference... · impression (methods and materials, see Fig. 1) that avoid s the operator

Abstract— The paper presented herein describes the

development of an advanced robotized system applied to in

vitro implant dentistry research. To the biomedical community

this paper shows the possibilities of industrial robots to help

the research. Robots are special suitable to the biomedical field

specially when integrated with advanced sensors and

technologies, which may facilitate both the programming tasks

and the data acquisition. Robotics researchers will find in this

paper one of the first applications of programming by

demonstration with real users with a novel explicit robot

programming technique making use of multi-camera vision

combined with speech recognition. This programming method

targets users who have minimal robot experience but aims at

‘teaching’ the robot to execute a specific task.

Keywords – programming-by-demonstration; service-oriented

architectures; dentistry

I. INTRODUCTION

esearch in implant dentistry is significantly constrained

by the availability of reliable in vivo data come from

clinical cases. The inherent differences between patients

and the high variability of the implanting procedure turns

collected data more imprecise and difficult to work with. To

overcome this limitation we developed a versatile robotic

system capable of executing several research procedures

mimicking all the process of the implant technique and

some of the living functions such as mastication.

Robotics in medical research is not a novelty. Robots

perform different tasks like automatic dispensing in

pharmaceutical labs, mechanical tests of spinal joints [1] or

stress/strain analysis in implant dentistry [2]. The choice of

the robot as the principal element of the research system

was driven by the needs of flexibility and repeatability.

Assembled with a force/torque sensor and with appropriate

software [3] this system presents high repeatability: less

than 0.01 mm in position and less than 0.1 N in force. The

global accuracy of the system enables highly reliable

Manuscript received January 31, 2012.

procedures that are being used, with minimum data, to

validate different tests regarding misalignment caused by

impression techniques, implants placing and mastication

force distribution under numerous scenarios.

It should be noticed that the complexity associated with

the research procedures promotes the interleaving between

the researcher (dentist) and the robot: the robot executes

operations when high degree of repeatability is required,

like drilling, or highly repetitive tasks like photo-shooting

p.e. On the other hand, the human (researcher) executes

tasks non-critical for the reliability of the research

procedures and also complex manipulation tasks like

screwing the prostheses p. e..

Robot programming in this scenario is mainly a two step

procedure: In the first step the position of one feature is

determined by the researcher and afterwards the robots

performs the operation with a set of different parameters.

For mastication for example the researcher determines the

position of the tool in the teeth and the robot applies the

force in that position with predefined directions and forces.

Our first experiments shown that the second step can easily

be performed independently by the researcher with the help

of appropriate software. Nonetheless, for the first step the

presence of a robotics expert is mandatory. The alternative

is the researcher program himself the robot, which

represents a great challenge since it has non expertise in the

field, is focused on other demanding tasks (research) and

have to move his attention to do extensive reprogramming

and adjustment.

Although offline programming tools have their important

(growing) market, industrial robots are still very dependent

on online teach pendant programming. This technology has

suffered a great evolution over the past years [4] but the

basic concept is still the same, and it is not well adapted to

the scenario described above. Therefore, new paradigms for

online robot programming are needed, with special focus on

explicit programming techniques that can bring the worker

or researcher closer to the robot. Recent developments

On the use of Robotics in implant dentistry research.

Germano Veiga1*, Francisco Caramelo2, Pedro Malaca3, Pedro Brito2,4 and J N Pires1,

1Mechanical Engineering Department, Faculty of Sciences and Technology, University of Coimbra, Portugal

2 Institute of Biomedical Research on Light and Image -Faculty of Medicine, University of Coimbra, Portugal

3Sarkkis Robotics Lda, Coimbra, Portugal

4Institute Health Sciences of Porto, Porto, Portugal

*Present address: INESC Porto, Porto, Portugal

R

The Fourth IEEE RAS/EMBS International Conferenceon Biomedical Robotics and BiomechatronicsRoma, Italy. June 24-27, 2012

978-1-4577-1198-5/12/$26.00 ©2012 IEEE 1596

Page 2: On the Use of Robotics in Implant Dentistry Researchvigir.missouri.edu/~gdesouza/Research/Conference... · impression (methods and materials, see Fig. 1) that avoid s the operator

presented by a robot manufacturer have embedded high

speed force control feedback into the robot controller [5],

which provides good perspectives in terms of advanced

sensor integration, crucial to improve the man/machine

interaction. These systems have been around in research for

a long time but their industrialization opens possibilities in

mass products with advanced teaching techniques, namely

Programming-by-demonstration [6][7].

The original contribution of the present work seeks two

audiences, the research community in the biomedical area

and the robotics researchers dealing with human-robot

interaction. To the biomedical community this paper shows

the possibilities of industrial robots to perform research

specially when combined with advanced sensors and human

robot interface technologies. Furthermore, a novel

technique for the measurement of gaps between implants

and the prosthesis is described. The robotics researchers

will find in this paper one of the first applications of

programming by demonstration with real non-expertise

users, and the presentation of a novel explicit robot

programming technique making use of multi-camera vision

fused with speech recognition.

II. ROBOTICS IN IMPLANTOLOGY RESEARCH

A. Comparative study of impression materials

We developed a method for measuring the fitness of

prosthesis and the corresponding dental implants. Since a

robot arm is use for executing the dental impressions and

part of the measuring procedure we gain high precision and

reproducibility. This “in vitro” method for evaluating

impression (methods and materials, see Fig. 1) that avoids

the operator for executing both the impression and the

measuring.

Fig. 1 Robot performing the impression procedure.

The proposed method is based on the use of a robot arm for

performing the impression. For measuring the gap, between

a realistic mandible made of a photoelastic polymer and the

prosthesis obtain from the impression, we use the robot to

move the high quality photo camera and take 6 zoom photos

of each of the implants (Fig. 2).

Fig. 2 Left: schematic representation of the photo arrangement for

measuring the gap between the plaster mandibles and the zirconia dental

prosthesis. Right: photo of the actual arrangement.

Finally the results were analyzed with specifically

developed software (Fig. 3).

Fig. 3 Photo of one implant and part of the prosthesis. The gap is well

defined as the space between the red and the blues lines. The distance

between the blue and the green line is used as reference.

B. Photoelastic analysis of implant arrangement.

The mastication force distribution inside the mandible is

decisive in the durability of dental implants. The purpose of

this study is to do a qualitative descriptive analysis of stress

patterns around four different implants embedded in photo-

elastic mandible.

The industrial robot equipped with a force torque sensor

produces force controlled movements that mimic

mastication forces that are propagated through the implants.

Photo-elastic mandible acrylic resin models were die casted

with four implants placed vertically and parallel between

mentonian hole. Zirconia prosthesis was screwed to the

implants and afterwards subjected to different loads (100N

150N and 200N) in 12 points with an axial force and 30ª

tilted (vestibular and lingual face).

Photoelastic analysis was accomplished using a reflection

polariscope. The fringe patterns produced in the

photoelastic resin for each implant and load were

photographed with a digital camera. Fringe concentrations

and the highest fringe order were recorded and described

for the apical, central, and coronal regions of all the

implants for each load scenario.

1597

Page 3: On the Use of Robotics in Implant Dentistry Researchvigir.missouri.edu/~gdesouza/Research/Conference... · impression (methods and materials, see Fig. 1) that avoid s the operator

Fig. 4 Left: Robot performing mastication simulation under polarized

light. Right: Results for strain stress analysis. .

III. PROGRAMMING-BY-DEMONSTRATION

The interaction technologies used in the PbD system

presented in this work are an infrared stereo vision system

and a voice recognition system. The orchestration of these

systems together with the robot is achieved using a service-

oriented platform and a high level programming

environment. The description of the methodology and

implementation is presented in sections III.A, III.B, III.D

A. Stereo Vision

The definition of trajectories is one of the crucial aspects in

robot programming. Thus, accurate detection of position in

a known frame is essential to accomplish this task.

Although, there are several ways of performing such

assignment artificial vision seems to be the easiest to be

employed. However, when just one camera is used the

problem of finding the correspondence between one point

in the image and the point in the real world is undetermined

since the dimension related to depth is lost. If other

information is employed, such as other images taken from

different angles, the 3D information can be recovered by

proper triangulation.

In a broader sense, three-dimensional reconstruction

depends on the initial information that is available. If the

extrinsic and the intrinsic parameters of the camera are

known the triangulation can be solved unambiguously.

However, if only the intrinsic parameters are known,

reconstruction is still possible but lacking one scaling

factor. Even if intrinsic and extrinsic parameters are

unknown reconstruction can be still performed based on the

known correspondences but up to an unknown global

projective transformation [8].

In the present work we were interested in recovering the

complete 3D information, which was achieved by

triangulation [8].

Figure 1 shows a schematic representation of the process:

lines l and r passing through the points in the images ( pl

and rp ) and the corresponding center of projection ( Cl

and rC ) would intersect at point P in the absence of noise.

However, the presence of noise inhibits in most cases that

the intersection exists. Therefore, P must be computed as

the point that simultaneously minimizes the distance to both

lines l and r .

pl

rp

rC

Cl

P

r

l

Fig. 1. General case of triangulation used in stereo vision in which the

rays do not intersect in space. P is the midpoint of the line that is

perpendicular to both rays l and r. (Adapted from [8]).

Point P can be compute by:

2

T

e da p T b R pP

+ += (1)

where a and b are calculated from:

( )T T

e e d da p c p R p T b R p+ × = + . (2)

R and T are defined by:

T

rR R R=l

T

rT T R T= −l

(2)

rR and Rl are the rotation matrix that represent the

orientation of right and left cameras, respectively and rT

and Tl are the position of the centers of projection.

In order to determine the parameters of the cameras we

used the method presented by [9] that was implemented in

C#. In this method, a planar calibration object is placed in

different orientations and their images are used to obtain the

camera parameters. The parameters are initially estimated

with a closed form solution and then refined using an

optimization step. Radial lens distortions are also modeled.

1598

Page 4: On the Use of Robotics in Implant Dentistry Researchvigir.missouri.edu/~gdesouza/Research/Conference... · impression (methods and materials, see Fig. 1) that avoid s the operator

Fig. 5 The use of the infrared pen to define positions for mastication

simulation in a prosthesis held on a photo-elastic mandible.

B. Robot Jogging and robot services

The use of one single infrared LED provides significant

advantages in terms of maneuverability in small spaces and

in the required simplicity of use, but limits the jogging of

the robot to translations. To overcome this issue we

developed a simple strategy that uses the definition of a

rotation vector making use of two points, according to the

Fig. 6. The tool center point and the two points define a

plane and an angle. The robot reorients the specified angle

around the normal vector of the plane.

Fig. 6 Reorienting the robot with two points

It should be noticed that the translation movements of the

robot can be made in either incremental or point-to-point

mode (absolute positioning). The incremental movements

permit to overcome some accuracy limitation of the stereo

vision system.

The robot services were automatically generated from the

robot program using software previously developed in the

lab [10]. This software reads properly tagged robot

programs and automatically generates network services.

C. Speech recognition system.

With the introduction of the language models, of which

the n-gram [12] is by far the most used, speech recognition

systems achieved remarkable efficiency levels.

Despite the need for further work on the recognition of

conversational speech, speech recognition technologies

regarding read speech are mature and ready to complement

other means of human-machine interaction. These advances

leaded to the development of several commercial products

that are now reaching wider markets.

However, from a practical point of view, the most robust

(speaker-independent) voice-enabled systems present in the

market are still working in command mode, which means

that the recognition is made against a predefined set of

commands defined with special purpose grammars [13]

[14].

In this work we use voice recognition grammars with a set

of commands that allowed the researcher to complement the

operation through the stereo vision system using voice

commands to change the jogging mode and activate other

functions (like photo shooting p.e.). The implementation of

the network service for the voice recognition also uses

automatic generation. Instead of robot programs this

automatic generation uses the grammar from the voice

recognition system (see [11] for details.).

D. Orchestration of services.

The concept behind the orchestration of services (or

composition in a larger sense, according to [15]) is derived

from the one introduced by distributed component-oriented

programming, which, like in many other WS*-related

technologies, was extended and standardized using XML

specifications. From the evolution process, a standard has

emerged as the most promising: Business Process

Execution Language for Web Services (BPEL4WS) [16].

The integration of the voice recognition system and the

infrared vision system was based on a service-oriented

architecture using two well know platforms (UPnP [17] and

DPWS [18]). In these platforms, network services are

composed by events and actions and their integration of the

programming-by-demonstration system is a partially

unsolved problem. In the context of constrained purpose

devices (like TV’s, p.e.) the use of such architectures is a

success, but in the flexible scenario of a robotic workcell

their major advantages is their integration with flexible

integration languages/platforms. As described in [11], in

spite of the existence of long-term consistent efforts on the

specification of languages for the orchestration of services

[19] [20], the event-driven nature of the robotic setup and

the inadequacy of business level orchestration languages

1599

Page 5: On the Use of Robotics in Implant Dentistry Researchvigir.missouri.edu/~gdesouza/Research/Conference... · impression (methods and materials, see Fig. 1) that avoid s the operator

(like BPEL4WS) leads us to the use of state-chart based

language and the respective environment previously

developed on our lab (SidneyChart [21]) to orchestrate our

services.

Fig. 7 SidneyChart ([21]) interface with the statechart of the PbD system.

The easy-of-use of the SidneyChart allowed the

researcher (dentist) to perform himself some orchestration

whenever a service changed, due to the change of the voice

recognition grammar p.e..

IV. RESULTS

Concerning the stereo vision system the absolute error

obtained with the present setup makes difficult the use of

the robot to the intended application in absolute mode, but

the incremental mode performance was satisfactory from

the point of view of the researcher.

The performance of the voice recognition system is very

high (>95%), due to the low number of commands used and

the quality of the microphone.

V. CONCLUSION

The evaluation of the developed equipment in this work

can be made from different perspectives. From the

implantologist/researcher point of view the equipment

provides the flexibility and repeatability required for the

research procedures: 0.01mm in position, less than 0.1 N in

force and when equipped with a high resolution camera is

able to measure down to 1 micrometer of gap between

implant and prosthesis. The global accuracy achieved with

this system enables highly reliable procedures that are

currently being used to validate different tests regarding

implant misalignment caused by different impression

techniques and mastication force distributions under

numerous implants placing scenarios.

From the Human-robot interaction perspective this paper

describes a programming-by-demonstration technique and

its evaluation with a real user (the researcher). The results

are very satisfactory. The researcher is now able to define

new research tasks independently and even perform some

orchestration of network services. This experiment shows

that with proper tools in terms of human robot interface and

advanced software tools for the integration of resources

robots can perform a wide variety of tasks in new markets.

ACKNOWLEDGMENT

REFERENCES

[1] H. de Visser, C. Rowe, and M. J. Pearcy, “Robotic

Testing Facility for the Measurement of the Mechanics

of Spinal Joints,” Proc ImechE Part H: J Eng in Med,

vol. 221, no. 3, pp. 221-227, 2007.

[2] J. N. Pires, F. J. Caramelo, P. Brito, J. Santos, and M.

F. Botelho, “Robotics in implant dentistry: stress/strain

analysis. System overview and experiments,”

Industrial Robot: An International Journal, vol. 33,

pp. 373-380, Sep. 2006.

[3] J. N. Pires and J. M. G. Sa da Costa, “Object-oriented

and distributed approach for programming robotic

manufacturing cells,” Robotics and Computer-

1600

Page 6: On the Use of Robotics in Implant Dentistry Researchvigir.missouri.edu/~gdesouza/Research/Conference... · impression (methods and materials, see Fig. 1) that avoid s the operator

Integrated Manufacturing, vol. 16, pp. 29-42, Feb.

2000.

[4] A. Kazi, J. Bunsendal, D. Haag, R. Baum, and R.

Bischoff, “Next Generation Teach Pendants for

Industrial Robots,” Advances in Human-Robot

Interaction, 2005. [Online]. Available:

http://www.springerlink.com/content/ac15e5b777e373

d5.

[5] ABB, “ABB IRC5 Application Manual Force Control

Machining FC,” Vasteras, 3hAC027595-001, 2007.

[6] Pires, J. N., Veiga, G., and Araujo, R., “Programming

by demonstration in the coworker scenario for SMEs,”

Industrial Robot, Emerald, vol. 36, no. 1, 2008.

[7] J. Hofschule, “Lead Through Programming,” 2008.

[Online]. Available:

http://www.smerobot.org/09_calendar/20071108_Press

etagABB-Ladenburg_LeadThrough.pdf. [Accessed:

21-May-2009].

[8] E. Trucco and A. Verri, Introductory Techniques for

3-D Computer Vision. Prentice Hall PTR, 1998.

[9] Z. Zhang, “A Flexible New Technique for Camera

Calibration,” IEEE Trans. Pattern Anal. Mach. Intell.,

vol. 22, no. 11, pp. 1330-1334, 2000.

[10] G. Veiga and J. . Pires, “Using robot programming

languages to specify service contracts,” in Proceedings

of the CONTROLO’2008 Conference, 2008.

[11] G. Veiga, P. Malaca, K. Nilsson, and J. N. Pires,

“Separation of concerns on the orchestration of

operations in flexible manufacturing,” Assembly, vol.

32, no. 1, 2011.

[12] C. Manning and H. Schütze, Foundations of

Statistical Natural Language Processing. The MIT

Press, 1999.

[13] “Voice Extensible Markup Language (VoiceXML)

Version 2.0,” 2010. [Online]. Available:

http://www.w3.org/TR/voicexml20/. [Accessed: 24-

Nov-2010].

[14] “Speech Recognition Grammar Specification Version

1.0,” 2010. [Online]. Available:

http://www.w3.org/TR/speech-grammar/. [Accessed:

23-Nov-2010].

[15] C. Pautasso and G. Alonso, “Visual composition of

web services,” 2003, pp. 92-99.

[16] “OASIS Web Services Business Process Execution

Language (WSBPEL) TC,” 2010. [Online]. Available:

http://www.oasis-

open.org/committees/tc_home.php?wg_abbrev=wsbpel

. [Accessed: 18-Dec-2008].

[17] “Universal Plug and Play.” [Online]. Available:

http://www.microsoft.com/whdc/device/media/upnp.ms

px. [Accessed: 01-Dec-2008].

[18] “OASIS Devices Profile for Web Services (DPWS),”

2010. [Online]. Available: http://docs.oasis-

open.org/ws-dd/ns/dpws/2009/01. [Accessed: 09-Mar-

2010].

[19] F. Jammes, H. Smit, J. L. M. Lastra, and I. M.

Delamer, “Orchestration of service-oriented

manufacturing processes,” presented at the Emerging

Technologies and Factory Automation, 2005. ETFA

2005. 10th IEEE Conference on, 2005, vol. 1, p. 8 pp.-

624.

[20] A. W. Colombo, F. Jammes, H. Smit, R. Harrison, J.

L. M. Lastra, and I. M. Delamer, “Service-oriented

architectures for collaborative automation,” presented

at the Industrial Electronics Society, 2005. IECON

2005. 31st Annual Conference of IEEE, 2005, p. 6 pp.

[21] P. Malaca, G. Veiga, and N. Pires, “SidneyChart: A

statechart GUI for SOA orchestration in autonomous

industrial systems,” in Industrial Informatics (INDIN),

2011 9th IEEE International Conference on, 2011, pp.

238-243.

1601