| アイテムタイプ |
学術雑誌論文 / Journal Article(1) |
| 公開日 |
2025-08-27 |
| タイトル |
|
|
タイトル |
ChatGPT-4 Responses on Ankle Cartilage Surgery Often Diverge from Expert Consensus: A Comparative Analysis. |
|
言語 |
en |
| 言語 |
|
|
言語 |
eng |
| キーワード |
|
|
言語 |
en |
|
キーワード |
Artificial intelligence/ChatGPT/ankle cartilage injury/bone marrow stimulation/consensus statement |
| 資源タイプ |
|
|
資源タイプ |
journal article |
| アクセス権 |
|
|
アクセス権 |
open access |
| 著者 |
横江, 琢示
WEKO
34429
e-Rad_Researcher
50895894
| ja |
横江, 琢示
宮崎大学
|
| ja-Kana |
ヨコエ, タクジ
|
| en |
Yokoe, Takuji
University of Miyazaki
|
Search repository
Roversi, Giulia
Sevivas, Nuno
亀井, 直輔
WEKO
35566
e-Rad_Researcher
70444685
| ja |
亀井, 直輔
宮崎大学
|
| ja-Kana |
カメイ, ナオスケ
|
| en |
Kamei, Naosuke
University of Miyazaki
|
Search repository
Diniz, Pedro
Pereira, Hélder
|
| 抄録 |
|
|
内容記述タイプ |
Abstract |
|
内容記述 |
There are few studies that have evaluated whether large language models, such as ChatGPT, can provide accurate guidance to clinicians in the field of foot and ankle surgery. This study aimed to assess the accuracy of ChatGPT's responses regarding ankle cartilage repair by comparing them with the consensus statements from foot and ankle experts as a standard reference./The open artificial intelligence (AI) model ChatGPT-4 was asked to answer a total of 14 questions on debridement, curettage, and bone marrow stimulation for ankle cartilage lesions that were selected at the 2017 International Consensus Meeting on Cartilage Repair of the Ankle. The ChatGPT responses were compared with the consensus statements developed in this international meeting. A Likert scale (scores, 1-5) was used to evaluate the similarity of the answers by ChatGPT to the consensus statements. The 4 scoring categories (Accuracy, Overconclusiveness, Supplementary, and Incompleteness) were also used to evaluate the quality of ChatGPT answers, according to previous studies./The mean Likert scale score regarding the similarity of ChatGPT's answers to the consensus statements was 3.1 ± 0.8. Regarding the results of 4 scoring categories of the ChatGPT answers, the percentages of answers that were considered "yes" in the Accuracy, Overconclusiveness, Supplementary, and Incompleteness were 71.4% (10/14), 35.7% (5/14), 78.6% (11/14), and 14.3% (2/14), respectively./This study showed that ChatGPT-4 often provides responses that diverge from expert consensus regarding surgical treatment of ankle cartilage lesions./Level V, expert opinion. |
|
言語 |
en |
| 書誌情報 |
en : Foot & ankle orthopaedics
巻 10,
号 3,
p. 24730114251352494,
発行日 2025-07
|
| 出版者 |
|
|
出版者 |
Sage |
| ISSN |
|
|
収録物識別子タイプ |
EISSN |
|
収録物識別子 |
24730114 |
| DOI |
|
|
関連タイプ |
isVersionOf |
|
|
識別子タイプ |
DOI |
|
|
関連識別子 |
https://doi.org/10.1177/24730114251352494 |
| 著者版フラグ |
|
|
出版タイプ |
VoR |