bims-librar Biomed News
on Biomedical librarianship
Issue of 2025–07–27
24 papers selected by
Thomas Krichel, Open Library Society



  1. SAGE Open Nurs. 2025 Jan-Dec;11:11 23779608251361012
       Background: Researchers have traditionally used commercial databases such as Scopus and Web of Science to track scholarly literature and ascertain the impact of publications in terms of citations. In recent years, new open data sources, such as OpenAlex, have appeared, which broaden the array of tools accessible for exploring academic literature.
    Objective: This study compares the coverage of scholarly nursing literature by OpenAlex, Scopus, and Web of Science in terms of journals, articles, and citations indexed.
    Methods: A list of nursing journals indexed by OpenAlex, Scopus, and Web of Science was compiled to compare the titles covered in terms of country of publication, publisher, and open access availability. Citations of nursing articles published between 2000 and 2022 in the three sources were also analyzed.
    Results: OpenAlex indexes nearly all the journals covered by Scopus and Web of Science, plus additional sources from countries outside the Western world. In terms of individual articles, OpenAlex indexes more citations than Scopus and Web of Science.
    Conclusion: OpenAlex is a comprehensive source for tracking nursing literature. Currently, its main limitation lies in the lack of a user-friendly web interface, as the need to use an API for data retrieval can complicate the search process for end users.
    Keywords:  bibliometrics; citation analysis; databases
    DOI:  https://doi.org/10.1177/23779608251361012
  2. Cureus. 2025 Jun;17(6): e86543
       BACKGROUND: The internet has become a primary source of health information for the public, with important implications for patient decision-making and public health outcomes. However, the quality and readability of this content vary widely. With the rise of generative artificial intelligence (AI) tools such as ChatGPT and Gemini, new challenges and opportunities have emerged in how patients access and interpret medical information.
    OBJECTIVE: To evaluate and compare the quality, credibility, and readability of consumer health information provided by traditional search engines (Google, Bing) and generative AI platforms (ChatGPT, Gemini) using three validated instruments: DISCERN, JAMA Benchmark Criteria, and Flesch-Kincaid Readability Metrics.  Methods: Twenty health-related webpages from each platform were collected using a standardized query across Google, Bing, Gemini, and ChatGPT. Each source was assessed independently by two reviewers using the DISCERN instrument and the adapted JAMA benchmark criteria. Readability was evaluated using the Flesch Reading Ease and Grade Level scores. One-way ANOVA with Bonferroni correction was used to compare platform performance, and Cohen's Kappa measured inter-rater reliability.  Results: Google achieved the highest mean scores for both quality and credibility (DISCERN: 3.33 ± 0.53; JAMA: 3.70 ± 0.44), followed by Bing, Gemini, and ChatGPT. ChatGPT received the lowest scores across all quality measures. Readability analysis revealed no statistically significant differences between platforms; however, all content exceeded recommended reading levels for public health information. Cohen's Kappa indicated strong inter-rater agreement across DISCERN items.  Conclusion: Google remains the most reliable source of high-quality, readable health information among the evaluated platforms. Generative AI tools such as ChatGPT and Gemini, while increasingly popular, exhibited significant limitations in accuracy, transparency, and complexity. These findings highlight the need for improved oversight, transparency, and user education regarding AI-generated health content.
    Keywords:  chatgpt; consumer health information; discern; gemini; generative ai; google; misinformation; readibility; rosacea; search engines
    DOI:  https://doi.org/10.7759/cureus.86543
  3. Cureus. 2025 Jun;17(6): e86491
      Background The purpose of this study was to evaluate the potential role of artificial intelligence, specifically ChatGPT-4, in generating patient education materials (PEMs) for total knee arthroplasty (TKA). The aim of our study was to compare the quality and readability of PEMs for TKA generated by ChatGPT-4 with those created by human experts to assess the potential for the use of AI in patient education. Materials and methods We assessed the quality and readability of TKA PEMs produced by ChatGPT-4 and five reputable human-generated websites. Readability was compared using Flesch-Kincaid Grade Level and Flesch Reading Ease. The quality of information was compared using the DISCERN criteria. Results ChatGPT-4 PEMs demonstrated a significantly higher reading grade level and lower reading ease score compared to human-generated PEMs (<0.001). Conclusions The utility of ChatGPT-4 for producing TKA PEMs is promising. Notably, the quality and reliability are as good as human-generated resources. However, it is currently limited by readability issues, leading to a recommendation against its use. Future AI enhancements should prioritize readability to ensure information is more accessible. Effective collaboration between AI developers and healthcare professionals is vital for improving patient education outcomes.
    Keywords:  arthrosplasty; artificial intelligence; large language models; patient education; total knee replacement
    DOI:  https://doi.org/10.7759/cureus.86491
  4. Knee. 2025 Jul 17. pii: S0968-0160(25)00173-5. [Epub ahead of print]56 495-502
       BACKGROUND: Chat Generative Pre-trained Transformer (ChatGPT) has the potential to enhance healthcare delivery by providing accessible, high-quality medical information. However, no prior studies have evaluated its ability to provide personalised and patient-specific information. Our study aims to evaluate ChatGPT's ability to provide personalised responses to frequently asked questions (FAQs) regarding anterior cruciate ligament (ACL) reconstruction.
    METHODS: We curated a list of 13 clinically relevant FAQs regarding ACL reconstruction and created four distinct patient profiles with varying demographics, comorbidities and activity demands. Each patient profile was entered separately as prompts into the latest version of ChatGPT (GPT-4o). ChatGPT's responses to the FAQs were recorded and assessed independently by two orthopaedic consultants who graded the responses on a scale of 0-2 (0, 1, 2) based on their accuracy and degree that it was personalised to each patient profile.
    RESULTS: ChatGPT performed well in providing accurate and personalised responses, achieving a mean overall accuracy score of 1.85 and mean overall personalisation score of 1.90. It was able to provide personalised and patient-specific responses by customizing its recommendations to cater to the situation of each distinct patient - achieving mean personalisation scores of 1.96, 1.88, 1.92 and 1.81 for the four respective patient profiles.
    CONCLUSIONS: In our study, ChatGPT demonstrated the ability to adapt to distinct patient profiles and provide highly personalised responses regarding ACL reconstruction. Despite potential ethical and regulatory challenges, the ability of AI chatbots to deliver individualized information and counselling is an exciting new frontier in personalised medicine which warrants further research.
    Keywords:  Anterior cruciate ligament; Artificial intelligence; ChatGPT; Chatbot; Frequently asked questions; Personalised
    DOI:  https://doi.org/10.1016/j.knee.2025.06.018
  5. JMIR Med Inform. 2025 Jul 24. 13 e68980
       Background: The integration of artificial intelligence (AI) into clinical workflows holds promise for enhancing outpatient decision-making and patient education. ChatGPT, a large language model developed by OpenAI, has gained attention for its potential to support both clinicians and patients. However, its performance in the outpatient setting of general surgery remains underexplored.
    Objective: This study aimed to evaluate whether ChatGPT-4 can function as a virtual outpatient assistant in the management of puerperal mastitis by assessing the accuracy, clarity, and clinical safety of its responses to frequently asked patient questions in Turkish.
    Methods: Fifteen questions about puerperal mastitis were sourced from public health care websites and online forums. These questions were categorized into general information (n=2), symptoms and diagnosis (n=6), treatment (n=2), and prognosis (n=5). Each question was entered into ChatGPT-4 (September 3, 2024), and a single Turkish-language response was obtained. The responses were evaluated by a panel consisting of 3 board-certified general surgeons and 2 general surgery residents, using five criteria: sufficient length, patient-understandable language, accuracy, adherence to current guidelines, and patient safety. Quantitative metrics included the DISCERN score, Flesch-Kincaid readability score, and inter-rater reliability assessed using the intraclass correlation coefficient (ICC).
    Results: A total of 15 questions were evaluated. ChatGPT's responses were rated as "excellent" overall by the evaluators, with higher scores observed for treatment- and prognosis-related questions. A statistically significant difference was found in DISCERN scores across question types (P=.01), with treatment and prognosis questions receiving higher ratings. In contrast, no significant differences were detected in evaluator-based ratings (sufficient length, understandability, accuracy, guideline compliance, and patient safety), JAMA benchmark scores, or Flesch-Kincaid readability levels (P>.05 for all). Interrater agreement was good across all evaluation parameters (ICC=0.772); however, agreement varied when assessed by individual criteria. Correlation analyses revealed no significant overall associations between subjective ratings and objective quality measures, although a strong positive correlation between literature compliance and patient safety was identified for one question (r=0.968, P<.001).
    Conclusions: ChatGPT demonstrated adequate capability in providing information on puerperal mastitis, particularly for treatment and prognosis. However, evaluator variability and the subjective nature of assessments highlight the need for further optimization of AI tools. Future research should emphasize iterative questioning and dynamic updates to AI knowledge bases to enhance reliability and accessibility.
    Keywords:  AI; Artificial intelligence; ChatGPT; DISCERN score; Flesch-Kincaid readability; general surgery outpatient clinics; patient safety; puerperal mastitis; response quality; treatment adherence
    DOI:  https://doi.org/10.2196/68980
  6. Aesthetic Plast Surg. 2025 Jul 21.
       BACKGROUND: This study aimed to evaluate the performance of four large language models (LLMs)-ChatGPT, Gemini, Copilot, and Claude-in responding to upper eyelid blepharoplasty-related questions, focusing on medical accuracy, clinical relevance, response length, and readability.
    METHODS: A set of queries regarding upper eyelid blepharoplasty, covering six categories (anatomy, surgical procedure, additional intraoperative procedures, postoperative monitoring, follow-up, and postoperative complications) were posed to each LLM. An identical prompt establishing clinical context was provided before each question. Responses were evaluated by three ophthalmologists using a 5-point Likert scale for medical accuracy and a 3-point Likert scale for clinical relevance. The length of the responses was assessed. Readability was also evaluated using the Flesch Reading Ease Score, Flesch-Kincaid Grade Level, Coleman-Liau Index, Gunning Fog Index, and Simple Measure of Gobbledygook grade.
    RESULTS: A total of 30 standardized questions were presented to each LLM. None of the responses from any LLM received a score of 1 regarding medical accuracy for any question. ChatGPT achieved an 80% 'highly accurate' response rate, followed by Claude (60%), Gemini (40%), and Copilot (20%). None of the responses from ChatGPT and Claude received a score of 1 regarding clinical relevance, whereas 10% of Gemini's responses and 26.7% of Copilot's responses received a score of 1. ChatGPT also provided the most clinically 'relevant' responses (86.7%), outperforming the other LLMs. Copilot generated the shortest responses, while ChatGPT generated the longest. Readability analyses revealed that all responses required advanced reading skills at a 'college graduate' level or higher, with Copilot's responses being the most complex.
    CONCLUSION: ChatGPT demonstrated superior performance in both medical accuracy and clinical relevance among evaluated LLMs regarding upper eyelid blepharoplasty, particularly excelling in postoperative monitoring and follow-up categories. While all models generated complex texts requiring advanced literacy, ChatGPT's detailed responses offer valuable guidance for ophthalmologists managing upper eyelid blepharoplasty cases.
    LEVEL OF EVIDENCE V: This journal requires that authors assign a level of evidence to each article. For a full description of these Evidence-Based Medicine ratings, please refer to the Table of Contents or the online Instructions to Authors www.springer.com/00266 .
    Keywords:  Artificial intelligence; Blepharoplasty; Large language models; Oculoplastic surgery
    DOI:  https://doi.org/10.1007/s00266-025-05071-9
  7. Chiropr Man Therap. 2025 Jul 21. 33(1): 28
       BACKGROUND: Rising low back pain (LBP) prevalence and increased patient reliance on online health resources necessitate critically evaluating how chiropractic websites represent common musculoskeletal conditions. This study analyzed LBP-related content shared by South African chiropractic websites to assess the credibility of available information.
    METHODS: A cross-sectional content evaluation was conducted between 2 June 2024 and 21 July 2024 on 333 South African chiropractic websites, identified using a Google search cross-referenced with the Allied Health Professions Council of South Africa registry. Four key areas were evaluated, including (1) contributors to, (2) diagnostic approaches for, (3) possible treatment approaches for and 4) indicators for seeking professional treatment for LBP. Sociodemographic factors such as educational background, sex, and practice ownership structure were considered. A pilot ensured data collection standardization. Content and statistical analysis explored information, citation frequency and identified trends in chiropractic online health communication.
    RESULTS: The analysis revealed significant variability of LBP-related content across chiropractic websites. Of the 333 websites analyzed, most chiropractors were trained at UJ (58.1%) and DUT (29.2%). Sex distribution showed 55.3% were female-owned. The highest concentration of chiropractic websites was in Gauteng, the Western Cape, and KwaZulu-Natal. Regarding the 4 key areas, mechanical contributors to LBP were predominant (65.2%), diagnostic approaches (56.7%), treatment approaches (79.6%) and explicit guidance on when to seek care (47.4%) was mentioned in the websites. Only 13.5% of websites cite sources of the information shared. Chiropractors trained at UJ and DUT emphasized evidence-based treatments, while those trained internationally referenced alternative methods. Sex differences showed males focusing on diagnostic tools and treatment protocols, while females emphasized holistic care and lifestyle modifications.
    CONCLUSION: This study highlights the need for improved LBP information quality and consistency on South African chiropractic websites. The structured subjective content evaluation revealed notable gaps in content depth, specificity, and evidence-based citations. Sociodemographic factors, including educational background and sex, influenced the presentation of information. Future research should prioritize developing standardized and evidence-based guidelines for chiropractic websites to improve access to online health information and patient education and ensure accurate, reliable health information delivery.
    Keywords:  Chiropractic; Consumer health information; Internet; Low back pain; Public health
    DOI:  https://doi.org/10.1186/s12998-025-00591-2
  8. Arthrosc Sports Med Rehabil. 2025 Jun;7(3): 101112
       Purpose: To analyze the content of the most common "People Also Ask" questions regarding anterior cruciate ligament (ACL) injuries and ACL surgery and to assess the scientific validity of the associated websites that provide answers.
    Methods: Search terms were entered into Google regarding ACL tear and ACL surgery. The 300 most asked questions for each query were categorized with the Rothwell classification. The associated webpage information was collected from the "People Also Ask" section, and qualitative analysis was performed with the Journal of the American Medical Association (JAMA) benchmark criteria.
    Results: The Rothwell classification of questions for ACL tear/ACL surgery included fact (53%/54%), policy (37%/30%), and value (16%/10%). Questions related to technical details were significantly more common for the ACL tear group than ACL surgery (P = .015). Academic websites were significantly more common in the ACL tear group (P = .006), and commercial websites were significantly more common in the ACL surgery group (P = .045). The mean JAMA score for all the websites was 1.22 ± 1.1. Websites classified as journal had the highest mean JAMA scores (3.8 ± 0.4), with legal websites and medical practices websites having the lower JAMA scores (0 ± 0, 0.5 ± 0.7).
    Conclusions: The most frequently asked questions on Google pertaining to ACL tear and ACL surgery are related to indications/management, technical details, and timeline of recovery. The quality of information provided by webpages, as measured by JAMA benchmark score, was poor (mean JAMA score 1.22). Academic and medical practice webpages were the most accessed for ACL tear and surgery, respectively.
    Clinical Relevance: It is important to continually analyze the validity and quality of information patients find on the Internet. The information gained from this study may help surgeons understand the type of information patients may find about ACL injuries and ACL surgeries.
    DOI:  https://doi.org/10.1016/j.asmr.2025.101112
  9. PEC Innov. 2025 Jun;6 100367
       Objective: To evaluate the medical information available on the Web in Japanese to patients undergoing splenectomy.
    Methods: Japanese websites on splenectomy were identified by conducting a search on two Internet search engines. Scales were used to score readability, quality, understandability, and actionability. Correlation coefficients for the scale scores were calculated.
    Results: 31 Japanese websites were included in the analysis. "The increased lifetime risk of infection post-splenectomy" and "the requirement to carry a patient card or another form of identification indicating post-splenectomy" was mentioned in 90.3 % (28/31) and 3.2 % (1/31) of websites, respectively. The mean (±standard deviation) grade on the Japanese Readability was 9.8 (±0.9). The actionability of 22 websites was less than 30 %. Positive correlations were found between readability and DISCERN scores (γ = 0.37; 95 % confidence interval [CI], 0.01 to 0.64, p < 0.05) and between content score and actionability (γ = 0.49; 95 %CI, 0.07 to 0.69, p < 0.05).
    Conclusions: Japanese information on websites available to patients undergoing splenectomy was insufficient. The dissemination of information on specific infection prevention measures is needed but must be available at an appropriate readability level.
    Innovation: Creating patient education materials using content score items may promote splenectomy patients coping infection prevention behaviors in Japan.
    Keywords:  Internet; Patient education; Patient education materials; Post-splenectomy
    DOI:  https://doi.org/10.1016/j.pecinn.2024.100367
  10. BMC Oral Health. 2025 Jul 19. 25(1): 1207
       BACKGROUND: With the growing use of Artificial Intelligence (AI) in healthcare, language models like ChatGPT are increasingly being used to provide health information. Given smoking's significant impact on oral health, including periodontal disease, tooth loss, oral cancer, and impaired healing-this study assesses ChatGPT's effectiveness in answering inquiries about the effects of smoking on oral health.
    METHODS: A total of 119 smoking-related dental health questions were generated using online tools and categorized into five domains: periodontal conditions, teeth and health, oral hygiene and breath, oral soft tissues, and oral surgery. ChatGPT 3.5 responses were evaluated for usefulness, readability (FKG index and, SMOG index), quality (Global Quality Scale), reliability (CLEAR tool), and understandability/actionability (PEMAT). Statistical analyses included ANOVA, Kruskal-Wallis, Mann-Whitney U, and Spearman correlation tests using SPSS v20.
    RESULTS: ChatGPT responses were rated as "very useful" (36.1%) or "useful" (42.0%) for most questions. The oral hygiene and breath category received the highest usefulness scores (p =.002). Readability scores indicated that 35.3% of responses were very difficult to read. Surgery-related questions were significantly harder to read (p <.001). The overall quality and reliability of responses was moderate to good. Although understandability was generally high, actionability was limited, with only 23.5% of responses achieving high actionability (> 70%).
    CONCLUSION: ChatGPT provides reliable, high-quality information on smoking's impact on oral health, suitable for general health education. However, it lacks actionable, personalized guidance, limiting its effectiveness in supporting behavioral change and health management.
    Keywords:  Artificial intelligence (Ai); ChatGPT; Dental health; Oral health; Smoking
    DOI:  https://doi.org/10.1186/s12903-025-06377-5
  11. Front Public Health. 2025 ;13 1575801
       Background: YouTube is becoming an increasingly popular platform for health education; however, its reliability for surgical patient education remains largely unexplored. Given the global prevalence of preoperative anxiety, it becomes essential to ensure accurate information online.
    Objectives: The objective is to assess tools/instruments used to evaluate YouTube videos on surgical procedures created to educate patients or health consumers.
    Methods: In June 2023, a comprehensive literature search was conducted on PubMed, PsycINFO, CINAHL, and Scopus. Primary studies with empirical data that evaluate English YouTube videos to educate patients about surgical procedures in all specialties were included. Two reviewers independently completed title/abstract and full text screening, and data extraction in duplicate. The data extracted includes the number of videos evaluated, assessment tools, outcomes of significance, specific objectives, and features examined.
    Results: A total of 41 studies were included in the review. The most commonly used evaluation tools were DISCERN (21 studies), the Global Quality Scale (11 studies), and the JAMA benchmark criteria (11 studies). Notably, 23 studies used a unique assessment instrument, and several studies employed more than one tool concurrently. Of the total studies included, 88% of the articles determined that patients were not adequately educated by YouTube videos per the ratings of the assessment tools, and 19 out of 41 articles mentioned that videos from professional sources were most useful.
    Conclusions: This systematic review suggests that the educational qualities in YouTube videos are substandard. Patients should be cautious when relying solely on YouTube videos for medical guidance. Surgeons and medical institutions are encouraged to direct patients to high-quality patient education sources and create accessible medical content. As there is variability in the quality assessment tools used for evaluation, a standardized approach to creating and assessing online medical videos would improve patient education.
    Keywords:  YouTube videos; health education; patient education; psychometrics; quality assessment; social media; surgical procedure
    DOI:  https://doi.org/10.3389/fpubh.2025.1575801
  12. Dent J (Basel). 2025 Jun 30. pii: 298. [Epub ahead of print]13(7):
      Objectives: The increasing reliance on online platforms for dental education necessitates an assessment of the quality and reliability of available resources. This study aimed to evaluate YouTube videos as educational tools for Class IV resin composite restorations. Methods: The first 100 YouTube videos were screened, and 73 met the inclusion criteria. The videos were evaluated using the Video Information and Quality Index (VIQI) and specific content criteria derived from the dental literature. Videos with a score below the mean were identified as low-content videos. Results: No significant differences were noted between high- and low-content videos when examining the number of views, number of likes, duration, days since upload, viewing rate, interaction index, and number of subscribers (p > 0.05). The high-content videos demonstrated higher mean values compared with the low-content videos in flow (4.11 vs. 3.21; p < 0.0001), accuracy (4.07 vs. 3.07; p < 0.0001), quality (4 vs. 2.66; p < 0.0001), and precision (4.16 vs. 2.86; p < 0.0001). The overall VIQI score was significantly higher (p < 0.0001) for high-content videos (Mean 16.34; SD 2.46) compared with low-content videos (Mean 11.79; SD 2.96). For content score, high-content videos (Mean 9.36; SD 1.33) had a higher score (p < 0.0001) than low-content videos (Mean 4.90; SD 2.04). The key areas lacking sufficient coverage included occlusion, shade selection, and light curing techniques. Conclusions: While a significant portion of YouTube videos provided high-quality educational content, notable deficiencies were identified. This analysis serves as a call to action for both content creators and educational institutions to prioritize the accuracy and completeness of online dental education.
    Keywords:  YouTube; composite; dental; education
    DOI:  https://doi.org/10.3390/dj13070298
  13. J Pediatr Nurs. 2025 Jul 21. pii: S0882-5963(25)00256-8. [Epub ahead of print]85 63-68
       PURPOSE: The aim of this study was to evaluate the content, reliability and quality of YouTube videos on insulin pen injection in children.
    DESIGN AND METHODS: This descriptive study examines 44 videos on YouTube that were searched with the keyword "insulin pen injection in children" and met the selection criteria. The content of the videos was assessed using the Insulin Pen Injection Application Control Form in Children, their reliability was assessed using the modified DISCERN tool and their quality was assessed using the Global Quality Scale (GQS).
    RESULTS: This study has determined that the average GQS assigned by evaluators is 3.81 ± 0.95 and the modified DISCERN score is 3.15 ± 1.43. The study showed that 75 % of the analyzed videos were useful, the GQS and modified DISCERN scores of the useful videos were statistically significantly higher than the useless videos, and there were significant differences in variables such as day, view ratio, duration (p < 0.05). In addition, it was determined that videos published by health institutions had higher GQS and modified DISCERN scores compared to independent publishers.
    CONCLUSION: The evidence presented in this study suggests that videos published by health organizations are a more reliable and high-quality source.
    APPLICATION TO PRACTICE: The reliability and quality of YouTube content on insulin pen injection practices in children should be improved. In this context, it is recommended that nurses responsible for education and patient care take more responsibility in this regard.
    Keywords:  Children; Insulin pen injection; YouTube
    DOI:  https://doi.org/10.1016/j.pedn.2025.07.013
  14. Arthrosc Sports Med Rehabil. 2025 Jun;7(3): 101150
       Purpose: To measure the quality and reliability of information from YouTube videos related to lateral epicondylitis (LE).
    Methods: The first 50 videos obtained from a YouTube search for "lateral epicondylitis" were evaluated. Video reliability was assessed using the Journal of the American Medical Association (JAMA) benchmark criteria (range 0-5). The videos' reliability and quality was evaluated using the Global Quality Score (GQS) (range 0-4), and the Lateral Epicondylitis Specific Score (LESS). Analysis of variance was used to examine differences in reliability and quality across content types and uploaders. Multivariate stepwise regressions were conducted to understand how specific video characteristics influence JAMA benchmark criteria, GQS, and LESS scores.
    Results: The 50 videos had a total viewership of 15,491,927, with an average of 309,838 views per video (range 1,492-2,814,670). Most videos focused on disease-specific information (40%). Nonphysician medical professionals created most of the content (56%), followed by physicians (28%). The mean JAMA benchmark criteria, GQS, and LESS scores were 2.69 (±.59) of 4.0, 2.64 (±.87) of 5.0, and 4.91 (±1.0) of 7.0, respectively. Analysis showed that academic sources had greater LESS scores (P = .035); physicians had higher JAMA benchmark criteria scores than non-physicians (P = .025); nonphysicians had greater GQS than commercials (P = .031); and nonphysicians had greater VPI than physicians (P = .020). Linear regression analysis showed longer duration videos correlated with greater GCS scores (β: 0.656, P < .001) and greater LESS (β: 0.614, P < .001), whereas commercial content types had with lower GCS (standardized beta -0.280, P = .045).
    Conclusions: Although YouTube videos related to lateral epicondylitis have millions of views, the information covered had low to moderate quality and reliability as measured by JAMA benchmark criteria, GQS, and LESS.
    Clinical Relevance: Lateral epicondylitis is a common condition that patients may research online. As the content on YouTube is continually increasing and its popularity remains high, it is important to continually investigate the quality of information available on this platform.
    DOI:  https://doi.org/10.1016/j.asmr.2025.101150
  15. J Orthop Surg Res. 2025 Jul 22. 20(1): 695
       BACKGROUND: Osteonecrosis, or avascular necrosis (AVN), is a degenerative bone disorder caused by insufficient blood supply, primarily affecting weight-bearing joints. With increasing reliance on platforms like YouTube for health information, evaluating the quality of content related to AVN is crucial to prevent misinformation. This study aims to assess the quality and educational value of YouTube videos on AVN.
    METHODS: A systematic search was conducted on November 17, 2024, using the terms "avascular necrosis" and "osteonecrosis." From an initial pool of 143,000 videos, 70 relevant videos were selected based on specific inclusion criteria. Two orthopedic surgeons independently reviewed these videos, assessing various characteristics and categorizing them by content and the uploader's background. The quality was evaluated using the DISCERN score, JAMA criteria, and Video Power Index (VPI). Inter-rater reliability was assessed using Intraclass Correlation Coefficients (ICC), which yielded a value of 0.903 (95% CI: 0.856 to 0.937), indicating nearly perfect agreement among reviewers.
    RESULTS: The mean video length was 11.6 ± 16.5 min, with a total view count averaging 23,087 ± 46,868. The mean DISCERN, JAMA, and GQS scores were 45.7 ± 13.9, 2.6 ± 0.7, and 3.0 ± 0.9, respectively. Videos uploaded by doctors had slightly higher quality scores, but overall, many videos were rated as average or poor. A significant positive correlation was found between the VPI and quality scores (p < 0.05), while older videos tended to have lower quality scores.
    CONCLUSIONS: Health-related videos on platforms like YouTube frequently do not meet high-quality standards, despite some differences in assessments. This emphasizes the need for better quality control and standardization in health content production. Collaboration among medical professionals, content creators, and platforms can significantly enhance online health education quality.
    Keywords:  Avascular necrosis; DISCERN; GQS; JAMA; Medical information reliability; Online health education; Osteonecrosis; YouTube
    DOI:  https://doi.org/10.1186/s13018-025-06116-0
  16. JMIR Infodemiology. 2025 Jul 21. 5 e64891
       Background: Stroke has become a leading cause of death and disability worldwide, resulting in a significant loss of healthy life years and imposing a considerable economic burden on patients, their families, and caregivers. However, despite the growing role of online videos as an emerging source of health information, the credibility and quality of stroke prevention education videos, especially those in Chinese, remain unclear.
    Objective: This study aims to assess the basic characteristics, overall quality, and reliability of Chinese-language online videos related to public health education on stroke prevention.
    Methods: We systematically searched and screened stroke prevention-related video resources from 4 popular Chinese domestic video platforms (Bilibili, Douyin, Haokan, and Xigua). General information, including upload date, duration, views, likes, comments, and shares, was extracted and recorded. Two validated evaluation tools were used: the modified DISCERN questionnaire to assess content reliability and the Global Quality Scale (GQS) to evaluate overall quality. Finally, Spearman correlation analysis was conducted to examine potential associations between general video metrics and their quality and reliability.
    Results: After searching and screening, a total of 313 eligible videos were included for analysis: 68 from Bilibili, 74 from Douyin, 86 from Haokan, and 85 from Xigua. Among these, 113 (36.1%) were created by health care professionals, followed by news agencies (n=95, 30.4%) and general individual users (n=40, 12.8%). The median scores for the modified DISCERN and GQS were 2 and 3, respectively, suggesting that the included stroke prevention-related videos were relatively unreliable and of moderate quality. Most videos focused on primary stroke prevention and commonly recommended adopting a healthy diet; engaging in physical activity; and managing blood pressure, glucose, and lipid levels. Additionally, videos with longer durations and more comments tended to be more reliable and of higher quality. A positive association was also observed between video quality and reliability.
    Conclusions: Overall, the quality and reliability of Chinese-language online videos as a source of stroke prevention information remain unsatisfactory and should be approached with caution by viewers. To address this issue, several measures should be implemented, including establishing an online monitoring and correction system, strengthening the video review process through collaboration with health care professionals, and encouraging more selective and cautious sharing of controversial content. These steps are essential to help curb the spread of online misinformation and minimize its ongoing impact.
    Keywords:  credibility; online video; prevention; public health education; quality; stroke
    DOI:  https://doi.org/10.2196/64891
  17. Arthrosc Sports Med Rehabil. 2025 Jun;7(3): 101101
       Purpose: To assess the validity and informational value of TikTok content related to the anterolateral ligament (ALL) of the knee.
    Methods: The terms "anterolateral ligament of the knee," "ALL reconstruction," "LET," "lateral extra-articular tenodesis," "Lemaire," and "Segond fracture" were used as keywords for an extensive search of video content on the TikTok video platform in August 2024. The video duration and the numbers of likes, shares, saves, and views were recorded for each video. The quality and reliability assessments of the video content were conducted via the DISCERN instrument, Journal of the American Medical Association benchmark criteria, and Global Quality Score.
    Results: A total of 55 videos were included in the analysis, of which 4 (7.3%) were posted by physiotherapists; 7 (12.7%), doctors of osteopathic medicine; 38 (69.1%), patients; and 6 (10.9%), medical doctors. Most of the information concerned patient experience (36%-65.5%). The mean length of the videos was 39.56 ± 49.31 seconds. The mean number of views was 16,444.44 ± 49,509.91, whereas the mean number of likes was 739.05 ± 2,558.24; comments, 9.40 ± 20.06; shares, 10.47 ± 33.59; and saves, 44.02 ± 113.35. The mean DISCERN score, Journal of the American Medical Association score, and Global Quality Score were 15.45 ± 1.56, 0.04 ± 0.19, and 1.07 ± 0.26, respectively, indicating the poorest quality for all 3 indices.
    Conclusions: Most of the videos on TikTok about the ALL of the knee are posted by patients. Videos created by health professionals are significantly more valid and reliable than those created by patients, but all the videos have low educational value.
    Clinical Relevance: In light of the increasing number of patients who use social media for medical information, it is imperative for orthopaedic health care professionals to acknowledge the limitations of TikTok videos concerning the ALL as potential informational resources for their patients.
    DOI:  https://doi.org/10.1016/j.asmr.2025.101101
  18. Int J Dermatol. 2025 Jul 23.
       INTRODUCTION: Vitiligo, an autoimmune disorder causing skin depigmentation, significantly impacts quality of life. With over 1 billion users, TikTok has become a major platform for health information dissemination. However, its engagement-driven algorithm raises concerns about misinformation. This study evaluates the accuracy of the most-liked vitiligo-related TikTok videos and examines the platform's role in health education.
    METHODS: We retrieved TikTok videos using "#Vitiligo" and analyzed the first 100 videos that met inclusion criteria. Videos were categorized into healthcare providers (HP) and non-healthcare providers (NHP). Content characteristics and engagement metrics were extracted. Three independent reviewers assessed quality using multiple assessment tools-Patient Education Materials Assessment Tool (PEMAT), modified DISCERN (mDISCERN), Video Information and Quality Index (VIQI), and Global Quality Score (GQS). A novel tool, V-TRACE, was developed to evaluate (1) clinical aspects (pathophysiology, clinical presentation, treatment, and autoimmune comorbidities) and (2) social aspects (body image and representation of people of color).
    RESULTS: Of 100 videos analyzed, 25% were from HP, demonstrating significantly higher quality across all metrics. HP videos focused on education (96% vs. 20%, p < 0.01) and scored higher on mDISCERN (2.66 vs. 1.66, p < 0.01), GQS (2.66 vs. 1.6, p < 0.01), and VIQI (12.33 vs. 10.00, p < 0.01). NHP produced 75% of videos, received higher engagement (p = 0.04), and had a more positive tone (30.7% vs. 12%, p = 0.05).
    CONCLUSION: Vitiligo-related TikTok content lacks reliability. HPs produce higher-quality content, while NHPs dominate engagement and potentially spread misinformation. There is a need for more evidence-based, engaging content to improve patient education on the platform.
    Keywords:  TikTok; adolescents; health education; social media; vitiligo
    DOI:  https://doi.org/10.1111/ijd.17962
  19. Arthrosc Sports Med Rehabil. 2025 Jun;7(3): 101102
       Purpose: To assess the validity and informational value of the material about the posterior cruciate ligament (PCL) provided on TikTok.
    Methods: The terms "posterior cruciate ligament" and "PCL" were used as keywords for an extensive search of the video content on the TikTok video platform. The video duration and the numbers of likes, shares, saves, and views were recorded for each video. The quality and reliability assessments of the video content were conducted via the DISCERN instrument, Journal of the American Medical Association benchmark criteria, and Global Quality Score.
    Results: A total of 48 videos were included in the analysis, of which 18 (37.5%) were posted by physiotherapists; 14 (29.2%), by private users; and 16 (33.3%), by medical doctors/doctors of osteopathic medicine. Most of the information concerned clinical examination (12-25%) or patient experience (12-25%). Video content was focused on education in 26 videos (54.2%), patient experience in 11 (22.9%), and rehabilitation in 11 (22.9%). Most of the videos used music as the background audio (29-60.4%). The mean length of the videos was 29.29 ± 26.95 seconds. The mean number of views was 565,863.02 ± 2,686,462.45, whereas the mean numbers of likes, comments, shares, and saves were 15,533.58 ± 73,143.51, 127.85 ± 671.49, 661.38 ± 3,049.96, and 1,473.90 ± 7,523.37, respectively. The mean DISCERN score, Journal of the American Medical Association score, and Global Quality Score were 16.15 ± 2.53, 0.15 ± 0.36, and 1.08 ± 0.28, respectively, indicating the poorest quality for all 3 indices.
    Conclusions: Videos about the PCL on the TikTok platform are not comprehensive and have low educational value.
    Clinical Relevance: Given the growing number of patients using social media to seek information about their medical concerns, it is crucial for orthopaedic health care practitioners to recognize the constraints of TikTok videos that focus on the PCL as possible sources of information for their patients.
    DOI:  https://doi.org/10.1016/j.asmr.2025.101102
  20. Aesthetic Plast Surg. 2025 Jul 23.
       BACKGROUND: Breast Implant Illness (BII) is a term used to describe a constellation of nonspecific symptoms attributed to breast implants and is increasingly discussed on social media platforms like TikTok. Understanding online discourse surrounding BII is important to better appreciate patient experiences and recognize potential impacts on patient decision-making.
    OBJECTIVES: This study aims to assess available information on TikTok regarding BII by examining both video content, quality, and engagement while identifying prevailing trends in the discourse.
    METHODS: TikTok was queried using BII-related hashtags. Relevant videos with over 10,000 views were included. Video engagement metrics, creator type, content, and purpose were analyzed. DISCERN and Global Quality Scale (GQS) scores were utilized to assess the quality of health information.
    RESULTS: The analyzed videos accumulated over 116 million views, with patients being the predominant content creators (73.9%). Engagement metrics did not differ significantly between creator types or board certification status, but verified accounts received significantly higher engagement. Personal experience videos were more common (45.7%) than entertainment (28.3%) and education (26.1%) videos, yet entertainment-focused videos received significantly more views and likes (p < 0.05). Overall, video quality was low, with plastic surgeons producing significantly higher DISCERN and GQS scores than patients (p < 0.001).
    CONCLUSION: Social media serves as an important space for BII discourse, but the predominance of anecdotal content over evidence-based knowledge raises concerns about misinformation. Plastic surgeons' videos were higher quality but did not achieve greater engagement, highlighting the need for greater provider participation to strengthen accurate and appropriate patient information.
    LEVEL OF EVIDENCE V: This journal requires that authors assign a level of evidence to each article. For a full description of these Evidence-Based Medicine ratings, please refer to the Table of Contents or the online Instructions to Authors www.springer.com/00266.
    Keywords:  Autoimmune; Breast implant illness; Explantation; Implant; Social media; TikTok
    DOI:  https://doi.org/10.1007/s00266-025-05074-6
  21. Ann Surg Oncol. 2025 Jul 19.
       BACKGROUND: Patients are increasingly turning to online social media networks for healthcare information. This study evaluated content related to oncoplastic breast surgery found on TikTok.
    METHODS: The top 50 videos were retrieved from subsequent TikTok searches for 'oncoplastic breast surgery', 'oncoplastics', and 'oncoplastic surgery'. These videos were compiled and the 50 with the most views were selected. Video characteristics were abstracted, and the videos were analyzed for thematic content by three surgeons and graded for educational value by eight individuals of varying backgrounds.
    RESULTS: We focused our analysis on the two most common creator types: surgeons and patients. Both groups showed comparable audience engagement patterns, with no significant difference in the number of views, likes, or comments. However, subscribers, saves, and shares differed significantly, with surgeons showing higher numbers (p < 0.05). Analyzing thematic content by creator type highlights different priorities between the groups. Surgeon videos received significantly higher ratings on questions about providing accurate medical information (p < 0.01) and explaining surgery outcomes clearly (p < 0.05). Surgeons and trainees consistently noted these differences, with 70% of their ratings showing significance in favor of surgeon content. In contrast, non-healthcare raters showed fewer significant distinctions, with only 40% of their ratings identifying differences between creator groups.
    CONCLUSION: While patient creators focus heavily on personal experiences and survivorship, surgeon-created content emphasizes medical accuracy and detailed surgical information. Surgeons can use these findings to better understand the types of information their patients are accessing and to tailor their own educational efforts accordingly.
    Keywords:  Breast cancer; Content; Health care information; Online medical information; Patient education; Social media
    DOI:  https://doi.org/10.1245/s10434-025-17858-w
  22. Front Public Health. 2025 ;13 1592093
       Introduction: As an important social group, mothers possess unique physiological and psychological characteristics. They may rely on Internet sources for obtaining health information about themselves and their children.
    Methods: Based on the Elaboration Likelihood Model (ELM) and Technology Acceptance Model (TAM), the influencing factors model of maternal online health information searching behavior is constructed. We collect 903 valid sample data through the questionnaire survey in China. Using SPSS software to empirically analyze the influence of different influencing factors on maternal online health information searching behavior.
    Results: Perceived usefulness mediated between the Internet information quality and online health information-seeking behavior, accounting for 59.46% of the variance. Anxiety mediated between maternal stress and online health information-seeking behavior, accounting for 19.40% of the variance.
    Conclusion: Internet information quality positively affected mothers' perception of the usefulness of the information. Perceived usefulness and anxiety played a partial mediating role in the central and peripheral paths, respectively. Furthermore, social support effectively moderated the influence of maternal stress on their anxiety.
    Keywords:  anxious emotion; maternal; online health information-seeking behavior; perceived usefulness; social support
    DOI:  https://doi.org/10.3389/fpubh.2025.1592093
  23. Arch Physiother. 2025 Jan-Dec;15:15 195-205
       Introduction: Spine-related leg pain (SRLP), also commonly referred to as sciatica, is characterized by many unanswered questions. Nowadays, patients search for information online to find answers not provided by clinicians. Therefore, this work aimed to (i) collect the main questions and keywords searched by the Italian general population related to SRLP, (ii) analyze the content of the most clicked web pages, (iii) investigate its alignment with evidence-based recommendations, and (iv) assess their credibility and (v) readability.
    Methods: SEMrush was used to collect questions, keywords, search volumes, and web pages. Credibility was assessed through the QUality Evaluation Scoring Tool (QUEST). Readability was evaluated by the Gulpease index. Thematic and content analyses were performed.
    Results: Monthly search volumes ranged from 4,400 to 33,100. QUEST scores ranged from 1 to 14 out of 28. Readability scores ranged from 20 to 47 out of 100. Common themes highlighted the roles of medications and physiotherapy, the value of doing effortless activities, and the importance of sleep. Content analysis showed that 11 codes (41%) were aligned with, 7 (26%) were contradicted by, and 26 (96%) were entirely omitted from at least one web page.
    Conclusion: The findings suggest that inaccurate, poorly credible, and difficult-to-read information on SRLP are common online, leading to plausible targets for the educational intervention of clinicians dealing with SRLP patients. Clinicians are now more aware of the questions that their patients could be asking to Google and the answers it provides about SRLP.
    Keywords:  Consumer Health; Disinformation; Health Education; Health Literacy; Public Health; Sciatica
    DOI:  https://doi.org/10.33393/aop.2025.3360
  24. Geriatrics (Basel). 2025 Jul 01. pii: 87. [Epub ahead of print]10(4):
      Background/Objectives: Visual impairment (VI) represents a significant health challenge among older adults, particularly due to their limited understanding of health information. This study aimed to investigate and compare the health information-seeking behavior (HISB) of older adults with VI across different stages of eye care. Methods: A cross-sectional comparative study was conducted in 248 older adults with VI in various stages of eye care, including the pre-visit stage (n = 84), treatment stage (n = 83), and follow-up stage (n = 81) at the Shenzhen Eye Hospital from July to October 2024. Participants completed an HISB questionnaire encompassing four dimensions: attitude, needs, sources, and barriers. Results: The overall mean score of HISB varied significantly among the different eye care stages. The treatment stage had the highest scores (3.70 ± 0.32), particularly in attitudes toward health information and information needs. Poor self-reported overall health facilitated HISB in each stage. In the pre-visit stage, higher income was associated with increased HISB, while a lack of internet access reduced it. In the treatment stage, higher education was associated with higher HISB, whereas moderate-to-severe VI and certain income levels were associated with negative effects. In the follow-up stage, rural residence and spousal or child caregiving emerged as key facilitators of HISB. Conclusions: The treatment stage is a critical period for HISB engagement in older adults with VI. Healthcare providers should consider stage-specific factors of HISB to optimize health information delivery.
    Keywords:  eye care; health behavior; health information; older adults; vision impairment
    DOI:  https://doi.org/10.3390/geriatrics10040087