Calling all #medlib friends!

I built a tool that makes it easier to build proximity searches in PubMed, among other things!

You can check it out at https://pairwise-pubmed.streamlit.app/

You can see a 15 minute presentation I gave about it, including a demo, at the PNC-MLA conference, at https://www.youtube.com/watch?v=cWFvpcVXhbw&list=PLkdeJJs0vOyFADjyFALAj-OKd7uFh3RMg&index=5&t=2657s

"Caught in the Web of Words:
Do LLMs Fall for Spin in Medical Literature?"

We evaluated 22 LLMs and found that they are across the board more susceptible to "spin" than humans. They might also propagate spin into their outputs: We find evidence, e.g., that
LLMs implicitly incorporate spin into plain language summaries that they generate.

https://arxiv.org/abs/2502.07963

#research #medicine #LLM #AItools
#science #medlib

Caught in the Web of Words: Do LLMs Fall for Spin in Medical Literature?

Medical research faces well-documented challenges in translating novel treatments into clinical practice. Publishing incentives encourage researchers to present "positive" findings, even when empirical results are equivocal. Consequently, it is well-documented that authors often spin study results, especially in article abstracts. Such spin can influence clinician interpretation of evidence and may affect patient care decisions. In this study, we ask whether the interpretation of trial results offered by Large Language Models (LLMs) is similarly affected by spin. This is important since LLMs are increasingly being used to trawl through and synthesize published medical evidence. We evaluated 22 LLMs and found that they are across the board more susceptible to spin than humans. They might also propagate spin into their outputs: We find evidence, e.g., that LLMs implicitly incorporate spin into plain language summaries that they generate. We also find, however, that LLMs are generally capable of recognizing spin, and can be prompted in a way to mitigate spin's impact on LLM outputs.

arXiv.org

#MedLib #JobAd #AcademicLib

We're looking for a new Medicine/Nursing Liaison to join at our Peoria campus. You'll be part of a good local team and a solid team of nursing/med liaisons across our campuses.

Happy to answer questions about the institution, I'm not on the search committee.

https://uic.csod.com/ux/ats/careersite/1/home/requisition/14975?c=uic&sq=1034019

Clinical Assistant Professor & Regional Health Sciences Librarian

Position SummaryAs a member of the University Library, working at LHS Peoria, the successful candidate will provide general health sciences reference ...

Bluesky

Bluesky Social

Trip database search application can
now generate mindmaps for you. The branches in the mindmap are clickable weblinks that will take you to articles from the Trip database related to that particular node in the mindmap.

https://www.tripdatabase.com

#research #medicine #mindmap
#medlib

Trip Medical Database

Evidence-based answers for health professionals | Searching sources such as systematic reviews, clinical guidelines and RCTs

"Artificial intelligence versus human researcher performance for systematic literature searches: a study focusing on the surgical management of base of thumb arthritis"

Interesting very recent paper comparing a manual systematic literature search to searches done with Elicit, Consensus, and ChatGPT tools.

https://www.oaepublish.com/articles/2347-9264.2024.99

#research #medicine #Elicit #Consensus #medlib

Artificial intelligence versus human researcher performance for systematic literature searches: a study focusing on the surgical management of base of thumb arthritis

Aim: In the digital age, artificial intelligence (AI) platforms have gradually replaced traditional manual techniques for information retrieval. However, their effectiveness in conducting academic literature searches remains unclear, necessitating a comparative assessment. This study examined the efficacy of AI search engines (Elicit, Consensus, ChatGPT) vs. manual search for literature retrieval, focusing on the surgical management of trapeziometacarpal osteoarthritis.Methods: The study was executed per the Cochrane Handbook for Systematic Reviews and PRISMA guidelines. AI platforms were given relevant keywords and prompts, while manual searches used PubMed, Cochrane CENTRAL, Web of Science, and Scopus databases from January 1901 to April 2024. The study focused on English-language randomized controlled trials (RCTs) comparing surgical management of trapeziometacarpal osteoarthritis (TMCJ OA). Two independent evaluators screened and extracted data from the studies. Primary outcomes involved the quality and relevancy of studies chosen by both search methods, evaluated by false positive rates and number of studies, including outcomes of interest.Results: The manual search yielded the most results (6,018), followed by Elicit (4,980), Consensus (3,436), and ChatGPT (6). Elicit identified the highest number of RCTs (205) but also had the greatest false positive rate (94%). Ultimately, the manual search identified 23 suitable studies, Elicit found 10, Consensus found 9, and ChatGPT identified only 1. No additional studies were found by AI search engines that were not discovered in the manual search.Conclusion: The findings highlight the potential advantages and drawbacks of AI search engines for literature searches. While Elicit was prone to error, Consensus and ChatGPT were less comprehensive. Significant enhancements in the precision and thoroughness of AI search engines are required before they can be effectively utilized in academia.

Plastic and Aesthetic Research
NOT-OD-25-047: 2024 NIH Public Access Policy

NIH Funding Opportunities and Notices in the NIH Guide for Grants and Contracts: 2024 NIH Public Access Policy NOT-OD-25-047. OD

If there's a #medlib out there that needs a research project, I need someone to do an analysis of the AAHSL stats over the past 15 years. Not the stats themselves -- specifically the instrument. The questions or some aspect of the definitions has changed EVERY SINGLE YEAR since I've become aware of them (so 14+ years so far) and the effect this has on any kind of longitudinal comparison or benchmarking, even internally has to be impacted.

Dana-Farber Cancer Institute finds main GPT-4 concerns include falsehoods, high costs https://www.zdnet.com/article/dana-farber-cancer-institute-finds-main-gpt-4-concerns-include-falsehoods-high-costs/

Interesting, and a new-to-me journal. A preprint of the paper is at https://osf.io/preprints/osf/bqv4f

#AI #ChatGPT #MedLib

Dana-Farber Cancer Institute finds main GPT-4 concerns include falsehoods, high costs

The renowned medical institute trialed letting 12,500 staff members use a private instance of GPT-4, a first-of-its-kind implementation in medicine.

ZDNET

"PubTator 3.0: an AI-powered Literature Resource for Unlocking Biomedical Knowledge"

PubTator 3.0 ( https://www.ncbi.nlm.nih.gov/research/pubtator3/ ) is a biomedical literature
resource using state-of-the-art AI techniques to offer semantic and relation searches for key
concepts like proteins, genetic variants, diseases, and chemicals. It currently provides over one billion entity and relation annotations across approximately 36 million PubMed abstracts ...

https://browse.arxiv.org/abs/2401.11048

#search #medlib #COVID

PubTator3 - NCBI - NLM - NIH