AI Can’t Help Real People - But Pass Med Exam w 95%

AI Can’t Help Real People - But Pass Med Exam w 95%

1.943 Lượt nghe
AI Can’t Help Real People - But Pass Med Exam w 95%
We need to code new AI systems w/ improved Human-AI interaction, especially on medical topics and health care. We need to create new training datasets for pre-training new VLMs for clinical settings and improve medical reasoning abilities (w/ dedicated data training sets) for incomplete human information transfer in medical domains. New question to solve for insurance companies: What company is responsible regarding AI liabilities and high insurance costs: the hospital or Anthropic (w/ their "Ai model welfare" when AI models become self-conscious)? Are Ai models liable themself (nice idea, Anthropic - smile) or their creators? Your thoughts? LLMs ace benchmark tests but flunk real-world patient interactions. All rights w/ authors: "Clinical knowledge in LLMs does not translate to human interactions" by Andrew M. Bean1, Rebecca Payne2,3, Guy Parsons1, Hannah Rose Kirk1, Juan Ciro4, Rafael Mosquera5,6, Sara Hincapi´e Monsalve5,6, Aruna S. Ekanayaka7, Lionel Tarassenko8, Luc Rocher1, Adam Mahdi1 from 1 Oxford Internet Institute, University of Oxford, Oxford, UK. 2 Nuffield Department of Primary Care,University of Oxford, Oxford, UK. 3 North Wales Medical School, Bangor University, Bangor, UK. 4 Contextual AI, Mountain View, USA. 5 MLCommons, San Francisco, USA. 6 Factored AI, Palo Alto, USA. 7 Birmingham Women’s and Children’s NHS Foundation Trust, Birmingham, UK. 8 Institute of Biomedical Engineering, University of Oxford, Oxford, UK. #healthcareinnovation #healthcareprofessionals #airesearch #clinic #medicalcoding #logicalthinking