Article Text
Abstract
Components of artificial intelligence (AI) for analysing social big data, such as natural language processing (NLP) algorithms, have improved the timeliness and robustness of health data. NLP techniques have been implemented to analyse large volumes of text from social media platforms to gain insights on disease symptoms, understand barriers to care and predict disease outbreaks. However, AI-based decisions may contain biases that could misrepresent populations, skew results or lead to errors. Bias, within the scope of this paper, is described as the difference between the predictive values and true values within the modelling of an algorithm. Bias within algorithms may lead to inaccurate healthcare outcomes and exacerbate health disparities when results derived from these biased algorithms are applied to health interventions. Researchers who implement these algorithms must consider when and how bias may arise. This paper explores algorithmic biases as a result of data collection, labelling and modelling of NLP algorithms. Researchers have a role in ensuring that efforts towards combating bias are enforced, especially when drawing health conclusions derived from social media posts that are linguistically diverse. Through the implementation of open collaboration, auditing processes and the development of guidelines, researchers may be able to reduce bias and improve NLP algorithms that improve health surveillance.
- ethics- medical
- ethics- research
- ethics
- decision making
- information technology
Data availability statement
Data sharing not applicable as no datasets generated and/or analysed for this study. No data are available.
Statistics from Altmetric.com
Data availability statement
Data sharing not applicable as no datasets generated and/or analysed for this study. No data are available.
Footnotes
Contributors SDY and LF contributed to the design and the formulation of the main arguments. LF, SK and SDY were responsible for the drafting and editing of the paper. LF, SK and SDY contributed to the final editing and approved the final version of the manuscript. LF is the guarantor and takes full responsibilty for all aspects of the manuscript.
Funding This study was funded by National Institutes of Allergy and Infectious Diseases (NIAID, grant number: 5R01AI132030-05), National Center for Complementary and Integrative Health (NCCIH), National Institute on Minority Health and Health Disparaties (NIMHD), and National Institute on Drug Abuse (NIDA).
Competing interests None declared.
Provenance and peer review Not commissioned; externally peer reviewed.
Read the full text or download the PDF:
Other content recommended for you
- Expert artificial intelligence-based natural language processing characterises childhood asthma
- Applying natural language processing and machine learning techniques to patient experience feedback: a systematic review
- Applications and challenges of AI-based algorithms in the COVID-19 pandemic
- Artificial intelligence in orthopaedics surgery: transforming technological innovation in patient care and surgical training
- Artificial intelligence (AI) and global health: how can AI contribute to health in resource-poor settings?
- Artificial intelligence-based mining of electronic health record data to accelerate the digital transformation of the national cardiovascular ecosystem: design protocol of the CardioMining study
- Artificial intelligence and inflammatory bowel disease: practicalities and future prospects
- HbA1c recording in patients following a first diagnosis of serious mental illness: the South London and Maudsley Biomedical Research Centre case register
- Potential applications and implications of large language models in primary care
- Future of machine learning in paediatrics