Introduction: Artificial Intelligence (AI) has transformative potential in medical sciences, promising increased accuracy in diagnosis, disease progression prediction, and personalized treatments. However, the growing emphasis on technical aspects has raised significant human and ethical challenges related to doctor-patient communication, privacy, accountability, and trust. This study aims to examine these two contrasting aspects—opportunities and challenges—while emphasizing the necessity of integrating artificial intelligence with human values in medicine.
Method: This research was conducted as a narrative review by searching scientific databases, including PubMed, Scopus, and Web of Science, using relevant keywords. Articles published between 2014 and 2024 that examined the human, communicative, and ethical aspects of AI applications in medicine were collected and analyzed through qualitative content analysis.
Results: The findings revealed that while AI automates repetitive tasks and analyzes vast amounts of data, allowing physicians to focus more on patient care, significant challenges were identified in several areas: deficiencies in establishing empathetic and nonverbal communication, concerns regarding data privacy and security, ambiguity in legal and ethical accountability in the event of errors, and the risk of algorithmic bias, which could lead to inequalities in healthcare delivery.
Conclusion: The successful implementation of AI in medicine requires moving beyond a purely technical perspective and adopting a human-centered approach. AI should function as an assistive tool for physicians rather than replace their human judgment and communication. To ensure the future of this collaboration, it is essential to develop transparent ethical frameworks, train healthcare professionals, and design efficient systems that uphold human values.
Type of Study:
Short Communication |
Subject:
Artificial Intelligence in Healthcare Received: 2025/04/25 | Accepted: 2025/06/2