Perspectives
Counterpoint
Artificial Intelligence and Regulatory Realities in Drug Development: A Pragmatic View
Isaac R. Rodriguez-Chavez
4Biosolutions Consulting
A

rtificial intelligence’s (AI) potential to accelerate drug development by leveraging data sets for better trial outcomes and faster therapy delivery is clear. However, realizing this vision faces complexities in the regulatory, intellectual property (IP), and patient privacy landscapes. Balancing innovation with data confidentiality, IP rights, and patient autonomy is key, alongside addressing algorithmic bias with transparency and accountability measures such as explainable AI (XAI).

The Data Surge and the Regulatory Framework

The pharmaceutical industry’s clinical trials generate vast data sets submitted to agencies such as the US Food and Drug Administration (FDA). While this data offers significant innovation potential, accessing it is more than a technological issue. Legal and FDA regulatory frameworks protect confidential information, IP, and patient privacy (e.g., 21 CFR Part 20, 21 CFR Part 21, 21 CFR 20.88, DSCSA, FISMA, and FDA Confidentiality Commitment Agreements). Confidentiality obligations under the Federal Food, Drug, and Cosmetic Act and the Trade Secrets Act, which prevent the FDA from disclosing sponsors’ commercial secrets, challenge open access to FDA data. Furthermore, the Freedom of Information Act (FOIA) includes exemptions protecting trade secrets contained within government-stored information. Releasing sponsor data broadly, even if anonymized, could lead to legal challenges and erode the trust between regulatory agencies and the pharmaceutical industry, potentially hindering future collaborations and data sharing initiatives.

Navigating Intellectual Property and Data Anonymization

The pharmaceutical industry invests heavily in drug development, generating valuable data that are considered IP. Clinical trial data, including protocols, reports, and patient-level information, is vital for companies seeking a competitive edge and return on investment.

While AI promises pattern discernment from data sets, legal and ethical obligations regarding patient data require careful attention. Compliance with the Health Insurance Portability and Accountability Act (HIPAA) is essential, yet reidentification risks remain, especially in rare disease studies due to small population sizes, unique population characteristics, extensive data sharing, and patient participation in online support forums. The use of artificial intelligence/machine learning (AI/ML) adds complexity; deidentifying data to meet HIPAA requirements while retaining AI/ML utility is challenging. International data transfer implications, especially compliance with the General Data Protection Regulation (GDPR) in the European Union, which mandates data anonymization and individual rights, present additional challenges. Techniques such as differential privacy and federated learning can aid GDPR compliance by minimizing reidentification risks and enabling analysis without direct data access. Addressing AI/ML prediction risks such as error propagation and algorithm opacity is crucial. It’s essential to address potential risks associated with relying on AI/ML predictions, including the propagation of errors, overfitting, and the “black box” nature of certain algorithms, which may obscure the reasoning behind AI-generated predictions.

Path of Least Regulatory Resistance: Sponsor-Led Initiatives

Given the regulatory hurdles associated with direct data sharing by the FDA, a more pragmatic approach involves empowering sponsor-led data-sharing initiatives. Companies are legally permitted to share their own deidentified clinical trial data, subject to appropriate privacy safeguards and ethical considerations. This approach aligns with existing legal frameworks and respects the IP rights of pharmaceutical companies. These initiatives can be built upon existing collaborative frameworks within the pharmaceutical industry, fostering a culture of data sharing while mitigating legal and regulatory risks. Creating a federated learning system, in which data are stored with sponsors but can be analyzed collectively, may be a middle ground that addresses legal and privacy concerns while enabling innovative possibilities. To ensure equitable outcomes, it is crucial to address potential biases in clinical trial data used to train AI/ML models and promote diversity in data sets to avoid exacerbating health disparities across different demographic groups.

Specific examples of these initiatives include:

  • TransCelerate BioPharma’s Data Sharing Collaboration: This initiative facilitates the sharing of clinical trial data among member companies to accelerate drug development and improve patient outcomes. They have multiple initiatives focused on developing data-sharing agreements and procedures.
  • Vivli, Center for Global Clinical Research Data Sharing: This independent, global data-sharing platform allows researchers to request access to clinical trial data from various sponsors. Vivli promotes responsible data sharing to advance scientific discovery and improve human health.

However, there is still a need to implement independent oversight mechanisms for data-sharing initiatives to mitigate potential conflicts of interest and ensure transparency and accountability in data access and utilization.

The Nascent Frontier of Artificial Intelligence Regulation and Validation

The application of AI in drug development is still a relatively new field, and regulatory guidelines are evolving. The FDA is actively working to develop frameworks for evaluating the safety and effectiveness of AI/ML-based medical products. Using AI to predict clinical trial outcomes will require a thorough validation process to ensure that the AI models are reliable, accurate, and do not introduce unintended biases.

The FDA has published guidance on its framework for regulatory considerations of AI/ML-based devices (see below) and continues to explore the implications of AI for drug development. Any AI-driven tools that analyze clinical trial data would need to be rigorously validated and demonstrate compliance with regulatory standards. Interdisciplinary collaboration among AI experts, clinicians, ethicists, and regulatory specialists is crucial for responsible AI implementation, ensuring comprehensive oversight and alignment with healthcare objectives. AI has the potential to significantly enhance clinical trial design and patient recruitment by optimizing inclusion criteria, predicting patient outcomes, and streamlining trial processes, while also addressing challenges such as data quality, algorithmic bias, and regulatory compliance.

Notable FDA resources include:

Consent: The Cornerstone of Ethical Data Use

In our enthusiasm for AI innovation, we must not overlook the fundamental principle of patient consent. Patients who participate in clinical trials provide consent for their data to be used for specific research purposes outlined in the informed consent document. Broadening the use of this data for general public access or AI processing without explicit consent raises serious ethical and legal concerns.

Retroactively seeking consent from patients whose data is already stored in large databases is a logistical and ethical challenge. It’s important that informed consent processes clearly articulate how patient data may be used in future research, including AI-driven analysis, while providing patients with the option to opt out of such uses. This is even more critical when considering clinical trials conducted worldwide, where diverse data protection laws may apply (e.g., GDPR). Trials are increasingly global, and adherence to local regulations is paramount. To address the challenge of obtaining prospective consent for AI/ML applications that were unforeseen at the time of initial data collection, explore the implementation of dynamic consent platforms. These platforms enable granular control over data usage, allowing participants to specify preferences for secondary research, data sharing, and AI-driven analysis, while also providing mechanisms for periodic review and modification of consent choices.

Call for Collaborative Realism

The path to realizing the vision of unlocking FDA’s clinical trial data sets with AI requires a pragmatic and legally sound approach. Current regulatory frameworks, IP rights, and ethical considerations surrounding patient data necessitate a more nuanced strategy than simply opening up deidentified FDA databases for public consumption. The path forward likely lies in sponsor-led data-sharing initiatives, guided by clear regulatory frameworks and ethical principles. By fostering a collaborative environment where industry, regulatory agencies, and researchers work together, we can unlock the potential of AI to accelerate drug development while upholding the highest standards of data privacy, security, and ethical conduct. Such collaborations need to be built with the understanding that FDA’s core mission is to regulate and approve safe and effective drugs; any redirection of this role would require substantial legislative changes and a fundamental shift in the relationship between the FDA, pharmaceutical companies, and the public.

References available upon request.