Author(s): Sai Shivani Parimi, Bishnu Sarker, Alexis Lamar, Hari Chandana Vadde
Recent advances in artificial intelligence have led to the emergence of multi-agent
health systems capable of integrating diverse data sources and supporting complex
decision-making in healthcare.These systems demonstrate the potential to analyze multimodal
health data and generate personalized insights but their application has largely focused
on individual-level care and often overlooks broader population health challenges,particularly
those related to health equity.In conditions such as Type 2 Diabetes (T2D), traditional
risk stratification models frequently rely on clinical indicators alone,failing to
account for the significant impact of social determinants of health (SDOH) on patient
outcomes.This study proposes a conceptual multi-agent AI framework designed to enhance
population-level diabetes risk stratification by integrating SDOH into electronic
health record (EHR)-based models.The framework extends existing multi-agent architectures
by introducing specialized and coordinated agents with distinct capabilities.A population
orchestrator agent manages task allocation and system coordination, a data science
agent enables SDOH-integrated predictive modeling and subgroup analysis;a domain expert
agent ensures clinically grounded,evidence-based interpretation,and a health coach
agent translates insights into culturally sensitive, actionable interventions.In addition,the
framework introduces supporting agents,including health equity agent for bias monitoring,
a data quality agent for ensuring data completeness, and an SDOH extraction agent
for structuring unstructured social data.The proposed framework emphasizes equitable
identification of at-risk populations, particularly individuals with moderate clinical
risk but high social vulnerability.This work is conceptual and has not yet been implemented
or empirically validated.Future research is required to evaluate model performance,
system integration feasibility, and real-world impact.Despite these limitations, the
framework provides a foundation for advancing equitable,AI-driven population health
interventions.
Author(s): Laxmi Shravani Mamidala, Ashwitha Reddy Venna, Gayathri Orsu, Lakshmi
Triveni Muthyala, Tharun Swaminathan Ravi Kumar
Blood cancer, especially leukemia, is one of the most dangerous diseases, and early,
accurate diagnosis is critical for improving patient outcomes. Conventional microscopic
analysis of blood smear images is time-consuming, highly subjective, and prone to
variability. This research proposes an automated Blood Cancer Detection System using
a combination of machine learning, deep learning, and computer vision techniques to
identify and classify abnormal blood cells.
The proposed system uses a hybrid artificial intelligence approach. Deep features are extracted from microscopic images using a pre-trained ResNet50 convolutional neural network, and these features are classified using machine learning algorithms such as Logistic Regression, Support Vector Machine (SVM), Random Forest, XGBoost, and K-Means clustering. To improve classification, YOLOv8 is used for accurate object detection and localization of blood cells. Model interpretability is enhanced using Grad-CAM explainability techniques, which highlight the most important regions of the image for the classification decision.
The dataset contains 5000 labeled microscopic images of blood cells, representing the five most prominent cell types essential for leukemia diagnosis. Experimental results show excellent performance of the supervised models, with SVM and Logistic Regression achieving about 98–99% accuracy, while YOLOv8 demonstrates strong accuracy in blood cell detection.
As an additional contribution, this research proposes an intelligent clinical decision-support system that allows users to upload images, display predictions, analyze explainable heatmaps, and compare predictions across models. Future work may include integration with real-time clinical databases, the use of federated learning for privacy-preserving collaboration, and multimodal patient data to further improve prediction accuracy. Overall, this system shows the potential of combining deep learning, machine learning, and explainable AI to accelerate biomedical image analysis and support faster, more accurate blood cancer diagnosis.
Google research has recently conducted a study (Heydari et al., 2025) that underscores the future of personalized healthcare with the help of multi-agent architectures. Multi-agent systems assign tasks to specialized agents, including data science, clinical, and health coaching, to more accurately domain-knowledge decision-making, compared to single AI agents, which are usually constrained with large, complex datasets and do not have coordinating capabilities with other agents in the system. This solution helps to integrate different sources of data, such as electronic health records and wearables, to provide more individualized and proactive care. Building on this concept, we developed a blueprint for a mental health solution that extends the architecture by introducing additional agents, including a health equity agent and a resource allocation agent, aimed at addressing disparities in care and optimizing the distribution of mental health resources. Overall, this multi-agent framework offers a more scalable, equitable, and patient-centered alternative to traditional single-agent models.
The personal health agent is headed by an Orchestrator that coordinates specialized agents: Data Science Agent predicts based on the datasets: CDC records, EHRs, Social Vulnerability Index (SVI). Domain Expert Agent uses DSM-5 knowledge and evidence-based practice to provide insight. Health Coach Agent delivers personalized interventions. The Health Equity and Resource Allocation Agents collaborate to define disparities and populations at risk with the help of socioeconomic and access-related data. The resource allocation agent represents the most efficient option for distributing healthcare resources based on the needs and socioeconomic status of the patient and local services. This system integrates the input of user-reported symptoms, wearable data, hospital data, and socioeconomic factors to generate personalized recommendations.
The model aims to offer individual and patient-centered solutions that meet personal needs and contribute to effective mental health management.
The rapid growth of scientific publications has placed significant pressure on traditional peer-review systems, making it difficult to maintain consistent, timely, and high-quality evaluations. Human reviewers often face challenges such as limited time, reviewer fatigue, and subjective biases, which can lead to variability in review quality. To address these issues, this paper introduces ProComp, a multi-agent large language model (LLM) framework designed to automate structured scholarly paper evaluation by emulating the roles of expert peer reviewers.
ProComp decomposes the review process into multiple specialized evaluation agents, each responsible for assessing a distinct aspect of a research paper. The framework employs three expert reviewer agents that independently analyze problem formulation and motivation, methodological soundness and technical rigor, and novelty and research contribution. Each agent produces structured assessments and quantitative scores aligned with common academic review criteria. To enhance contextual understanding and support informed evaluation, the system incorporates a Retrieval-Augmented Generation (RAG) component that retrieves related work and citation context from external scholarly databases.
A fourth agent acts as a comparative meta-reviewer, aggregating the outputs of the expert agents to generate a consolidated evaluation and final recommendation. The proposed framework is evaluated on a dataset of 194 research papers, comparing multiple LLM backends and analyzing their agreement with human review signals. Results demonstrate that the multi-agent architecture produces structured, interpretable evaluations and maintains meaningful correlation with human judgments. Ablation analysis further investigates the influence of individual reviewer agents on final scoring decisions.
Overall, ProComp demonstrates the potential of collaborative LLM agents to support scalable, transparent, and consistent automated scholarly evaluation, providing a promising direction for augmenting future academic peer-review systems.
Author(s): Abhinay Peeta, Venkata Haritha Sai Sri Animilli & Dr Sayed K. Shah
Biometric authentication systems based on conventional modalities such as fingerprints
or facial recognition are increasingly vulnerable to spoofing, environmental variability,
and privacy concerns. In response, this work introduces a novel tongue-based biometric
identification framework designed for enhanced security and robustness within biomedical
and AI-driven authentication contexts. Leveraging the intrinsic morphological stability
of the tongue, its protected intra-oral environment, and its inherent difficulty to
forge or replicate, we propose this modality as a next-generation biometric trait
for personal identification.
At the heart of our system lies a deep learning architecture: a convolutional neural network (CNN) that automatically extracts high-dimensional, discriminative features capturing fissure patterns, surface texture, and geometric tongue shape. To further boost performance and anti-spoofing resilience, we integrate a feature-level fusion strategy that combines these learned embeddings with complementary hand-crafted descriptors thereby forming a unified, highly discriminative feature vector. Experimental validation demonstrates that this hybrid approach significantly improves identification accuracy while reinforcing the modality’s robustness against adversarial or spoofing attempts. The results affirm the viability of the tongue as a reliable biometric modality and highlight the value of combining biomedical imaging with advanced AI techniques for secure authentication.
Author(s): Ananya Agarwal, Ampana Jayaram, Sravan Kumar Bhashyam, Harshith Harikrishnan,
Kewei Sha, Cliff Whitworth
The evolution of autonomous driving is shifting toward cognitive intelligence through
Vision-Language-Action (VLA) models that move beyond simple question-answering toward
direct trajectory generation. This research presents a multi-platform framework using
the Quanser QCar2 for real-time local LLM reasoning and the RobiX platform for large-scale
3D SLAM. By integrating these systems, we are developing a high-quality data pipeline
for ""Small Expert Reasoner"" models designed for efficient in-vehicle hardware execution.
Our methodology incorporates cutting-edge architectures, including OpenDriveVLA for
spatial grounding, AutoVLA for adaptive ""dual-mode"" reasoning, and DriveMLM for
aligning linguistic decisions with behavioral planning states. Utilizing the 64K-clip
LMDrive dataset, synthetic environmental stressors (rain, snow, fog), simulated sensor
data, and VLM-generated clips, we bridge the ""execution gap"" between conceptual
intent and low-level actuator control.
Crucially, this research systematically addresses four ""unsolved dilemmas"" currently
preventing Level 5 autonomy. First, we tackle numerical and spatial reasoning deficits,
specifically the waypoint ""coarseness"" inherent in discrete tokenizers. Second,
we confront the temporal reasoning and alignment gap in modeling dynamic scene evolution,
such as sequential turn signals. Third, we mitigate causal confusion and the interpretability-trust
paradox, where models may generate plausible but incorrect linguistic hallucinations
for their actions. Finally, we address the scarcity of rare edge case data by utilizing
generative world models to simulate scenarios that deterministic rules cannot capture.
By fusing deterministic sensor logic with the sophisticated reasoning of local LLMs,
we aim to deliver an autonomous experience that is collaborative, intuitive, and physically
grounded.
Clinical chatbot assistants are increasingly deployed to improve patient outcomes and support informed healthcare decision-making, yet existing solutions fail to leverage the full spectrum of electronic health record (EHR) data types. Most existing systems are limited in scope primarily leveraging structured data or isolated text inputs and often lack multimodal integration and contextual reasoning.
We present CareLens, a large language model (LLM)-based clinical chatbot that integrates structured and unstructured EHR data including clinical notes and medical images through a multimodal retrieval-augmented generation (RAG) framework. While LLMs are powerful for reasoning tasks, they often lack factual accuracy, especially for numerical queries (e.g., retrieving basic statistics such as counts, averages, etc.). CareLens overcomes this limitation by utilizing an intent detection layer to classify queries as either factual- or reasoning-based. The system routes factual queries to a structured query engine to retrieve accurate numerical results and explains these results in clinical terms via LLM. For reasoning with unstructured text and images, a fine-tuned NLP model and domain-specific vision model are used.
CareLens offers a scalable and interpretable solution that synthesizes and presents factual and contextual insights from diverse EHR data types spanning free-text narratives, diagnostic codes, and imaging while supporting natural human interaction via speech and text. This chatbot aims to enhance clinician decision-making and accelerate clinical research by enabling seamless navigation of complex patient information.
Diabetic Retinopathy (DR) is a leading cause of preventable blindness worldwide, yet early diagnosis remains difficult due to scarce annotated data and limited access to ophthalmic expertise. AI-based screening systems using Convolutional Neural Networks (CNNs) show promise for DR prediction through retinal image analysis, but they require thousands of labeled images—an expensive, time-intensive demand unsuitable for resource-constrained settings. Semi-supervised learning (SSL) offers a solution by leveraging limited labeled data alongside abundant unlabeled images; however, existing SSL methods for DR detection suffer from unreliable pseudo-labels.
We propose a novel SSL approach that adapts Area Under Margin Self-Training (AUM-ST) from natural language processing to medical imaging for DR detection. We replace BERT with BEIT (Bidirectional Encoder Representation from Image Transformers) to effectively process retinal fundus images while retaining AUM-ST’s reliability assessment mechanism. To preserve lesion features essential for diagnosis, we incorporate saliency-preserving strategies that avoid aggressive augmentations such as random rotations and crops.
Experimental validation on a real-world dataset of 757 retinal fundus images demonstrates superior performance over baselines including ResNet-18, ResNet-50, DenseNet-121, MobileNet-v2, and EfficientNet-B0. Our method achieves notable gains in both accuracy and F1 score when training with fewer than 50 labeled samples per class, substantially outperforming supervised approaches that require extensive annotations for clinical-grade performance.
These results highlight the practical viability of our framework for scalable DR screening in resource-limited environments, including rural healthcare facilities and developing regions. More broadly, this work advances SSL in healthcare AI by addressing annotation costs and limited clinical expertise, paving the way for reliable, accessible medical imaging solutions.
The increasing use of artificial intelligence (AI) across all use contexts has created a need for AI curricula. The few available AI literacy frameworks focus on using and understanding large language model AIs, such as audiovisual generation and chatbot-based AI, but these are not the only use cases of AI. EdgeAI is more commonplace than teachers realize, and is less emphasized in the AI literacy literature. However, it has unique affordances for learning.
This multi-stage design-based research project has undertaken both hardware and curriculum design. Phase one developed a custom learning board which includes a microprocessor, discrete sensors, and LCD panel output. This board is unique; design choices include utilizing five individual sensors in order to highlight the affordances and limitations of each, highlighting inputs versus outputs, and for durability in classroom environments.
Implementation and feedback from the initial implementation at a museum in Florida lead to streamlining background knowledge teaching in Module 1 and expanding hands-on opportunities. Initial feedback from high school implementations of a condensed curriculum include providing more student-centered materials for classes that provide more self-paced learning, the addition of screen-free and gamified activities to emphasize AI concepts to lead into the final AI for Social Good design activity.
Phase two is iterative curriculum design, implementation, teacher feedback, and revision. Multiple stages of teacher feedback will be continually solicited during the design and implementation phases in Texas classrooms during the 2025-2026 and 2026-2027 school years. The modules are: Introduction to IoT and Embedded Systems, Exploring the microcontroller without AI, AI Integration to the microcontroller - Edge Impulse, Challenging AI Hardware skills, and AHA! Social contribution.
The UNT College of Information and Dallas AI invite student researchers and educators to submit abstracts and poster proposals for InfoFORCE 2026: Information and AI Weekend, an interdisciplinary event focused on exploring the Future Opportunities in Research, Collaboration, and Education (FORCE) through the lens of AI and information.
This event welcomes submissions that highlight the intersection of AI with the disciplines represented within the College of Information:
*College of Information faculty can submit in collaboration with COI students.
Selection Process:
For questions, contact CI-Events@unt.edu or visit https://ci.unt.edu/infoforce.
We look forward to celebrating student research at InfoFORCE!