InfoFORCE 2026: Accepted Research Posters



Friday, April 10th  

A Conceptual Multi-Agent AI Framework for Integrating Social Determinants of Health into Equitable Diabetes Risk Stratification

Author(s): Sai Shivani Parimi, Bishnu Sarker, Alexis Lamar, Hari Chandana Vadde

Recent advances in artificial intelligence have led to the emergence of multi-agent health systems capable of integrating diverse data sources and supporting complex decision-making in healthcare.These systems demonstrate the potential to analyze multimodal health data and generate personalized insights but their application has largely focused on individual-level care and often overlooks broader population health challenges,particularly those related to health equity.In conditions such as Type 2 Diabetes (T2D), traditional risk stratification models frequently rely on clinical indicators alone,failing to account for the significant impact of social determinants of health (SDOH) on patient outcomes.This study proposes a conceptual multi-agent AI framework designed to enhance population-level diabetes risk stratification by integrating SDOH into electronic health record (EHR)-based models.The framework extends existing multi-agent architectures by introducing specialized and coordinated agents with distinct capabilities.A population orchestrator agent manages task allocation and system coordination, a data science agent enables SDOH-integrated predictive modeling and subgroup analysis;a domain expert agent ensures clinically grounded,evidence-based interpretation,and a health coach agent translates insights into culturally sensitive, actionable interventions.In addition,the framework introduces supporting agents,including health equity agent for bias monitoring, a data quality agent for ensuring data completeness, and an SDOH extraction agent for structuring unstructured social data.The proposed framework emphasizes equitable identification of at-risk populations, particularly individuals with moderate clinical risk but high social vulnerability.This work is conceptual and has not yet been implemented or empirically validated.Future research is required to evaluate model performance, system integration feasibility, and real-world impact.Despite these limitations, the framework provides a foundation for advancing equitable,AI-driven population health interventions. 

Accelerating Blood Cancer Research Through Big Data and Machine Learning

Author(s):  Laxmi Shravani Mamidala, Ashwitha Reddy Venna, Gayathri Orsu, Lakshmi Triveni Muthyala, Tharun Swaminathan Ravi Kumar

Blood cancer, especially leukemia, is one of the most dangerous diseases, and early, accurate diagnosis is critical for improving patient outcomes. Conventional microscopic analysis of blood smear images is time-consuming, highly subjective, and prone to variability. This research proposes an automated Blood Cancer Detection System using a combination of machine learning, deep learning, and computer vision techniques to identify and classify abnormal blood cells.

The proposed system uses a hybrid artificial intelligence approach. Deep features are extracted from microscopic images using a pre-trained ResNet50 convolutional neural network, and these features are classified using machine learning algorithms such as Logistic Regression, Support Vector Machine (SVM), Random Forest, XGBoost, and K-Means clustering. To improve classification, YOLOv8 is used for accurate object detection and localization of blood cells. Model interpretability is enhanced using Grad-CAM explainability techniques, which highlight the most important regions of the image for the classification decision.

The dataset contains 5000 labeled microscopic images of blood cells, representing the five most prominent cell types essential for leukemia diagnosis. Experimental results show excellent performance of the supervised models, with SVM and Logistic Regression achieving about 98–99% accuracy, while YOLOv8 demonstrates strong accuracy in blood cell detection.

As an additional contribution, this research proposes an intelligent clinical decision-support system that allows users to upload images, display predictions, analyze explainable heatmaps, and compare predictions across models. Future work may include integration with real-time clinical databases, the use of federated learning for privacy-preserving collaboration, and multimodal patient data to further improve prediction accuracy. Overall, this system shows the potential of combining deep learning, machine learning, and explainable AI to accelerate biomedical image analysis and support faster, more accurate blood cancer diagnosis.

Becoming Your Authentic Self in the Age of AI: Identity, Agency, and Digital Selfhood in Information Environments
Author(s):  Unknown

As artificial intelligence increasingly mediates how individuals learn, communicate, and access information, questions of identity and authenticity have become central to human experiences and expression. This poster explores how AI-driven information environments influence an individual’s ability to develop, express, and sustain an authentic sense of self. Grounded in humanistic psychology, Self-Determination Theory, narrative identity theory, and information science, this work synthesizes current research to examine the complex relationship between AI systems, digital information ecosystems, and identity formation. A conceptual model is proposed in which AI systems shape the information environment, which in turn influence identity processes such as self-reflection, autonomy, and narrative construction, ultimately affecting authentic self-expression. Findings from the literature suggest that AI tools can support self-awareness and creativity, yet algorithmic curation may also narrow identity exploration and dilute personal voice. Implications are offered for educators, AI designers, and researchers seeking to promote agency, transparency, and critical information literacy in AI-mediated contexts. This work highlights the importance of intentional design and reflective practice to ensure that authenticity—and the human capacity for meaning-making—remains possible in an increasingly algorithmic world.

Convolution Neural Network: A revolutionary turn in the history of computation. Automatic Detection of Defects and Quality Prediction in the Textile Industry
Author(s):  Kamrun Nahar, Sai Harshith Kolavasi, Krishna Mogaveera, Maimuna Patwary

This research focuses on the automatic detection of defects fabrics and quality prediction in the textile industry. Our study analyzed the finished textile product using a deep learning framework based, Convolution Neural Network or CNN in short. CNN is an artificial intelligence model that processes data and extracts meaningful insights. Just as human neurons receive input through our nervous system, CNNs receive input from cameras or sensors and process it through layered 2D and 3D structures. These networks can be trained to classify images, test patterns, and generate accurate outputs. Thus, we developed a supervised Convolutional Neural Network (CNN) Model that can detect and categorize and classify the defects in textile fabrics. High quality defective and non-defective images of various kinds were used to train and validate the CNN model architecture, to detect the good quality fabric automatically for inventory and final distribution.

We aim to create a system that can automatically detect, classify, and predict textile defects. Raw data was collected, cleaned, verified, and finally classified for its defect identification according to the industry standard. This project seeks not only to improve the quality assurance process in the textile industry but also potentially to revolutionize defect prediction and prevention. Integrating machine learning into textile manufacturing is crucial for enabling smart and automated quality control. By combining deep learning with domain expertise, we can create a system for textile industry’s employees or workers who mostly are work on finish products and test excellence that enhances fabric production quality, reduces material waste, minimizes costly errors in the production cycle, and support the evolution of intelligent textile manufacturing environment.

Edublox: Using Web Scraping and AI to Map, Compare, and Analyze University Programs Across US
Author(s):  Ampana Jayaram, Lizal Adhikari, Dr. Haihua Chen

If you've ever tried to compare academic programs across multiple universities, you already know how frustrating it can be. Every school has a different website, a different catalogue format, and a different way of presenting the same information. There is no single place where a student, advisor, or researcher can search across thousands of institutions, compare them side by side, and get personalised guidance until now.
Edublox was built to solve exactly that problem. The project collects academic program data from 6,186 accredited U.S. universities and brings it together into one clean, searchable interface. The challenge was that university websites are notoriously inconsistent some use modern JavaScript frameworks that traditional scrapers cannot read, others publish their catalogues as PDFs, and many have broken or missing links altogether. To address this, the system uses Crawl4AI, an open-source browser automation tool that operates like a real human browsing the web, rendering pages fully before extracting content. The pipeline automatically detects program listings, follows promising sub-pages, parses PDF catalogues, and recovers gracefully from failures. Each university entry is enriched with real institutional data including tuition costs, graduation rates, median graduate earnings, and Carnegie Classification.

What makes Edublox go beyond a simple directory is its AI comparison engine. Users can describe their personal preferences field of study, budget, location, career goals and receive an intelligent analysis.
This project sits at the intersection of information science, learning technologies, and applied AI and it raises a question worth asking: why should the data students need to make one of the most important decisions of their lives be this scattered and hard to find in the first place?

Efficient Depression Detection from Social Media Posts Using Hybrid Deep Learning
Author(s):  Ananya Agarwal, Bishnu Sarker

Depression is one of the most prevalent and debilitating mental health disorders worldwide, affecting emotional well-being and cognitive functioning, and the World Health Organization estimates that over 280 million people globally (around 5–6% of adults) suffer from depression, making it a leading cause of disability worldwide. Despite this widespread impact, more than 75% of affected individuals in low- and middle-income countries receive no treatment. We propose a depression detection framework in this study using a large-scale social media dataset collected from X (Twitter), Reddit, and Facebook, comprising approximately 244,388 posts, evenly balanced between depressive (label = 1) and non-depressive (label = 0) classes. To ensure data quality and semantic consistency, an extensive preprocessing pipeline was applied, including punctuation removal, stop-word elimination, and URL-link removal. For traditional machine learning models, textual features were extracted using TF-IDF vectorization, while deep learning models leveraged tokenized and padded sequences to capture contextual dependencies. Through careful preprocessing and developing hybrid deep learning architectures, we propose a CNN + logistic regression model that achieved an accuracy and an F1 score of 0.95, while a bidirectional CNN-LSTM captured the nuanced flow of emotional expression at 93% accuracy—both outperforming traditional approaches like linear SVM at 93.24%. Despite these promising results, several challenges remain, including dataset bias arising from platform-specific language, limited linguistic and cultural diversity, the dynamic and evolving nature of social media language, and the difficulty of distinguishing transient emotional expression from clinically significant depression. 

Learning Technologies in Sales Training: B2B Pitch Effectiveness
Author(s):  Kostiantyn Savchenko, Viktoriia Savchenko

This research explores the application of learning technologies in transforming B2B sales training through a structured, consultative approach. Traditional six-step B2C sales presentations have proven insufficient in engaging business partners, fostering long-term motivation, and driving active retail sales. In response, a 12-step B2B sales presentation model was developed, integrating two distinct consultative sales cycles (B2C and B2B), enhanced question-handling, and an embedded micro-training module using role-play and immediate feedback. This structure leverages principles of adult learning, performance-based training, and attention-driven content delivery to improve knowledge retention, engagement, and sales outcomes. Field experiments with companies such as 3M and Omega Bittner demonstrated significantly higher buy-in rates and measurable sales growth, even during market crises. The model has since been successfully adopted across multiple industries in Ukraine, illustrating the potential of learning technologies to reshape sales enablement strategies and improve business performance in B2B environments.

PolicyPulse.AI An Intelligent Insurance Monitoring Platform Using RAG, BERT, Pre-Trained AI Agents, and Python Full-Stack Development
Author(s):  Dhiresh Venkateshwarlu Katuri, Sri Ramya Karuturi, Srinivas Kalyan Rajarapu

PolicyPulse.AI is a smart insurance assistant and monitoring platform designed to help users manage health insurance and vehicle insurance in a single application. Existing insurance systems are often fragmented, difficult to understand, and mostly reactive, making it hard for users to track policies, understand coverage, and make timely decisions. This project addresses that gap by developing an AI-powered prototype that combines Retrieval-Augmented Generation (RAG), BERT-based document understanding, pre-trained AI agents, and Python full-stack development to deliver a more proactive and user-friendly experience.

The system allows users to upload insurance documents, which are processed using BERT transformers for document breakdown, contextual understanding, and extraction of key information such as coverage details, exclusions, premium information, and claim-related terms. This processed content is then integrated into a RAG pipeline, enabling the application to provide accurate, context-aware answers to user queries based on the actual policy content rather than generic generated responses. To further enhance intelligence, pre-trained AI agents are used to perform specialized tasks such as policy explanation, recommendation generation, renewal reminders, and risk-aware guidance.

The prototype is developed using a Python full-stack approach, with Streamlit as the frontend interface for rapid development, interactive dashboards, and user-friendly visualization. The initial version focuses on core features such as document upload, policy summarization, RAG-based question answering, insurance monitoring, and personalized recommendations. By combining document intelligence, retrieval-based reasoning, and agent-driven support, PolicyPulse.AI aims to transform insurance management from a static process into an interactive, intelligent, and proactive digital experience.


ProComp: A Problem-Driven Multi-Agent Framework for Comparative Paper Ranking
Author(s):  Laxmi Shravani Mamidala, Shree Harshini Mamidala,  Haoxuan Zhang, Professor Dr. Ting Xiao

The rapid growth of scientific publications has placed significant pressure on traditional peer-review systems, making it difficult to maintain consistent, timely, and high-quality evaluations. Human reviewers often face challenges such as limited time, reviewer fatigue, and subjective biases, which can lead to variability in review quality. To address these issues, this paper introduces ProComp, a multi-agent large language model (LLM) framework designed to automate structured scholarly paper evaluation by emulating the roles of expert peer reviewers.

ProComp decomposes the review process into multiple specialized evaluation agents, each responsible for assessing a distinct aspect of a research paper. The framework employs three expert reviewer agents that independently analyze problem formulation and motivation, methodological soundness and technical rigor, and novelty and research contribution. Each agent produces structured assessments and quantitative scores aligned with common academic review criteria. To enhance contextual understanding and support informed evaluation, the system incorporates a Retrieval-Augmented Generation (RAG) component that retrieves related work and citation context from external scholarly databases.

A fourth agent acts as a comparative meta-reviewer, aggregating the outputs of the expert agents to generate a consolidated evaluation and final recommendation. The proposed framework is evaluated on a dataset of 194 research papers, comparing multiple LLM backends and analyzing their agreement with human review signals. Results demonstrate that the multi-agent architecture produces structured, interpretable evaluations and maintains meaningful correlation with human judgments. Ablation analysis further investigates the influence of individual reviewer agents on final scoring decisions.

Overall, ProComp demonstrates the potential of collaborative LLM agents to support scalable, transparent, and consistent automated scholarly evaluation, providing a promising direction for augmenting future academic peer-review systems.

Retrieval-Augmented Phishing Detection: Leveraging Context Engineering for Cyber Literacy in K-12 Education
Author(s):  Tayiba Raheem, Indira Priyaharshini Sundarrajan, Gahangir Hossain

Phishing remains a critical cybersecurity threat that bypasses traditional keyword-based detection through increasingly sophisticated social engineering. This study investigates whether Retrieval-Augmented Generation can improve phishing email classification by grounding a language model in dynamically retrieved threat intelligence. We constructed a dual-retriever pipeline combining semantic vector search and TF-IDF keyword matching to supply contextual examples to a Flan-T5 classifier. The system was evaluated across three detection approaches and tested for robustness against data poisoning attacks at varying injection rates. An interactive demonstration with confidence scores was developed to support K-12 cybersecurity literacy education, bridging the gap between AI-driven security  tools and student comprehension.​

SafeRoute - Route Based Accident Risk Classification Using Historical Accident Data and Live Weather Conditions
Author(s):  Jahnavi Kovelamudi, Anuhya Koyyana, Syam Sai Konakalla, Kavya Sai Kotha, Whitworth Clifford  

Route optimization in navigation systems focuses on speed and distance rather than safety. Although current maps provide certain safety features, there are no accident risk predictions, which are based on historical data and current circumstances. There are 7.7 million historical accident data points in the US, covering the period 2016-2023, which creates a unique opportunity for safer routes by combining this data with real-time data.
 
We are developing the SafeRoute that combines historical patterns for the severity of accidents with live weather data by a route-based risk classification. The initial examination of 7.7 million US accidents found in the Moosavi dataset indicates some crucial trends. 47% of the accidents are due to traffic signals and junctions. The averages on the severity at nighttime are 5% higher than in the daytime. Weather conditions such as freezing rain and heavy snow are associated with high severity accidents.
 
Our methodology incorporates classification models based with feature aggregation to represent infrastructure characteristics and temporal patterns of accidents. Models like ensemble and gradient boosting are used to classify the route segments based on their severity. In addition, the model offers an understanding of each prediction representing the factors that are included in the prediction of the route segment, such as infrastructure type, weather conditions, and time of day.
 
By providing clear, explainable route safety assessments, SafeRoute enables informed driver decisions and supports evidence-based urban traffic management policies. This work demonstrates the practical value of integrating historical accident patterns with real-time data to address a critical safety gap in modern transportation systems.

Stealth Myocardium: Prediction of Silent Myocardial Infarction (Heart Attack Without Symptoms) Using AI/ML Model
Author(s):  Preethi Dasari, Sai Sahasra Katuri, Jayakrishna Pamuru, Priyanka Mada, Yuvasri Kumar

Silent myocardial infarction is clinically significant yet underdiagnosed condition that occurs without noticeable symptoms, often remaining undetected until later identified through electrocardiographic or imaging findings. These unnoticed events can lead to severe complications, including recurrent myocardial infarction, revascularisation, rehospitalization, and increased risk of sudden cardiac death. This highlights a critical gap in early cardiovascular risk detection and preventive care.

The present study aims to address this hidden clinical burden by developing an artificial intelligence AI -driven model for early prediction of silent myocardial infarction. The motivation is to identify high-risk individuals before the onset of major cardiovascular events, thereby enabling timely intervention and improved clinical outcomes. 

Our project focuses on detecting silent myocardial infarction (MI) using AI/ML models by analyzing demographic data such as age, gender, and height, along with ECG data. We utilized the PTB-XL dataset, which consists of 12-lead ECG records, and classified cases into normal, silent myocardial infarction, and acute myocardial infarction.Our model is a two-branch deep learning (DL) architecture developed using TensorFlow and Keras. One branch processes ECG signals using Convolutional Neural Networks (CNNs), while the other branch incorporates demographic inputs. The outputs from both branches are combined and passed through a softmax layer to classify cases into the three target categories.The model was trained and evaluated using performance metrics including accuracy, precision, recall, and F1-score.

For real-world applicability, we developed a tool that converts paper based ECG images into digital signals using OpenCV, NumPy, waveform databases. The images are de-skewed and divided into 12 leads, after which the waveforms are extracted, filtered, and resampled. These processed digital signals are then passed into the CNN model for prediction.


Saturday, April 11th  

A Deep Fusion Framework for Next-Generation Tongue Biometrics: Secure, Stable, and Spoof-Resistant Identification (DFGT)

Author(s): Abhinay Peeta, Venkata Haritha Sai Sri Animilli & Dr Sayed K. Shah

Biometric authentication systems based on conventional modalities such as fingerprints or facial recognition are increasingly vulnerable to spoofing, environmental variability, and privacy concerns. In response, this work introduces a novel tongue-based biometric identification framework designed for enhanced security and robustness within biomedical and AI-driven authentication contexts. Leveraging the intrinsic morphological stability of the tongue, its protected intra-oral environment, and its inherent difficulty to forge or replicate, we propose this modality as a next-generation biometric trait for personal identification.

At the heart of our system lies a deep learning architecture: a convolutional neural network (CNN) that automatically extracts high-dimensional, discriminative features capturing fissure patterns, surface texture, and geometric tongue shape. To further boost performance and anti-spoofing resilience, we integrate a feature-level fusion strategy that combines these learned embeddings with complementary hand-crafted descriptors thereby forming a unified, highly discriminative feature vector. Experimental validation demonstrates that this hybrid approach significantly improves identification accuracy while reinforcing the modality’s robustness against adversarial or spoofing attempts. The results affirm the viability of the tongue as a reliable biometric modality and highlight the value of combining biomedical imaging with advanced AI techniques for secure authentication.

Advancing Cognitive Autonomy: A Multi-Platform Framework for Spatially-Grounded and Adaptive VLA Models

Author(s): Ananya Agarwal, Ampana Jayaram, Sravan Kumar Bhashyam, Harshith Harikrishnan, Kewei Sha, Cliff Whitworth

The evolution of autonomous driving is shifting toward cognitive intelligence through Vision-Language-Action (VLA) models that move beyond simple question-answering toward direct trajectory generation. This research presents a multi-platform framework using the Quanser QCar2 for real-time local LLM reasoning and the RobiX platform for large-scale 3D SLAM. By integrating these systems, we are developing a high-quality data pipeline for ""Small Expert Reasoner"" models designed for efficient in-vehicle hardware execution.
Our methodology incorporates cutting-edge architectures, including OpenDriveVLA for spatial grounding, AutoVLA for adaptive ""dual-mode"" reasoning, and DriveMLM for aligning linguistic decisions with behavioral planning states. Utilizing the 64K-clip LMDrive dataset, synthetic environmental stressors (rain, snow, fog), simulated sensor data, and VLM-generated clips, we bridge the ""execution gap"" between conceptual intent and low-level actuator control.
Crucially, this research systematically addresses four ""unsolved dilemmas"" currently preventing Level 5 autonomy. First, we tackle numerical and spatial reasoning deficits, specifically the waypoint ""coarseness"" inherent in discrete tokenizers. Second, we confront the temporal reasoning and alignment gap in modeling dynamic scene evolution, such as sequential turn signals. Third, we mitigate causal confusion and the interpretability-trust paradox, where models may generate plausible but incorrect linguistic hallucinations for their actions. Finally, we address the scarcity of rare edge case data by utilizing generative world models to simulate scenarios that deterministic rules cannot capture. By fusing deterministic sensor logic with the sophisticated reasoning of local LLMs, we aim to deliver an autonomous experience that is collaborative, intuitive, and physically grounded.

An Application Framework For Malignant Blood Disorder Detection From Peripheral Blood Smear Pathology Images Using Deep Learning
Author(s): Bishnu Sarker, Ananya Agarwal

Malignant blood disorders, including leukemias, lymphomas, and myelodysplastic syndromes, often manifest through subtle morphological abnormalities in peripheral blood that precede overt clinical symptoms, leading to delayed diagnosis and reliance on invasive bone marrow biopsy. Although peripheral blood smear microscopy is routinely performed, manual interpretation is time-consuming, subjective, and limited in its ability to detect early, low-prevalence malignant features. In this work-in-progress, we present a web application framework for identifying malignant blood disorder based on digitized peripheral blood smear images using deep learning. We have used transformer-based and convolutional neural network models that learn multi-scale representations of cellular morphology, spatial organization, and population-level patterns associated with malignancy. The long-term goal is to establish a scalable, objective, and interpretable image-based AI system for early detection and risk stratification of malignant blood disorders from routine peripheral blood smears.

CareLens: A Clinical Decision Support Assistant Leveraging Multimodal EHR Data
Author(s): Gowtham Vuppaladhadiam, Baby Jahnavi Kovelamudi, Samuel Sibbi Rayan Jerome, Ana D. Cleveland, K S M Tozammel Hossain

Clinical chatbot assistants are increasingly deployed to improve patient outcomes and support informed healthcare decision-making, yet existing solutions fail to leverage the full spectrum of electronic health record (EHR) data types. Most existing systems are limited in scope primarily leveraging structured data or isolated text inputs and often lack multimodal integration and contextual reasoning. 

We present CareLens, a large language model (LLM)-based clinical chatbot that integrates structured and unstructured EHR data including clinical notes and medical images through a multimodal retrieval-augmented generation (RAG) framework.  While LLMs are powerful for reasoning tasks, they often lack factual accuracy, especially for numerical queries (e.g., retrieving basic statistics such as counts, averages, etc.). CareLens overcomes this limitation by utilizing an intent detection layer to classify queries as either factual- or reasoning-based. The system routes factual queries to a structured query engine to retrieve accurate numerical results and explains these results in clinical terms via LLM. For reasoning with unstructured text and images, a fine-tuned NLP model and domain-specific vision model are used.

CareLens offers a scalable and interpretable solution that synthesizes and presents factual and contextual insights from diverse EHR data types spanning free-text narratives, diagnostic codes, and imaging while supporting natural human interaction via speech and text. This chatbot aims to enhance clinician decision-making and accelerate clinical research by enabling seamless navigation of complex patient information.

Enabling Federated Object Detection for Connected Autonomous Vehicles: A Deployment-Oriented Evaluation
Author(s): Komala Subramanyam Cherukuri, Kewei Sha, Zhenhua Huang

Object detection is crucial for Connected Autonomous Vehicles (CAVs) to perceive their surroundings and make safe driving decisions. Centralized training of object detection models often achieves promising accuracy, fast convergence, and simplified training process, but it falls short in scalability, adaptability, and privacy-preservation. Federated learning (FL), by contrast, enables collaborative, privacy-preserving, and continuous training across naturally distributed CAV fleets. However, deploying FL in real-world CAVs remains challenging due to the substantial computational demands of training and inference, coupled with highly diverse operating conditions. Practical deployment must address three critical factors: (i) heterogeneity from non-IID data distributions, (ii) constrained onboard computing hardware, and (iii) environmental variability such as lighting and weather, alongside systematic evaluation to ensure reliable performance. This work introduces the first holistic deployment-oriented evaluation of FL-based object detection in CAVs, integrating model performance, system-level resource profiling, and environmental robustness. Using state-of-the-art detectors, YOLOv5, YOLOv8, YOLOv11, and Deformable DETR, evaluated on the KITTI, BDD100K, and nuScenes datasets, we analyze trade-offs between detection accuracy, computational cost, and resource usage under diverse resolutions, batch sizes, weather and lighting conditions, and dynamic client participation, paving the way for robust FL deployment in CAVs.

Evidence Based Biomedical Question Answering with BioMCP Server
Author(s): Umme Israt Afroz

Accurate biomedical question answering depends on linking responses to trustworthy scientific evidence. Large language models (LLMs), while effective at generating fluent and informative responses, they often produce hallucinations that reduce reliability in sensitive areas like clinical decision support and biomedical research. To address this challenge, we propose a framework for evidence-grounded biomedical Q&A using the BioMCP server, an open-source Model Context Protocol (MCP) implementation available on GitHub. The BioMCP server provides a stable interface for retrieving relevant PubMed articles, which are dynamically linked to each sentence of an LLM-generated answer. Our approach involves querying BioMCP with biomedical questions, retrieving candidate literature passages, and integrating these as verifiable references that either support or contradict the generated statements. By prioritizing contradictory evidence, the system encourages critical evaluation and mitigates overconfidence in LLM outputs. This work demonstrates how MCP-based integration of PubMed with LLMs can improve transparency, reliability, and trustworthiness in biomedical question answering.

Low-Resource Diabetic Retinopathy Screening via AUM-ST with Vision Transformer
Author(s): S M Saiful Islam Badhon, Praneeth Rikka, Ana Cleveland, and K S M Tozammel Hossain

Diabetic Retinopathy (DR) is a leading cause of preventable blindness worldwide, yet early diagnosis remains difficult due to scarce annotated data and limited access to ophthalmic expertise. AI-based screening systems using Convolutional Neural Networks (CNNs) show promise for DR prediction through retinal image analysis, but they require thousands of labeled images—an expensive, time-intensive demand unsuitable for resource-constrained settings. Semi-supervised learning (SSL) offers a solution by leveraging limited labeled data alongside abundant unlabeled images; however, existing SSL methods for DR detection suffer from unreliable pseudo-labels.

We propose a novel SSL approach that adapts Area Under Margin Self-Training (AUM-ST) from natural language processing to medical imaging for DR detection. We replace BERT with BEIT (Bidirectional Encoder Representation from Image Transformers) to effectively process retinal fundus images while retaining AUM-ST’s reliability assessment mechanism. To preserve lesion features essential for diagnosis, we incorporate saliency-preserving strategies that avoid aggressive augmentations such as random rotations and crops.

Experimental validation on a real-world dataset of 757 retinal fundus images demonstrates superior performance over baselines including ResNet-18, ResNet-50, DenseNet-121, MobileNet-v2, and EfficientNet-B0. Our method achieves notable gains in both accuracy and F1 score when training with fewer than 50 labeled samples per class, substantially outperforming supervised approaches that require extensive annotations for clinical-grade performance.

These results highlight the practical viability of our framework for scalable DR screening in resource-limited environments, including rural healthcare facilities and developing regions. More broadly, this work advances SSL in healthcare AI by addressing annotation costs and limited clinical expertise, paving the way for reliable, accessible medical imaging solutions.

Measuring Cognitive Presence in Online Discussions: Automated Detection and Instructional Insights from a MOOC Context
Author(s): Fengjiao Tu, Ji Hyun Yu, Haihua Chen, Junhua Ding, Liu Dong, Chi-Jia Hsieh, Hannah Kim, Sunnie Lee Watson

This study develops and validates an interpretable transformer-based model for detecting cognitive presence (CP) in MOOC discussions to serve as a diagnostic foundation for instructional design and facilitation. Specifically, we built this model by fine-tuning the RoBERTa architecture with theory-derived structural cues from the Community of Inquiry framework to capture causal, synthesizing, and action-oriented language. Using data from six Social Work MOOCs (4,260 learners; 227,863 sentences), the model classified discourse into four CP phases and a fifth non-CP category. The cue-augmented RoBERTa achieved strong performance (macro F1 = 0.714; κ = 0.666), and SHAP analyses confirmed the model’s internal logic aligns with theory-consistent linguistic markers, ensuring the pedagogical transparency of the results. Learner-level hierarchical regressions revealed that while discussion quantity predicted behavioral persistence, higher-order CP indicators (Integration and Resolution) functioned as unique diagnostic proxies for motivation, satisfaction, and perceived knowledge transfer. These findings shift the role of CP analytics from descriptive observation to prescriptive action through a proposed Pedagogical Diagnostic & Engineering (PDE) framework. By mapping automated structural signals to specific instructional failure points, this framework offers concrete tactics for real-time scaffolding, such as Socratic prompts and practitioner spotlights, to move learners beyond exploratory monologues toward deeper knowledge construction and professional application.  

Personal Health Agent (PHA): Multiple-Agent AI for Mental Health Support
Author(s): Naaz Fatima Begum, Aishwarya Salla

Google research has recently conducted a study (Heydari et al., 2025) that underscores the future of personalized healthcare with the help of multi-agent architectures. Multi-agent systems assign tasks to specialized agents, including data science, clinical, and health coaching, to more accurately domain-knowledge decision-making, compared to single AI agents, which are usually constrained with large, complex datasets and do not have coordinating capabilities with other agents in the system. This solution helps to integrate different sources of data, such as electronic health records and wearables, to provide more individualized and proactive care. Building on this concept, we developed a blueprint for a mental health solution that extends the architecture by introducing additional agents, including a health equity agent and a resource allocation agent, aimed at addressing disparities in care and optimizing the distribution of mental health resources. Overall, this multi-agent framework offers a more scalable, equitable, and patient-centered alternative to traditional single-agent models.

The personal health agent is headed by an Orchestrator that coordinates specialized agents: Data Science Agent predicts based on the datasets: CDC records, EHRs, Social Vulnerability Index (SVI). Domain Expert Agent uses DSM-5 knowledge and evidence-based practice to provide insight. Health Coach Agent delivers personalized interventions. The Health Equity and Resource Allocation Agents collaborate to define disparities and populations at risk with the help of socioeconomic and access-related data. The resource allocation agent represents the most efficient option for distributing healthcare resources based on the needs and socioeconomic status of the patient and local services. This system integrates the input of user-reported symptoms, wearable data, hospital data, and socioeconomic factors to generate personalized recommendations. 

The model aims to offer individual and patient-centered solutions that meet personal needs and contribute to effective mental health management.

Pioneering a Comprehensive AI and Microelectronics Curriculum in High School Engineering Classes
Author(s): Megan Barnes, Clelia Perez, Woorin Hwang, Yessi Ambarwati, Anany Sharma, Talar Terzian, Maryam Babaee, Lauren Eutsler, Andrea Salgado, Pasha Antonenko, D. Donihue, S. Bhunia, B. Chatterjee

The increasing use of artificial intelligence (AI) across all use contexts has created a need for AI curricula. The few available AI literacy frameworks focus on using and understanding large language model AIs, such as audiovisual generation and chatbot-based AI, but these are not the only use cases of AI. EdgeAI is more commonplace than teachers realize, and is less emphasized in the AI literacy literature. However, it has unique affordances for learning. 

This multi-stage design-based research project has undertaken both hardware and curriculum design. Phase one developed a custom learning board which includes a microprocessor, discrete sensors, and LCD panel output. This board is unique; design choices include utilizing five individual sensors in order to highlight the affordances and limitations of each, highlighting inputs versus outputs, and for durability in classroom environments. 

Implementation and feedback from the initial implementation at a museum in Florida lead to streamlining background knowledge teaching in Module 1 and expanding hands-on opportunities. Initial feedback from high school implementations of a condensed curriculum include providing more student-centered materials for classes that provide more self-paced learning, the addition of screen-free and gamified activities to emphasize AI concepts to lead into the final AI for Social Good design activity. 

Phase two is iterative curriculum design, implementation, teacher feedback, and revision. Multiple stages of teacher feedback will be continually solicited during the design and implementation phases in Texas classrooms during the 2025-2026 and 2026-2027 school years. The modules are: Introduction to IoT and Embedded Systems, Exploring the microcontroller without AI, AI Integration to the microcontroller - Edge Impulse, Challenging AI Hardware skills, and AHA! Social contribution. 

Smart Learning: AI-Driven Cybersecurity Training for Entrepreneurs
Author(s): Esther Eze

Small business owners are facing more cyber attacks, but they often can't get affordable training that fits their needs. Current cybersecurity education usually has the same content for everyone, which doesn't work well because people learn differently, have different skill levels, and threats change fast. This poster looks at research on a new AI e-learning platform that personalizes cybersecurity training for entrepreneurs. This system uses machine learning to watch how students perform and find their weak spots. Then, it changes how hard the content is and how it's taught. It also uses natural language processing to give helpful feedback. Data tracks how each person is doing and suggests extra lessons when needed. By customizing the experience and constantly changing to fit the student, the goal is to give training that is relevant and gets results for each entrepreneur. Early tests show that using AI to adapt the training helps people remember what they learned and keeps them interested compared to regular online courses. The platform can also be easily expanded to reach more people, giving entrepreneurial communities without resources access to good cybersecurity education. This study pushes forward information technology and AI by showing how adaptive learning can help close the gap between cybersecurity experts and small business owners. The project also points out chances for AI to be used in job training and to help colleges, businesses, and small businesses work together to improve cybersecurity.

Socio-Technical Fit of AI Coordination Agents in Refugee Pediatric SAM Care – A SEIPS 2.0 Analysis​
Author(s): Tayiba Raheem, Gahangir Hossain

This study applies the SEIPS 2.0 framework to evaluate the socio- technical fit of an AI Coordination Agent designed for refugee pediatric Severe Acute Malnutrition (SAM) care. ​Using structured component mapping and scenario-based heuristic evaluation. We decompose the refugee SAM care work system into five SEIPS 2.0 components : People, Tasks, Tools, Organization, Environment . Then we assess how the Coordination Agent aligns with each. ​The evaluation reveals that AI-assisted coordination improves fit across People, Tasks, and Tools components. ​But introduces tensions in Organization and Environment dimensions, particularly infrastructure constraints and institutional readiness​. These findings yield actionable design requirements for AI systems in resource-constrained humanitarian settings, where socio-technical misalignment can compromise patient safety.​

The UNT College of Information and Dallas AI invite student researchers and educators to submit abstracts and poster proposals for InfoFORCE 2026: Information and AI Weekend, an interdisciplinary event focused on exploring the Future Opportunities in Research, Collaboration, and Education (FORCE) through the lens of AI and information.

This event welcomes submissions that highlight the intersection of AI with the disciplines represented within the College of Information:

  • Information Science
  • Library Science
  • Health Informatics
  • Data Science
  • Learning Technologies
  • Linguistics

*College of Information faculty can submit in collaboration with COI students.

Selection Process:

  • Selection: 24 posters total (12 posters per day) selected by College of Information Faculty
  • Judging: The faculty and invited industry partners will judge posters prior to the event. Twelve finalists will be selected and notified by March 30th. Awards will be given for 1st, 2nd, and 3rd place across the entire group of selected posters. The top 3 winners will be announced at 4:00pm on Saturday, April 11th following the last poster session.
  • Lightening Talk: Finalists will be invited to give a 7-10 minute lightening talk on the same day as your scheduled poster presentation. This invitation will come on March 30th along with your finalist notification. 
  • Selection Criteria:
    - Research quality and rigor - 25%
    - Clarity and organization - 25%
    - Poster design and visual communication - 10%
    - Use of sources and academic integrity - 15%
    - Overall impact - 25%

For questions, contact CI-Events@unt.edu or visit https://ci.unt.edu/infoforce.

We look forward to celebrating student research at InfoFORCE!