Scholarly Works, Computer Science
Permanent URI for this collection
Research articles, presentations, and other scholarship
Browse
Recent Submissions
- Mixed-reality art as shared experience for cross-device users: Materialize, understand, and exploreMoon, Hayoun; Saade, Mia; Enriquez, Daniel; Duer, Zachary R.; Moon, Hye Sung; Lee, Sang Won; Jeon, Myounghoon (Academic Press - Elsevier, 2024-10)Virtual reality (VR) has opened new possibilities for creative expression, while the 360-degree head-worn display (HWD) delivers a fully immersive experience in the world of art. The immersiveness, however, comes with the cost of blocking out the physical world, including bystanders without an HWD. Therefore, VR experiences in public (e.g., galleries, museums) often lack social interactivity, which plays an important role in forming aesthetic experiences. In the current study, we explored the application of a crossdevice mixed reality (MR) platform in the domain of art to enable social and inclusive experiences with artworks that utilize VR technology. Our concept of interest features co-located audiences of HWD and mobile device users who interact across physical and virtual worlds. We conducted focus groups (N=22) and expert interviews (N=7) to identify the concept’s potential scenarios and fundamental components, as well as expected benefits and concerns. We also share our process of creating In-Between Spaces, an interactive artwork in MR that encourages social interactivity among crossdevice audiences. Our exploration presents a prospective direction for future VR/MR aesthetic content, especially at public events and exhibitions targeting crowd audiences.
- CLOSUREX: Compiler Support for Correct Persistent FuzzingRanjan, Rishi; Paterson, Ian; Hicks, Matthew (ACM, 2025-02-03)Fuzzing is a widely adopted and pragmatic methodology for bug hunting as a means of software hardening. Research reveals that increasing fuzzing throughput directly increases bug discovery rate. The highest performance fuzzing strategy is persistent fuzzing, which reuses a single process for all test cases by looping back to the start upon completion, instead of exiting. This eliminates all process creation, initialization, and tear-down costs—which are on-par with execution cost. Unfortunately, persistent fuzzing leads to semantically inconsistent program states because process state changes from one test case remain for subsequent test cases. This semantic inconsistency results in missed crashes, false crashes, and overall incorrectness that undermines fuzzer effectiveness. We observe that existing fuzzing execution mechanisms exist on a continuum, based on the amount of state that gets discarded and restored between test cases. We present ClosureX, a fuzzing execution mechanism that sits at a new spot on this state restoration continuum, where only testcase- execution-specific state is reset. This fine-grain state restoration provides near-persistent performance with the correctness of heavyweight state restoration. We construct ClosureX as a set of LLVM passes that integrate with AFL++. Our evaluation on ten popular open-source fuzzing targets show that ClosureX maintains semantic correctness, while increasing test case execution rate by over 3.5x, on average, compared to AFL++. ClosureX also finds bugs more consistently and 1.9x faster than AFL++, with ClosureX discovering 15 0-day bugs (4 CVEs).
- Systematic CXL Memory Characterization and Performance Analysis at ScaleLiu, Jinshu; Hadian, Hamid; Wang, Yuyue; Berger, Daniel; Nguyen, Marie; Jian, Xun; Noh, Sam; Li, Huaicheng (ACM, 2025-03-30)Compute Express Link (CXL) has emerged as a pivotal interconnect for memory expansion. Despite its potential, the performance implications of CXL across devices, latency regimes, processors, and workloads remain underexplored. We present Melody, a framework for systematic characterization and analysis of CXL memory performance. Melody builds on an extensive evaluation spanning 265 workloads, 4 real CXL devices, 7 latency levels, and 5 CPU platforms. Melody yields many insights: workload sensitivity to sub-μs CXL latencies (140-410ns), the first disclosure of CXL tail latencies, CPU tolerance to CXL latencies, a novel approach (Spa) for pinpointing CXL bottlenecks, and CPU prefetcher inefficiencies under CXL.
- PhasePrint: Exposing Cloud FPGA Fingerprints by Inducing Timing Faults at RuntimeMahmod, Jubayer; Hicks, Matthew (ACM, 2025-03-30)Cloud FPGAs, with their scalable and flexible nature, are rapidly gaining traction as go-to hardware acceleration platforms for compute-intensive workloads. However, their increasing adoption introduces unique security challenges. The hardware-level access that FPGAs provide leads to many vulnerabilities, including the leakage of sensitive information through data remanence and the creation of analog-domain covert channels among users. A foundational requirement in these scenarios is the ability to target an individual FPGA; knowing this, cloud vendors prevent FPGA localization by restricting access to low-level information of the underlying hardware. Beyond aiding adversaries, FPGA localization enables defenders to strategically rotate FPGA usage, preventing prolonged exposure that can lead to confidential data leakage due to long-term data remanence. This paper introduces PhasePrint, a cloud FPGA localization approach using dynamic timing faults in functionally valid circuits. PhasePrint induces timing faults in a specially crafted circuit at runtime and infers delay characteristics from the resulting error pattern—without incorporating information sources blocked by cloud vendors. PhasePrint utilizes an FPGA’s internal clock synthesizer to derive a clock pair with a strict phase relationship. By adjusting the phase relationship of these clocks, PhasePrint intentionally causes timing faults at runtime that reveal manufacturing variations among FPGA chips. We transform these fault locations into feature vectors to create device signatures and train a multiclass classifier on a dataset from 300 unique FPGAs across four AWS geographic regions. This entirely on-chip signature extraction method achieves >99% accuracy, operates 13× faster, and costs 92% less than the state-of-the-art.
- Solid State Drive Targeted Memory-Efficient Indexing for Universal I/O Patterns and Fragmentation DegreesIm, Junsu; Kim, Jeonggyun; Oh, Seonggyun; Koo, Jinhyung; Park, Juhyung; Chwa, Hoon Sung; Noh, Sam H.; Lee, Sungjin (ACM, 2025-03-30)Thanks to the advance of device scaling technologies, the capacity of SSDs is rapidly increasing. Such increase, however, comes at the cost of a huge index table requiring large DRAM. To provide reasonable performance with less DRAM, various index structures exploiting locality and regularity of I/O references have been proposed. However, they provide deteriorated performance depending on I/O patterns and storage fragmentation. This paper proposes a novel approximate index structure, called AppL, which combines memoryefficient approximate indices and an LSM-tree that has an append-only and sorted nature. AppL reduces the index size to 6∼8-bits per entry, which is considerably smaller than the typical index structures requiring 32∼64-bits, and maintains such high memory efficiency irrespective of locality and fragmentation. By alleviating memory pressure, AppL achieves 33.6∼72.4% shorter read latency and 28.4%∼83.4% higher I/O throughput than state-of-the-art techniques.
- Enhancing Immersive Sensemaking with Gaze-Driven Recommendation CuesTahmid, Ibrahim Asadullah; North, Chris; Davidson, Kylie; Whitley, Kirsten; Bowman, Doug (ACM, 2025-03-24)Sensemaking is a complex task that places a heavy cognitive demand on individuals. With the recent surge in data availability, making sense of vast amounts of information has become a significant challenge for many professionals, such as intelligence analysts. Immersive technologies such as mixed reality offer a potential solution by providing virtually unlimited space to organize data. However, the difficulty of processing, filtering relevant information, and synthesizing insights remains. We proposed using eye-tracking data from mixed reality head-worn displays to derive the analyst’s perceived interest in documents and words, and convey that part of the mental model to the analyst. The global interest of the documents is reflected in their color, and their order on the list, while the local interest of the documents is used to generate focused recommendations for a document. To evaluate these recommendation cues, we conducted a user study with two conditions: a gaze-aware system, EyeST, and a “Freestyle” system without gaze-based visual cues. Our findings reveal that the EyeST helped analysts stay on track by reading more essential information while avoiding distractions. However, this came at the cost of reduced focused attention and perceived system performance. The results of our study highlight the need for explainable AI in human-AI collaborative sensemaking to build user trust and encourage the integration of AI outputs into the immersive sensemaking process. Based on our findings, we offer a set of guidelines for designing gaze-driven recommendation cues in an immersive environment.
- KHAIT: K-9 Handler Artificial Intelligence Teaming for Collaborative SensemakingWilchek, Matthew; Wang, Linhan; Dickinson, Sally; Feuerbacher, Erica N.; Luther, Kurt; Batarseh, Feras A. (ACM, 2025-03-24)In urban search and rescue (USAR) operations, communication between handlers and specially trained canines is crucial but often complicated by challenging environments and the specific behaviors canines are trained to exhibit when detecting a person. Since a USAR canine often works out of sight of the handler, the handler lacks awareness of the canine’s location and situation, known as the “sensemaking gap.” In this paper, we propose KHAIT, a novel approach to close the sensemaking gap and enhance USAR effectiveness by integrating object detection-based Artificial Intelligence (AI) and Augmented Reality (AR). Equipped with AI-powered cameras, edge computing, and AR headsets, KHAIT enables precise and rapid object detection from a canine’s perspective, improving survivor localization. We evaluate this approach in a real-world USAR environment, demonstrating an average survival allocation time decrease of 22%, enhancing the speed and accuracy of operations.
- Mental Models of Generative AI Chatbot EcosystemsWang, Xingyi; Wang, Xiaozheng; Park, Sunyup; Yao, Yaxing (ACM, 2025-03-24)The capability of GenAI-based chatbots, such as ChatGPT and Gemini, has expanded quickly in recent years, turning them into GenAI Chatbot Ecosystems. Yet, users’ understanding of how such ecosystems work remains unknown. In this paper, we investigate users’ mental models of how GenAI Chatbot Ecosystems work. This is an important question because users’ mental models guide their behaviors, including making decisions that impact their privacy. Through 21 semi-structured interviews, we uncovered users’ four mental models towards first-party (e.g., Google Gemini) and third-party (e.g., ChatGPT) GenAI Chatbot Ecosystems. These mental models centered around the role of the chatbot in the entire ecosystem.We further found that participants held a more consistent and simpler mental model towards third-party ecosystems than the first-party ones, resulting in higher trust and fewer concerns towards the thirdparty ecosystems. We discuss the design and policy implications based on our results.
- CLEAR: Towards Contextual LLM-Empowered Privacy Policy Analysis and Risk Generation for Large Language Model ApplicationsChen, Chaoran; Zhou, Daodao; Ye, Yanfang; Li, Toby; Yao, Yaxing (ACM, 2025-03-24)The rise of end-user applications powered by large language models (LLMs), including both conversational interfaces and add-ons to existing graphical user interfaces (GUIs), introduces new privacy challenges. However, many users remain unaware of the risks. This paper explores methods to increase user awareness of privacy risks associated with LLMs in end-user applications. We conducted five co-design workshops to uncover user privacy concerns and their demand for contextual privacy information within LLMs. Based on these insights, we developed CLEAR (Contextual LLM-Empowered Privacy Policy Analysis and Risk Generation), a just-in-time contextual assistant designed to help users identify sensitive information, summarize relevant privacy policies, and highlight potential risks when sharing information with LLMs. We evaluated the usability and usefulness of CLEAR across two example domains: ChatGPT and the Gemini plugin in Gmail. Our findings demonstrated that CLEAR is easy to use and improves users’ understanding of data practices and privacy risks. We also discussed LLM’s duality in posing and mitigating privacy risks, offering design and policy implications.
- A Comprehensive Indoor Environment Dataset from Single-Family Houses in the USAnik, Sheik Murad Hassan; Gao, Xinghua; Meng, Na (MDPI, 2025-03-05)The paper describes a dataset comprising indoor environmental factors such as temperature, humidity, air quality, and noise levels. The data were collected from 10 sensing devices installed in various locations within three single-family houses in Virginia, USA. The objective of the data collection was to study the indoor environmental conditions of the houses over time. The data were collected at a frequency of one record per minute for a year, combining to a total over 2.5 million records. The paper provides actual floor plans with sensor placements to aid researchers and practitioners in creating reliable building performance models. The techniques used to collect and verify the data are also explained in the paper. The resulting dataset can be employed to enhance models for building energy consumption, occupant behavior, predictive maintenance, and other relevant purposes.
- Deep Learning Ensemble Approach for Predicting Expected and Confidence Levels of Signal Phase and Timing Information at Actuated Traffic SignalsEteifa, Seifeldeen; Shafik, Amr; Eldardiry, Hoda; Rakha, Hesham A. (MDPI, 2025-03-07)Predicting Signal Phase and Timing (SPaT) information and confidence levels is needed to enhance Green Light Optimal Speed Advisory (GLOSA) and/or Eco-Cooperative Adaptive Cruise Control (Eco-CACC) systems. This study proposes an architecture based on transformer encoders to improve prediction performance. This architecture is combined with different deep learning methods, including Multilayer Perceptrons (MLP), Long-Short-Term Memory neural networks (LSTM), and Convolutional Long-Short-Term Memory neural networks (CNNLSTM) to form an ensemble of predictors. The ensemble is used to make data-driven predictions of SPaT information obtained from traffic signal controllers for six different intersections along the Gallows Road corridor in Virginia. The study outlines three primary tasks. Task one is predicting whether a phase would change within 20 s. Task two is predicting the exact change time within 20 s. Task three is assigning a confidence level to that prediction. The experiments show that the proposed transformer-based architecture outperforms all the previously used deep learning methods for the first two prediction tasks. Specifically, for the first task, the transformer encoder model provides an average accuracy of 96%. For task two, the transformer encoder models provided an average mean absolute error (MAE) of 1.49 s, compared to 1.63 s for other models. Consensus between models is shown to be a good leading indicator of confidence in ensemble predictions. The ensemble predictions with the highest level of consensus are within one second of the true value for 90.2% of the time as opposed to those with the lowest confidence level, which are within one second for only 68.4% of the time.
- Impossibility of adversarial self-testing and secure samplingBansal, Akshay; Singh Arora, Atul; Van Himbeeck, Thomas; Sikora, Jamie (American Physical Society, 2024-08-21)Self-testing is the task where spatially separated Alice and Bob cooperate to deduce the inner workings of untrusted quantum devices by interacting with them in a classical manner. We examine the task above where Alice and Bob do not trust each other which we call adversarial self-testing.We show that adversarial self-testing implies secure sampling—a simpler task that we introduce where distrustful Alice and Bob wish to sample from a joint probability distribution with the guarantee that an honest party’s marginal is not biased. By extending impossibility results in two-party quantum cryptography, we give a simple proof that both of these tasks are impossible in all but trivial settings.
- Low responsiveness of machine learning models to critical or deteriorating health conditionsPias, Tanmoy Sarkar; Afrose, Sharmin; Tuli, Moon Das; Trisha, Ipsita Hamid; Deng, Xinwei; Nemeroff, Charles B.; Yao, Danfeng Daphne (Springer Nature, 2025-03-11)Background: Machine learning (ML) based mortality prediction models can be immensely useful in intensive care units. Such a model should generate warnings to alert physicians when a patient’s condition rapidly deteriorates, or their vitals are in highly abnormal ranges. Before clinical deployment, it is important to comprehensively assess a model’s ability to recognize critical patient conditions. Methods: We develop multiple medical ML testing approaches, including a gradient ascent method and neural activation map. We systematically assess these machine learning models’ ability to respond to serious medical conditions using additional test cases, some of which are time series. Guided by medical doctors, our evaluation involves multiple machine learning models, resampling techniques, and four datasets for two clinical prediction tasks. Results: We identify serious deficiencies in the models’ responsiveness, with the models being unable to recognize severely impaired medical conditions or rapidly deteriorating health. For in-hospital mortality prediction, the models tested using our synthesized cases fail to recognize 66% of the injuries. In some instances, the models fail to generate adequate mortality risk scores for all test cases. Our study identifies similar kinds of deficiencies in the responsiveness of 5-year breast and lung cancer prediction models. Conclusions: Using generated test cases, we find that statistical machine-learning models trained solely from patient data are grossly insufficient and have many dangerous blind spots. Most of the ML models tested fail to respond adequately to critically ill patients. How to incorporate medical knowledge into clinical machine learning models is an important future research direction.
- 'Do I Have to Take This Class?': A Review of Ethics Requirements in Computer Science CurriculaWeichert, James; Kim, Dayoung; Zhu, Qin; Eldardiry, Hoda (ACM, 2025-02-12)ABET criteria for accreditation of undergraduate computer science (CS) degrees require universities to cover within their curricula topics including “local and global impacts of computing solutions on individuals, organizations, and society,” and to prepare their students to “make informed judgments in computing practice, taking into account legal, ethical, diversity, equity, inclusion, and accessibility principles” [1]. A growing body of research similarly identifies the need for CS programs to integrate ethics into their degree requirements, both through standalone ethics-related courses and embedded modules or case studies on the ethical impacts in ‘technical’ courses. The calls for increased attention to CS ethics education have become more pressing with the emergence of sophisticated consumer-ready AI technologies, which pose new ethical challenges in the forms of bias, hallucination, and autonomous decision-making. Yet it remains unclear whether current university curricula are adequately preparing future graduates to confront these challenges. This paper presents a systematic review of the degree requirements of 250 computer science bachelor’s degree programs worldwide. We categorize each program according to whether a CS-related ethics course is offered and/or required by the department, finding that almost half of all universities we review do not offer any computing ethics courses, and only 33% of universities require students to take an ethics course to obtain their degree. We analyze differences among public US, private US, and non-US universities and discuss implications for curricular changes and the state of undergraduate computing ethics education.
- Diary Study as an Educational Tool: An Experience Report from an HCI CourseFan, Jixiang; Haqq, Derek; Saaty, Morva; Wang, Wei-Lu; McCrickard, D. Scott (ACM, 2025-02-12)With the rapid advancement and widespread adoption of computer technology, it has become an indispensable component in the development of human society. Therefore, computer science education’s focus extends beyond merely teaching students to read and write code; it is crucial to assist them in gaining an accurate and deep understanding of the applications of technology in the real world, its conveniences, and potential risks. Furthermore, it involves exploring how to design, improve, and innovate computer technologies to meet practical demands. Consequently, Human-Computer Interaction (HCI) has grown increasingly significant in the curriculum of computer science. However, research indicates that computing students face numerous challenges in learning HCI. To enhance students’ ability to experience, discover, and understand user needs, the authors of this paper recommend incorporating diary studies in HCI education. In the field of HCI, diary studies are a method for collecting long-term data on user behavior and experiences in a natural environment. Participants are required to record their daily activities, product usage, encountered issues, and personal impressions over specific periods. This paper will detail the process and steps implemented in our diary studies and present student feedback and evaluations. Through this experience report, we hope to encourage more educators to adopt and refine the diary study methodology in their courses, thereby aiding computer science students in better understanding and embracing the concepts and knowledge of HCI.
- The Impact of Group Discussion and Formation on Student Performance: An Experience Report in a Large CS1 CourseWu, Tong; Tang, Xiaohang; Wong, Sam; Chen, Xi; Shaffer, Clifford A.; Chen, Yan (ACM, 2025-02-12)Programming instructors often conduct collaborative learning activities, such as Peer Instruction (PI), to enhance student motivation, engagement, and learning gains. However, the impact of group discussion and formation mechanisms on student performance remains unclear. To investigate this, we conducted an 11- session experiment in a large, in-person CS1 course. We employed both random and expertise-balanced grouping methods to examine the efficacy of different group mechanisms and the impact of expert students’ presence on collaborative learning. Our observations revealed complex dynamics within the collaborative learning environment. Among 255 groups, 146 actively engaged in discussions, with 96 of these groups demonstrating improvement for poor-performing students. Interestingly, our analysis revealed that different grouping methods (expertise-balanced or random) did not significantly influence discussion engagement or poor-performing students’ improvement. In our deeper qualitative analysis, we found that struggling students often derived benefits from interactions with expert peers, but this positive effect was not consistent across all groups.We identified challenges that expert students face in peer instruction interactions, highlighting the complexity of leveraging expertise within group discussions.
- Understanding the Effects of Integrating Music Programming and Web Development in a Summer Camp for High School StudentsManesh, Daniel; Jelson, Andrew; Altland, Emily; Freeman, Jason; Lee, Sang Won (ACM, 2025-02-18)This poster presents the development and implementation of a 10- day remix-based summer camp curriculum designed to introduce high school students, particularly a multinational cohort of young women, to programming through creative coding. The curriculum integrates music composition using EarSketch and web development with HTML and CSS. The camp aims to inspire participants to gain self-efficacy in programming and motivate them to explore STEM/computing careers. Preliminary results from surveys and interviews indicate increased confidence in programming skills. This ongoing research explores the impact of remixing as a gateway for transitioning into more general-purpose computing domains such as web development.
- RT-BarnesHut: Accelerating Barnes–Hut Using Ray-Tracing HardwareNagarajan, Vani; Gangaraju, Rohan; Sundararajah, Kirshanthan; Pelenitsyn, Artem; Kulkarni, Milind (ACM, 2025-03)The 𝑛-body problem involves calculating the effect of bodies on each other. 𝑛-body simulations are ubiquitous in the fields of physics and astronomy and notoriously computationally expensive. The naïve algorithm for 𝑛-body simulations has the prohibiting 𝑂(𝑛2) time complexity. Reducing the time complexity to 𝑂(𝑛 · lg(𝑛)), the tree-based Barnes–Hut algorithm approximates the effect of bodies beyond a certain threshold distance. Other than algorithmic improvements, extensive research has gone into accelerating 𝑛-body simulations on GPUs and multi-core systems. However, Barnes– Hut is a tree-traversal algorithm, which makes it a poor target for acceleration using traditional GPU shader cores. In contrast, recent work shows that, for tree-based computations, GPU ray-tracing (RT) cores dominate shader cores. In this work, we reformulate the Barnes–Hut algorithm as a ray-tracing problem and implement it with NVIDIA OptiX. Our evaluation shows that the resulting system, RT-BarnesHut, outperforms current state-of-the-art GPU-based implementations.
- Making Software Development More Diverse and Inclusive: Key Themes, Challenges, and Future DirectionsHyrynsalmi, Sonja; Baltes, Sebastian; Brown, Chris; Prikladnicki, Rafael; Rodriguez-Perez, Gema; Serebrenik, Alexander; Simmonds, Jocelyn; Trinkenreich, Bianca; Wang, Yi; Liebel, Grischa (ACM, 2025)Introduction: Digital products increasingly reshape industries, influencing human behavior and decision-making. However, the software development teams developing these systems often lack diversity, which may lead to designs that overlook the needs, equal treatment or safety of diverse user groups. These risks highlight the need for fostering diversity and inclusion in software development to create safer, more equitable technology. Method: This research is based on insights from an academic meeting in June 2023 involving 23 software engineering researchers and practitioners. We used the collaborative discussion method 1-2-4-ALL as a systematic research approach and identified six themes around the theme ?challenges and opportunities to improve Software Developer Diversity and Inclusion(SDDI)'. We identified benefits, harms, and future research directions for the four main themes. Then, we discuss the remaining two themes, Artificial Intelligence&SDDI and AI&Computer Science education, which have a cross-cutting effect on the other themes. Results: This research explores the key challenges and research opportunities for promoting SDDI, providing a roadmap to guide both researchers and practitioners. We underline that research around SDDI requires a constant focus on maximizing benefits while minimizing harms, especially to vulnerable groups. As a research community, we must strike this balance in a responsible way.
- Reexamining Technological Support for Genealogy Research, Collaboration, and EducationShan, Fei; Luther, Kurt (ACM, 2025)Genealogy, the study of family history and lineage, has seen tremendous growth over the past decade, fueled by technological advances such as home DNA testing and mass digitization of historical records. However, HCI research on genealogy practices is nascent, with the most recent major studies predating this transformation. In this paper, we present a qualitative study of the current state of technological support for genealogy research, collaboration, and education. Through semi-structured interviews with 20 genealogists with diverse expertise, we report on current practices, challenges, and success stories around how genealogists conduct research, collaborate, and learn skills. We contrast the experiences of amateurs and experts, describe the emerging importance of standardization and professionalization of the field, and stress the critical role of computer systems in genealogy education. We bridge studies of sensemaking and information literacy through this empirical study on genealogy research practices, and conclude by discussing how genealogy presents a unique perspective through which to study collective sensemaking and education in online communities.