CS4624: Multimedia, Hypertext, and Information Access
Permanent URI for this collection
This collection contains the final projects of the students in in the course Computer Science 4624: Multimedia, Hypertext, and Information Access, at Virginia Tech.
This course, taught by Professor Ed Fox, is part of the Human-Computer Interaction track, the Knowledge, Information, and Data track, and the Media/Creative Computing track. The curriculum introduces the architectures, concepts, data, hardware, methods, models, software, standards, structures, technologies, and issues involved with: networked multimedia (e.g., image, audio, video) information, access and systems; hypertext and hypermedia; electronic publishing; virtual reality. Coverage includes text processing, search, retrieval, browsing, time-based performance, synchronization, quality of service, video conferencing and authoring.
Browse
Browsing CS4624: Multimedia, Hypertext, and Information Access by Title
Now showing 1 - 20 of 294
Results Per Page
Sort Options
- 21st InventoryGarner, Elliot; Dean, Brandon; Mason, Brannon (2015-05-14)Currently, Network Infrastructure & Services (NI&S) takes inventory of equipment assigned to employees (computers, laptops, tablets, tools) and sends reports of higher value items to the Controller’s Office. All items have a VT tag number and a CNS number, which can currently only be matched up via an Oracle Forms interface. An inventory clerk must personally verify the existence and location of each piece of equipment. An improvement would be an app that scans an inventory number or bar code and the GPS location where it is scanned and the custodian of that equipment. This data could then be uploaded to a more accessible Google spreadsheet or similar web-based searchable table. The 21st Century Inventory app aims to solve this problem by employing barcode scanning technology integrated into a mobile app which would then send the accompanying asset ID to a CSV formatted output file. By directly tying a product’s asset ID to the user and their information, along with having the capability to scan a product’s barcode to simplify inventory lookup, saving product information to a CSV file, and giving the user the ability to edit the current information of a product in the application, we are providing a significant upgrade to a system that currently solely relies on an Oracle Forms interface.
- 3-Dimensional Weather VisualisationNimitz, Sarah; Forsyth, Duke; Knittle, Andrew (2016-05-04)Project deliverables are provided, including a detailed description of the creation of a polling and parsing system for keeping track of severe weather warnings, as delivered by the National Weather Service, and an interface to allow the user to view a representation of Doppler radar data in three dimensions. The report describes the roles of the team members, the work accomplished over the Spring 2016 semester, and the methods by which the team accomplished this work.
- 4624S14DSpaceEmbargoSchiefer, Jeb; Sharma, Paul (2014-05-07)DSpace [1] is an open source repository application used by many organizations and institutions. It provides a way to access and manage all kinds of digital documents. The 4624S14DSpaceEmbargo project was intended to extend the functionality of the ItemImport command line tool. Specifically the goal was to add the ability to embargo uploaded items until a specified date. This functionality was already implemented for the two web interfaces (XMLUI and JSPUI). DSpace is used by the Virginia Tech library in the form of VTechWorks [2]. The project was overseen initially by Keith Gilbertson and Zhiwu Xie who work for the Virginia Tech library. Near the end of the semester we were introduced to another software developer for the library, Jay Chen. We helped Jay set up the DSpace environment on his local computer and demonstrated to him how to use the ItemImport command line tool. Embargoes are used to limit access until a specified date. An embargo can be applied as a resource policy on an item, group, or bitstream level. An item level embargo restricts access to all of the files uploaded for a particular item. A group level embargoes submissions from anyone that is a member of the specified group. By default, the Anonymous group is the group that is used. A bitstream level embargo restricts access only on a specific file that is uploaded. The date format expected for setting an embargo must adhere to the ISO 8601 date format [3], specifically the YYYY-MM-DD, YYYY-MM, and YYYY variations. The deliverables for this project were the source and this documentation. Source code will be available on VTechWorks as well as GitHub. The GitHub repository [4] will be more up to date than the VTechWorks copy because we will continue some work on the project after the due date for this project based on feedback from the DSpace developers. The JIRA ticket for this feature to be implemented in DSpace 5.0 is DS-1996 [5]. [1] DuraSpace, “DSpace”, 2014, http://dspace.org/ [2] Virginia Tech, “VTechWorks”, 2014, http://vtechworks.lib.vt.edu/ [3] ISO, “Date and time format - ISO 8601,” 2014, http://www.iso.org/iso/home/standards/iso8601.htm [4] GitHub, “jebschiefer/DSpace,” 2014, https://github.com/jebschiefer/DSpace/ [5] DuraSpace JIRA, “[DS-1996] Embargo Support in ItemImport,” 2014, https://jira.duraspace.org/browse/DS-1996
- ABC Drone TeamBartal, Connor; Cooper, Jared (Virginia Tech, 2021-05-13)The ABC Sports Drone capstone team is an extension of the ABC Drone Project which is a group spearheaded by client Charles Kerr and in conjunction with the VT Club Ultimate team, Burn. The goal of the project as a whole is to provide high-quality footage and streaming of amateur sports to the masses. This capstone team is a subsection of the ABC Drone Project that has been tasked with creating software solutions and developing new techniques to help push this drone project to fruition. This report covers the progress of the capstone team in developing new routines for the drone, and the pivots that have been introduced as the team has received new data. The first goal that was tackled was identifying players on a field from an endzone-to-endzone view. This started with the analyzing of contours in addition to their position and attributes to determine if a contour was a player. Artifacts from off the field of play proved to be greatly troublesome, so a field bounding solution was created to eliminate as many artifacts as possible that were not on the field of play. Fairly good accuracy was achieved with this method (~75%), but the goal was set at 85%+ accuracy for identification. After experimenting with motion-detection and object persistence, the best course of action seemed to be identification via a convolutional neural network. No datasets were available that matched the application of this network, so an original dataset needed to be created. An application was developed that allowed for fairly quick extraction of data from sample videos. This data was fed to the neural network and constantly yields around 94% identification accuracy. Although the accuracy is high, it reduces frame rates to approximately 1 FPS. Some market interviews with actual coaches revealed a larger interest in post-processing capability than live-identification, so the client decided to pivot. A system that allows for speed-editing of footage has been developed, and a (proof of concept) companion application will allow coaches to easily track stats and pre-edit film via a GUI. The speed editing program takes in the footage and allows the coach to use a video game controller to create quick cuts to eliminate down time, as well as pan, tilt, and zoom on the footage to ensure the action is always framed. The edits are recorded in an edit-decision-list (EDL) file which is then sent in conjunction with the video file to Amazon Web Services. AWS takes the EDL file and original video and returns a fully-edited game film. With this method, a 90 minute game can be edited in 5 minutes or less. If coaches are recording stats during the game, the footage will also be annotated with important plays which are recorded on a similar EDL for gameplay statistics. Players will then have access to a program that will allow them to click their name to see the timestamps of all of their highlights.
- ADS Assessment VideoBeemsterboer, Christopher; Zebina, Tyler (Virginia Tech, 2019-05-19)In our Multimedia/Hypertext/Information Access capstone course, we worked with Adult Day Services to create a training video system to teach new instructors in their organization how to conduct recurring interviews with the adult clients. Adult Day Services is an organization at Virginia Tech that provides person-centered care to older adults who need assistance. Adult Day Services also aims to promote the physical, social, emotional, mental, and cognitive health of its participants, and they use a variety of assessments to measure overall well-being and participant progress. These assessments are conducted in the form of interviews, and the body language, tone, and speech of the interviewer are key to performing them successfully. The training video system we created covers five different types of assessments and is designed to efficiently train new instructors to conduct these interviews. We filmed an Adult Day Services instructor conducting interviews with five different participants, each completing the five assessments. We edited the footage and compiled all of the clips of each type of assessment together including transitions and titles. We later created a menu system which allows a user to select to play all of the training videos at once, or to play just the training video for a specific type of assessment. We have also included sub-categories within each type of assessment so the user can decide to view a specific participant as opposed to all. We delivered this project in the form of a Blu-ray .iso file on a USB drive which contains the menu system and the associated videos. We have also included instructions on how to download the VLC media player, which is the optimal software for viewing the contents on the .iso file. Finally, we have included our final presentation from our capstone course that goes over the final product as well as the lessons learned and our future plans.
- Adult Day Services Memory Masterclass Promotional VideoKulik, Maddie; Castillo, Pablo; Zurita, Jose (Virginia Tech, 2019-05-01)The goal of the project was to create a promotional video for Virginia Tech’s Adult Day Services center, specifically to advertise for their Memory Masterclass program. Adult Day Services is a center located within the Human Development and Family Sciences Department at Virginia Tech. They are licensed by the Department of Social Services to offer personal care, health monitoring, meals, therapeutic activities, dementia care, and recovery assistance. They serve typically 18 participants each operating day who average about 75 years of age. According to ADS’s mission statement, the center is dedicated to providing a center focused on the well-being and optimal functioning of its participants, a resource for caregiver support, an education opportunity for students, and a community among generations of children, college students, and adults. One of ADS’s main service offerings is their Memory Masterclass course. This course is offered in 6-week sessions to participants over 55 years of age who want to maximize their brain health. The focus of the course is to educate and serve people who have been diagnosed with Mild Cognitive Impairment (MCI). MCI is not a symptom or precursor to Alzheimer's or dementia, but rather a condition that occurs as aging changes brain function. In the 6-week course participants learn strategies for application to daily life that can strengthen brain reserve as you age and get connected with others who have similar concerns about memory. Our main objective was to create a promotional video that Adult Day Services could use on their website to inform and attract people to take the class. This project was broken up into several different stages. The first stage was to meet with our clients, Adult Day Service professionals, to gain a better understanding of the project requirements. Our clients described to us that they would like a video that showcased the active, healthy lifestyle of one of their Memory Masterclass participants. This would include footage of men and women doing outdoor activities, participating in class, and doing mentally-stimulating activities. From meeting with our clients, we came to realize that they wanted a specific type of aesthetic to their video - a combination of active and “homey” footage. An important goal for our clients was to have the video ready to be presented at an AARP event in mid-March, so the first stage of this project had to be completed by that deadline. The second stage was scheduling time to physically shoot the videos. This involved renting camera and sound equipment, coordinating with our clients and course participants, deciding on filming locations, and collecting the raw footage. Once we had shot all of the raw footage, the third stage was comprised of condensing, cleaning, and enhancing the raw footage to create a preliminary draft of the video. The video was delivered to the client, we received feedback, and have begun work revising the video to meet client specifications. The client will be able to use this video for advertising on the ADS website, as well as at different events where their services are promoted. The fourth stage of this project is what we are currently working on right now. Another recommendation was that we prepare another video that was a bit shorter, approximately 90 seconds long, that could be used as a shorter promotion. This shorter video will likely be a condensed version of highlights from the 4-minute video. The third stage was to revise the initial version of the video based on client feedback. This involved sitting down with our client and gaining specific insight as to what details they liked and what they wanted to have modified. After we acquired feedback, we were able to reshoot footage that was not preferable and take more shots of outdoor activities. The final version of the video incorporated footage from both stages of filming and incorporated the client's desired changes. This version of the video was also shown to an applicable user pool of Memory Masterclass students who gave us further feedback.
- AgInsuranceLLMsShi, Michael; Rajesh, Saketh; Truong, An; Hilgenberg, Kyle (2024-05-09)Our project is to develop a conversational assistant to aid users in understanding and choosing appropriate agricultural insurance policies. The assistant leverages a Large Language Model (LLM) trained on datasets from the Rainfall Index Insurance Standards Handbook and USDA site information. It is designed to provide clear, easily understood explanations and guidance, helping users navigate their insurance options. The project encompasses the development of an accessible chat interface, backend integration with a Flask API, and the deployment of the assistant on Virginia Tech's Endeavour cluster. Through personalized recommendations and visualizations, the assistant empowers users to make well-informed decisions regarding their insurance needs. Our project report and presentation outline the project's objectives, design, implementation, and lessons learned, highlighting the potential impact of this interactive conversational assistant in simplifying the complex process of selecting agricultural insurance policies.
- AI Aided AnnotationBishop, Jonah B. M.; David, Isaac; Lubana, Ishaandeep (Virginia Tech, 2022-05-11)Human annotation of long documents is a very important task in training and evaluation in NLP. The process generally starts with the human annotators reading over the document in its entirety. Once the annotator feels they have a sufficient grasp on the document, they can begin to annotate it. Specifically, annotators will look for questions that can be answered, and then write down the question and answer. In our client’s case, the chosen long documents are electronic theses and dissertations (ETDs) which are often 100-150 pages minimum, thereby making it a time consuming and expensive process to annotate. The ETDs are annotated on a chapter by chapter basis as content can vary significantly in each chapter. The annotations generated are then used to help evaluate downstream tasks such as summarization, topic modeling, and question answering. The system aids the annotators in the creation of a Knowledge Base that is rich with topics/keywords and question-answer pairs for each chapter in ETDs. The core of the system revolves around an algorithm known as the Maximal Marginal Relevance. By utilizing the MMR algorithm with a changeable lambda value, keywords, and a couple of other elements, we can identify sentences based on their similarity or diversity relative to a collection of sentences. This algorithm would greatly enhance the annotation process in ETDs by automating the process of identifying the most relevant sentences. Thus, annotators do not have to sift through the ETDs one sentence at a time, instead making a comprehensive summary as fast as the MMR algorithm can work. As a result, annotators can save many hours per ETD, resulting in more human generated annotations in a shorter amount of time. The final deliverables are the project, a final slideshow presenting our work throughout the semester, a final report, and a video demonstrating exactly how to use our platform. All of this is available here on VTechWorks in this report. Additionally, the project is being built using GitHub, making it free and available to the public to fork and modify in any way they see fit.
- AI-Assisted Annotation of Medical ImagesDewan, Suha; Zhou, Daodao; Huynh, Long; Guo, Zipeng (Virginia Tech, 2022-12-15)In digital image processing and computer vision, image segmentation is the process of partitioning a digital image into multiple image segments. More precisely, it’s the process of assigning a label to every pixel in an image so that pixels with the same label share certain characteristics. Image segmentation is an important step in almost any medical image study. Segments are used in images from microscopes that show us different types of cells and these cells contain hundreds of organelles and macromolecular assemblies. Cell Segmentation is the task of splitting a microscopic image domain into lots of different segments, which represent individual instances of cells, however, this requires enormous time for domain experts to label manually and thus the need for AI-Assisted annotation of medical Images. Our project will aid the annotators in receiving images quickly and easily through our web application and performing the predictions on these images.
- Airbnb ScrapingYu, Wang; Huang, Baokun; Liu, Han; Pham, Vinh; Nikolov, Alexander (Virginia Tech, 2020-05-13)Inside Airbnb is a project by Murray Cox, a digital storyteller, who visualized Airbnb data that was scraped by author and coder Tom Slee. The website offers scraped Airbnb data for select cities around the world; historically data is also available. We were tasked with creating visualizations with listing data over Virginia and Austria to see what impact Airbnb was having on the communities in each respective region. The choice was Virginia and Austria because our team was familiar with both regions, with parts of our team being familiar with Virginia and other parts being familiar with Austria. The eventual goal is to expand past analysis of these 2 regions and expand further to say the rest of the United States. Since July 2019, Tom Slee has abandoned the script2 to collect data. To collect data on Virginia and Austria, we needed to update the script to collect more recent data. We began inspecting the script and found it was not collecting as much data as it once was. This was almost certainly due to Airbnb’s website layout changing over time (a common nature of websites). After finding out how the script worked, we eventually found out the various problems related to the script and updated it to the new Airbnb website design. Doing so, we were able to get even more data than we thought possible such as calendar and review data. From there, we were able to begin our data collection process. During all the time fixing the script, our team was making mock visualizations to be displayed on a website for easy viewability. Once data collection was complete, the data was transferred over to be used for these mock visualizations. We visualized many things such as how many listings a single host had, how many listings were in a given county, etc. The main visualization created was to see where all the listings for Airbnb were on the map. We displayed this on a map. We also made maps to visualize availability, prices, and the number of reviews. Further, we created pie charts and histograms to represent Superhosts, instantly bookable listings, and price distributions. We expect that in the future the script and the data collected and visualized will be used by both future CS Students working on subsequent iterations of the project as well as Dr. Zach himself, our client.
- Analyzing Microblog Feeds to Trade StocksWatts, Joseph; Anderson, Nick; Asbill, Connor; Mehr, Joseph (Virginia Tech, 2017-05-10)The goal of this project is to leverage microblogging data about the stock market to predict price trends and execute trades based on these predictions. Predicting the price trends of stocks with microblogging data involves a complex opinion aggregation model. For this, we built upon previous research, specifically a paper called "CrowdIQ" submitted by a team consisting of some Virginia Tech faculty. This paper details a complicated method of aggregating an accurate opinion by modeling judge reliability and interdependence. Once the overall sentiment of the judges was deduced, we built trading strategies that take this information into account to execute trades. The first step of the project was a sentiment analysis of posts on a microblogging site named StockTwits. These messages can contain a label indicating a bullish or bearish sentiment, which will help indicate a specific position to take on a given stock. However, most users choose not to use these labels on their StockTwits. A classification of these unlabeled tweets is required to autonomously utilize StockTwits to drive the proposed trading strategies. With a working sentiment analysis model, we implemented the opinion aggregation model described by CrowdIQ. This can gauge an accurate market sentiment for a particular stock based on the collection of sentiments that are received from users on StockTwits. The next step was the creation of a trading simulation platform, including a complete virtual portfolio management system and an API for retrieving historical and current stock data. These tools allow us to run quick and repeatable tests of our trading strategies on historical data. We can easily compare the performance of strategies by running them with the same historical data. After we had a viable testing environment setup, we implemented trading strategies. This required research and analysis of other attempts at similar uses of microblogging data on predicting stock returns. The testing environment was focused on a set of stocks that is consistent with those used in CrowdIQ. The implementation of the CrowdIQ strategy served as a baseline against which we compared our results. Development of new trading strategies is an open-ended task that involved a process of trial and error. It is possible for a strategy to find success in 2014, but not perform quite as well in other years, because market climates can be fickle. To assess the dependence of the market climate on our strategy's success, we also tested against data for the year of 2015 and compared the performance. The final deliverable is a viable trading simulation environment coupled with various trading strategies and an analysis of their performance in the years of 2014 and 2015. The analysis of each strategy's performance indicated that our sentiment-based strategies perform better than the index in bullish markets like that of 2014, but, when they encounter a bear market, they typically make poor trading decisions which result in a loss of value.
- AnimalRescueSchneier, Neal; Highman, Zachary (2012-05-04)
- Anti-Poaching Drone ControlLyman, Matthew; Hudson, Matthew; Bishop, Cory (Virginia Tech, 2022-05-11)Our project assists the SeaQL Lab of Virginia Tech's Department of Fisheries and Wildlife Conservation. Working with the Marine Management Organisation of the UK, the Lab's project entails developing an autonomous drone swarm that can fly predetermined routes around the Chagos Archipelago and send alerts about potential poaching boats, based on machine learning image analysis in the drones' attached computing modules. The main goal of this project is to save the sharks and the ecosystem of those waters while decreasing the time, money, and effort for the local Coast Guard to perform regular monitoring. Instead, the drones will send detection alerts to a remote server being monitored by a ranger if it spots a potential poaching boat. Our report details our contributions to the overall project. Our team took responsibility for several smaller tasks integral to the overall project. First, we familiarized ourselves with the Robotic Operating System (ROS) to connect, calibrate, test, and record video using the cameras provided. ROS will control much of the drones' added functionality such as running the poaching boat detection algorithm, sending flight commands to the drones, and streaming video over a cellular connection. Next, we aided the larger project team in repairing one off-the-shelf drone for potential flight testing. After unsuccessful troubleshooting, we moved to help finish construction of the primary hexacopter. Finally, we wrote a script to start the 4G cellular connection automatically when a drone is powered on. The AntiPoachingDroneControlReport details this work amidst the larger project goals of the SeaQL Lab. The AntiPoachingDroneControlPresentation gives a brief summary of our project work and the lessons learned. This was presented to our CS4624: Multimedia, Hypertext, and Information Access class to summarize our project work and experiences.
- Apple Ridge Farms Corporate Retreat VideoVernon, Tyler; Dallachie, Charles; Mykich, Andrew; Duval, Matthew (2012-05-03)Apple Ridge Farms, a NPO in the Roanoke area, sponsors an academic summer camp for underprivileged youths in the Roanoke area. They also host corporate retreats and other events on their grounds in the off-season. They requested a short video for internet distribution to increase revenue from their corporate retreat portion of their business. We filmed the grounds on April 27th, 2012, and created a video for them using captured video, images, and audio, as well as images they provided.
- AppTrackWildlifeDiseasesJi, Shangzheng; Lyu, Jiarui; Vu, Justin; Zhang, Tenghui (Virginia Tech, 2021-05-07)Our project is to design a smartphone application and a website to report mange and other wildlife diseases in realtime. Our free smartphone app is designed for both professionals (e.g. hunters) and non-professionals. Our app provides a mini questionnaire to collect the users' familiarity with the mange, take photos of the wildlife species and potential disease, and get the geolocation and date of the photo. Then, all information collected will be saved to the firebase and used by the website. Our website will summarize the data and images collected and display them on the map. We submit the PDF and the PowerPoint of our final presentation. Our final presentation starts from project Introduction, then to the project design, timeline, work completed, iOS application, website, testing, future works, lessons learned, acknowledgment, and references. We also submit the PDF and the zip project dump from Overleaf of our final report. Our final report covers Executive Summary /Abstract, Introduction, Requirements, Design, Implementation, Testing/Evaluation/Assessment, Users' Manual, Developer’s Manual, Lessons Learned, and Acknowledgements.
- Arabic News Article SummarizationAyoub, Souleiman; Freeman, Julia (2015-05-14)This project involves taking Arabic PDF news articles to produce results from our new program that indexes, categorizes, and summarizes them. We fill out a template to summarize news articles with predetermined attributes. These values will be extracted using a named entity recognizer (NER) which will recognize organizations and people, topic generation using an LDA algorithm, and direct information extraction from news articles’ authors and dates. We use Fusion LucidWorks (a Solr based system) to help with the indexing of our data and provide an interface for the user to search and browse the articles with their summaries. Solr is used for information retrieval. The final program should enable end users to sift through news articles quickly.
- Artificial Immune System (AIS) Based Intrusion Detection System (IDS) for Smart Grid Advanced Metering Infrastructure (AMI) NetworksSong, Kevin; Kim, Paul; Tyagi, Vedant; Rajasekaran, Shivani (Virginia Tech, 2018-05-09)The Smart Grid is a large system consisting of many components that contribute to the bidirectional exchange of power. The reason for it being “smart” is because vast amounts of data are transferred between the meter components and the control systems which manage the data. The scale of the smart grid is too large to micromanage. That is why smart grids must learn to use Artificial Intelligence (AI) to be resilient and self-healing against cyber-attacks that occur on a daily basis. Unlike traditional cyber defense methods, Artificial Immune System (AIS) principles have an advantage because they can detect attacks from inside the network and stop them before they occur. The goal of the report is to provide a proof of concept that an AIS can be implemented on smart grid AMI (Advanced Metering Infrastructure) networks and furthermore be able to detect intrusions and anomalies in the network data. The report describes a proof of concept implementation of an AIS system for intrusion detection with a synthetic packet capture (pcap) dataset containing common Internet protocols used in Smart grid AMI networks. An intention of the report is to provide the necessary background for understanding the implementation in the later sections. The background section defines what a smart grid is and how its Advanced Metering Infrastructure (AMI) works, describing all three networks the AMI consists of. The Wide Area Network (WAN) is one of the three networks and we were scoping down to WAN for our project. The report goes on to discuss the current cyber threats as well as defense solutions related to the smart grid network infrastructure today. One of the most widely used defense mechanisms is the Intrusion Detection System (IDS), which has many important techniques that can be used in the AIS based IDS implementation of this report. The most commonly used AIS algorithms are defined. Specifically, the Negative Selection Algorithm (NSA) is used for our implementation. The NSA algorithm components used in the implementation section are thoroughly explained and the AIS based IDS framework is defined. A list of AIS usages/values in enterprise networks is presented as well as research on current NSA use in AIS implementations. The latter portion of the report consists of the design and implementation. Due to data constraints and various other limitations, the team wasn’t able to complete the initial implementation successfully. Therefore, a second implementation design was created, leading to the main implementation which meets the project’s objective. The implementation employs a proof of concept approach using a C# console application which performs all steps of an AIS on user created network data. In conclusion, the second implementation has the ability to detect intrusions in a synthetic dataset of “man-made” network data. This proves the AIS algorithm works and furthers the understanding that if the implementation was scaled up and used on real-time WAN network data it would run successfully and prevent attacks. The report also documents the limitations and problems one can run into when attempting to implement a solution of this scale. The ending sections of the report consists of the Requirements, Assessment, Assumptions, Results, and lessons learned followed by the Acknowledgments to MITRE Corporation which helped immensely throughout the development of the report.
- Assistive Voice AssistantSatnur, Abishek Ajai; Bruner, Charles (2024-05-09)This project is an extension of work that has been done in previous years on the sharkPulse website. sharkPulse was created due to the escalating exploitation of shark species and the difficulty of classifying shark sightings. Due to sharks’ low population dynamics, exploitation has only exacerbated the issue and made sharks the most endangered group of marine animals. sharkPulse retrieves sightings from several sources such as Flickr, Instagram, and user submissions to generate shark population data. The website utilizes WordPress , HTML, and CSS for the front end and R-Shiny, PostgreSQL, and PHP to connect the website to the back end database. The team was tasked with improving the general usability of the site by integrating dynamic data-informed visualizations. The major clients of the project are Assistant Professor Franceso Ferreti from the Virginia Tech Department of Fish and Wildlife Conservation and Graduate Research Assistant Jeremy Jenrette. The team established regular contact through Slack, scheduled weekly meetings online with both clients, and acquired access to all major code repositories and relevant databases. The team was tasked with creating dynamic and data-informed visualizations, general UI/UX improvements, and stretch goals for improving miscellaneous pages throughout the site. The team developed PHP scripts to model a variety of statistics by dynamically querying the database. These scripts were then sourced directly through the site via the Elementor WordPress module. All original requirements from the clients have been met as well as some stretch goals established later in the semester. The team created a Leaflet global network map of affiliate links which dynamically sourced the sharkPulse social network groups from an Excel spreadsheet and generated country border markers and links to each country’s social network sites as well as a Taxonomic Accuracy Table for the Shark Detector AI. The team created and distributed a survey form to collect user feedback on the general usability of the site which was compiled and sent to the client for future work.
- ATinstagramJeshong, Tashi; Joseph, Zubin; Barden, Mason; Halstead, Nicholas; Cho, Steve (Virginia Tech, 2022-05-09)For this project, we wanted to discover if and how hikers use the social media platform, Instagram, to talk about Leave No Trace (LNT) principles on the Appalachian Trail. Leave No Trace principles refer to a set of guidelines that hikers should follow in order to promote conservation on trails. The workflow to complete the project included: collecting relevant Instagram posts, performing sentiment analysis on these posts, and finally creating a series of graphs that show the different connections between posts. We started by utilizing Python, JSON objects, and Selenium to gather all of the Instagram posts with specific hashtags, such as “#AppalachianTrail” , “LeaveNoTrace”, and “LNT”. Selenium is used for the API calls, which retrieve the many Instagram posts. Information about each post, such as its geographic location, caption, and hashtag are extracted using JSON objects. The final two parts of the project include performing sentiment analysis on the collected posts and then visualizing the data in a variety of ways. For the sentiment analysis, we analyzed each caption of every post, and assigned it a score ranging from negative one to positive one. Negative one would represent a highly negative sentiment and positive one represents a highly positive sentiment. From there, we utilized the K-Means Clustering algorithm to gather posts with similar hashtags. For the visualizations, we displayed what tags occur in the same post, connections between different hashtags, and the geolocations of the different posts. The deliverables of our project include the source code that is used to scrape the Instagram posts, perform sentiment analysis, and visualize the data, along with several folders showing the results of our data collection. These results include the scraped Instagram posts, the sentiment analysis results, and the visualizations we created. These deliverables could help our client and those interested with research relating to Instagram, Leave No Trace principles, and the Appalachian Trail.
- Authoritative VenuesYoussef, Ali; Marku, Bella; Spicer, Tanner; Forst, Kyle (Virginia Tech, 2021)This submission details the progress made on the Authoritative Venues project. The goal of the Authoritative Venues project was to use machine learning algorithms to create a web application that can accurately recommend fitting ACM-related venues for Computer Science researchers trying to publish their work. By providing a ranked output list of publication venues related to a paper’s topic, we help researchers make more informed decisions about where to submit their work for publication. Additionally, we provide insight into the data collection, virtual machine setup, and website hosting process that allowed for this project to be easily accessible by anyone. This project is particularly useful for CS researchers wanting to gain insight into which ACM-related publication venue would best fit their paper. The recommender is hosted at authvenue.cs.vt.edu. On this website, there are two input fields that researchers can use to provide the title and abstract of their paper. Once this is inputted, researchers can submit this information and receive recommendations specifically catered to their work.