Penny King
projects contact
×
projects contact

Efficient Common Sense in LLMs via Knowledge Graph Compression

This project tackles a fundamental challenge in AI: teaching machines common sense. Specifically we explored how Large Language Models (LLM) deal with common sense and intuition. Commonsense typically comes easily to humans. While humans naturally understand that "dropping a glass cup will break it," machines struggle with this type of intuitive knowledge. To help with this, data scientists use knowledge graphs, which are structured graph networks of information that connect concepts to each other along with their relationships. But there's one “big” problem: these graphs are REALLY BIG. Feeding all this information into LLMs can overwhelm them, leading to confusion and poor performance. This mirrors much like how humans also get overwhelm with a huge overflow of information. So, we utilize Graph Transformers to compress these knowledge graphs to be smaller and effective at teaching common sense to LLMs.

Comparing Graph Neural Networks Against Transformer Based Graph Learning

This project explores the use of transformer-based graph learning models, specifically GraphGPS, and benchmarks their performance against state-of-the-art graph neural networks (GNNs) such as GCN, GIN, and GAT across diverse datasets like CORA, IMDB-BINARY, ENZYME, and PEPTIDES-FUNC. The study highlights the strengths and limitations of these architectures, focusing on their ability to capture local and global dependencies. While simpler GNNs excel in less complex tasks, GraphGPS demonstrates its advantage in handling long-range interactions, particularly in datasets requiring such dependencies, offering insights into advancing graph-based machine learning applications.

English-to-Japanese Translation Deep Learning Model

This project explores and compares the performance of four Seq2Seq architectures—CNNs, GRUs, LSTMs, and Transformers—on an English-to-Japanese translation task using the Japanese-English Subtitle Corpus (JESC). The models were evaluated based on their translation accuracy, with CNNs outperforming the others due to their ability to converge faster and effectively capture local connections in short sequences. The study highlights CNNs' suitability for handling Japanese's shorter sentence structures and suggests future directions, including testing on longer documents and other language pairs. The code for this project is available on GitHub.

Winning Hackathon Project for Image Anomaly Detection

This project won UCSD's 2024 DataHacks hackathon, which was judged by data science professors and researchers. In this project I collaborated with two other data science students, developing an autoencoder based neural network model. Ultimately this project is an image based anomaly detection. In the project team, I acted as the project team lead, designing and directing data collection and methodology, while also writing code to build and test the deep learning architectures.

Restaurant Rating Prediction and Recommender System

  • Project Link
  • Python project (sklearn, seaborn, numpy, matplotlib, pandas, tensorflow)

In this project I trialed multiple recommender system modeling methods for a prediction problem using a Google restaurant review dataset. Some of the methods involved techniques such as utilizing latent factors, while others involved no latent factors. In addition to testing multiple machine learning models, I also performed exploratory data analysis on the dataset, feature engineering, and explored past literature.

Features used to train the model consisted of natural language processing (TF-IDF) to figure out what words were associated with positive or negative reviews, as well as figuring out how to deal with the sparsity of the data we were using.

Travel Analysis

  • Project Link
  • Python project (pandas, sklearn, seaborn, statsmodels, numpy)

In this group project, I explored international traveler behavior coming into the US to analyze if the strength of the USD and the appearance of natural disasters and federal emergencies had a significant impact on tourism into the United States.

We tested multiple regression models in our analysis and looked into why some models may have performed better than others in the context of our data. In this project, I took a lead role in outlining key deadlines, assigning work, and the project design. I also wrote code for exploratory data analysis (EDA) and training and trialing the models.

Power Outage Model

This project focuses on predicting the cause of power outages. Sometimes power outages may occur due to unexpected reasons and the cause for it would be unclear.

This project consisted of building a decision tree-based classifier using information that is available shortly after a power outage. Aside from model testing and feature engineering, this project also consisted of EDA, looking at past research, and studying the fairness of our final model through hypothesis testing.

Breast Cancer Tissue Classification

  • Project Link
  • Python project (pandas, statsmodels, numpy, matplotlib, seaborn)

This project aimed to predict the malignancy of breast tumors using various information about the tumor’s cell nuclei’s shape and appearance using a dataset of processed tumor images. Being able to accurately predict a tumor cell’s malignancy using its nucleus would be a very important tool to help cancer patients go forward with treatment before they reach a more critical stage of cancer.

In our project, we aimed to classify breast tumor mass cells as either malignant or benign using a logistic regression classification model based on data about the tissue’s nuclei measurements. We used k-folds cross-validation for hyperparameter tuning, with stratified random sampling. In addition to studying past research, this project also consisted of EDA, feature engineering, hyperparameter selection, and model testing.