Ashe Neth

CV

About me:

I am an undergraduate computer science researcher at Worcester Polytechnic Institute (WPI, Class of 2025). I am currently looking for PhD positions for fall 2025. My primary research as of late surrounds creating high-performance platform-specific algorithms for machine learning and inference. My research interests include many forms of parallel computing such as high-performance computing, high-throughput computing, and multi-task computing, volunteer computing, resource-constrained computing, and distributed systems. My primary work so far has been on finding solutions to computing problems in resource-constrained environments and utilizing them in non-constrained environments. I aim to utilize the mindset that I have acquired while working with both resource-constrained and non-constrained systems to further the field of parallel computing.

Papers:

UnIT Pruner: Unstructured Inference-Time Pruning for Battery-Free Systems - Ashe Neth, Mohammad Nur Hossain Khan, Subrata Biswas, Asif Salekin, Bashima Islam

Preprint: link

In Review - The ACM Conference on Embedded Networked Sensor Systems (ACM SenSys 2025)

With deep neural networks (DNNs) being increasingly deployed on low-resource devices, achieving efficiency without sacrificing accuracy has become a critical challenge. Model pruning, commonly performed during training, is a common approach to reducing model size and computational cost, even in environments where computational resources are not limited. However, traditional pruning methods are inherently static, often leading to accuracy losses that hinder their applicability in dynamic, real-world settings. Although some inference-time pruning techniques exist, these methods are generally structured, introducing similar accuracy compromises as conventional pruning methods.In this paper, we propose a novel, unstructured pruning algorithm that performs adaptive, input-specific pruning during inference: Unstructured Inference-Time Pruner (UnIT Pruner). Unlike traditional approaches, UnIT Pruner dynamically skips redundant operations in real time based on the unique properties of each input, allowing for efficient computation without significant accuracy trade-offs. Our algorithm can complement existing methods, enhancing the balance between energy efficiency and accuracy and, in certain systems, even reducing latency. Experimental results show that our method reduces MAC operations by 70.38–87.39% with as low as 0.43% accuracy drop. Additionally, we achieve 74.4–94.7% less inference time and consume 74.2–96.5% less energy on microcontrollers when compared to other pruning methods. Our proposed algorithm achieves state-of-the-art inference time pruning performance, demonstrating its effectiveness in deploying DNNs in resource-constrained environments.

Pronoun Logic - Rose Bohrer and Ashe Neth

Queer in AI Workshop at The Nations of the Americas Chapter of the Association for Computational Linguistics (Queer in AI @ NAACL 2024)

arXiv: link

Historically, computational linguistics has been focused on rule-based semantics. This approach is still useful for parts of language that demand correctness such as with identity. This paper offers potential formalisms that would allow users to describe complex personal identities.

Particularly in transgender and nonbinary (TGNB) communities, it is an increasingly common practice to publicly share one's personal pronouns so that we may be gendered correctly in others' speech. Many of us have nuanced desires for how we are gendered, leading us to use more complex descriptions of our wishes; for example, the descriptor 'she/they'.

We observe that these descriptions of our wishes have the structure of a little language all their own. We thus propose formal logic as a tool for expressing one's personal pronouns and potentially other aspects of gender. We explore three potential logical foundations (linear logic, temporal logic, and free logic with definite descriptions) and their trade-offs.

Selected Projects:

CUDA Neural Network

Created a simple machine learning framework using CUDA.

Experimented with several novel optimizer algorithms.

Experimented with different ways of moving and storing memory on the GPU and its impact on processing speeds.

CUDA Ray Casting Engine

Created a raycasting library using my CUDA/OpenGL interoperability library for writing to the screen and CUDA for the internal logic.

Experimented with different ways of moving and storing memory on the GPU and its impact on processing speeds.

Library For CUDA/OpenGL Interoperability

Created a CUDA/OpenGL interoperability library for simplifying the process of exposing framebuffers to CUDA via OpenGL pixelbuffers.