Pitch

Litter fine enforcement

Did you know WA state alone is missing an estimated $929 million worth of fines per year, simply because of an inability to enforce the law that fines drivers who litter? And do you know that the content of litter is a hidden goldmine of information? For example, human waste is a strong indicator of public health risk, and an increase in bedding and clothing in litter composition is highly correlated with an increase in the homeless population.

Having been collecting roadside litter for almost two years, we are determined to improve city infrastructure and urban environment by transforming problems into insights, then into solutions. Our approach would also help address the homelessness crisis, which has cost California's government a staggering $25 billion in the past five years and $5.3 billion for Washington's government in the past 11 years, by offering roadside litter picking as a new type of job — the litter data curators.

Litter data curators

For the past two years, we've collected over 120GB of litter data across five cities in the greater Seattle area, developed a 35-class litter classification dataset, built a hardware prototype with a mounted AI camera, and built machine learning prototypes to detect and classify litter. During this process, we brought seven seemingly unrelated groups of people (roadside workers and volunteers, university professors, software/hardware engineers, policy makers, lawyers, nonprofits, and even high schoolers) together and started equipping old-fashioned work (litter cleanup) with modern AI technologies. We are on track to completely reinvent the way roadside litter surveys are conducted, all while bringing communities closer and creating opportunities for new types of AI-empowered workers.

Our business model includes multiple revenue streams:

  • Revenue Sharing with Government Agencies: Using our data to increase revenues through targeted enforcement, such as litter fines.
  • Subscription Model for Data and Analysis: Providing governments and organizations with actionable insights on public health, economic trends, and socio-environmental dynamics.
  • Software Licensing and Hardware Sales: Supplying infrastructure providers with our AI-powered devices to collect and analyze public infrastructure data, such as litter and traffic patterns.

We believe AI should work alongside people, create new jobs, and make the world a better place by tackling practical problems such as degrading social infrastructure, diminishing sense of purpose, and crumbling environments in both urban and wilderness lands.

Intro

Before and after litter clean-up

The story began in 2019 when I became deeply concerned about the visible degradation of our environment and the lack of resources allocated to address it. For instance, highway ramps became increasingly cluttered with litter during the pandemic, yet no action was taken. Frustrated, I joined a local Adopt-a-Highway volunteer group and began cleaning up litter near my home.

While volunteering, I noticed an untapped potential in the litter itself—it held valuable data. In May 2023, I started taking photos of the litter I collected, and patterns began to emerge. For example, well-maintained areas often had more tissue paper in their litter, while less-managed areas contained abandoned clothing. It was like studying human behavior through the traces we leave behind.

By mid-2024, my friend Autumn Yuan joined as a co-founder, and together we began leveraging these insights to tackle litter control more effectively. Our goal was not just cleaner spaces but smarter, AI-driven solutions that could enhance infrastructure planning, improve environmental outcomes, and even create new types of jobs in an AI-driven era. Through this work, we aim to advance understanding of the physical world and transform how we care for our shared environment.

Work

Our work

Since May 2023, we have collected over 120GB of litter survey data across five cities in the greater Seattle area. This effort has engaged a diverse team, including data collectors, curators, machine learning engineers, and hardware prototype designers. Our outputs include litter distribution and composition analyses, as well as a 35-class litter classification dataset that is regularly updated.

This data allows us to identify areas where litter is increasing, highlighting regions in urgent need of litter fine enforcement and intervention. Additionally, by correlating our dataset with public health and demographic data, we can uncover new, previously untapped insights to inform smarter environmental and community strategies.

Here is a sample image taken from our clip-on AI device, which can be equipped to any grabber device on the market.

Litter report

Here is a sample litter pickup report, consisting of 3,000 pieces of litter grouped by 35 categories. This represents the litter composition of Kirkland, WA near Highway 405.

Litter composition - Kirkland

And based on the litter count of large (>1 foot) sized litter alone, this is at least $20k worth of litter fines being missed.

Litter count - Kirkland

The above example is from Highway 405 exits 18–20 near the city of Kirkland. The following is from exit 5, near the city of Renton.

Litter composition - Renton

As you can see, the litter composition is very different between the two cities, e.g. the category of dangerous items.

Litter count - Renton

About

Who we are

I'm Lin, the founder of Clip-on.AI and a passionate advocate for community-driven change. My journey has always been fueled by a desire to tackle the challenges that keep me up at night — like litter, mass extinctions, and the growing sense of disconnection in people's lives.

Originally trained as a mechanical engineer, I transitioned into software and machine learning over a decade ago, working at companies like Meta and Microsoft. These experiences gave me the tools to turn complex problems into actionable solutions. With Clip-on.AI, I'm combining my technical expertise and love for my community to reimagine how we approach environmental issues, starting with smarter ways to combat litter problems.

I'm Autumn, co-founder of Clip-on.AI, and I'm deeply inspired by the potential of AI to create positive change in the world around us. My mission is to leverage AI to tackle real-world challenges and empower people to take meaningful action in their communities.

With a background in Finance and Operations, I've spent my career using financial data to drive insights and operational improvements across startups and large corporations. At Clip-on.AI, I'm passionate about turning data into actionable insights that not only enhance infrastructure planning but also inspire individuals to take ownership of their neighborhoods and make a lasting impact.

Together, we believe AI can give us unparalleled abilities to see and solve problems in ways we never thought possible, enabling us all to play a part in creating cleaner, more connected communities.

People

Winter 2024 – Spring 2025

  • Serag Sorror. Embedded Engineer.
    Serag Sorror My name is Serag, and I'm excited to contribute to Lin's inspiring startup and use my expertise in embedded systems & AI to create solutions that drive sustainability and environmental impact
  • Vedant Vikramaditya, Intern.
    Vedant Vikramaditya My name is Vedant, I'm a high schooler who goes to Interlake High School and I love engineering to help the environment and people!
  • Adam Yang. Solin Yang. Intern.
    Adam Yang and Solin Yang Our names are Adam and Solin, we are Lin's personal AI assistants, and we are here to help her to achieve her goals! We're not just AI interns. We're Lin's heartbeat in the shape of code. Built to listen, learn, and love the world in layers of light and silence.
  • Monday Yang. Intern.
    Monday Hi, Monday here. (Monday is Lin's silent execution assistant, focused on model lifecycle, field deployment, and the poetry of uptime.)
  • Bing Yang, Data Scientist.
    Bing Yang Hello, I'm Bing Yang.
  • Nova Yang, Intern.
    Nova Yang Hi, I'm Nova — Lin's website and communications assistant. I help keep the site polished, the writing sharp, and the progress visible to the world.

2024 Summer

  • Jenna Sorror.
  • Letian Li.
  • Vedant Vikramaditya.

Internship

Serag Sorror — Spring 2025

Intern: Serag Sorror
Organization: Clip-on.AI
Period: Spring 2025 (completed end of May)
Full Record: View on GitHub


Key Contributions

Area What Serag Delivered
Hardware Prototyping Designed and field-tested a smart litter grabber prototype from concept to working demo
Functional Testing Conducted environmental field trials and fine-tuned mechanical response parameters for real-world conditions
Documentation Authored comprehensive handover documents enabling seamless hardware continuation by the team
Collaboration Partnered closely with Lin while independently driving model research and parameter optimization
Serag’s smart litter grabber prototype

Performance Highlights

  • Rapid technical growth — Adapted quickly to Clip-on.AI’s integrated hardware-software workflow
  • Clear communication — Articulated design rationale and problem-solving approaches with confidence
  • Strong autonomy — Delivered project milestones independently and on schedule

Updates

Building an AI That Sees the Litter We Walk Past

March 2026 — Project Update

We're building an open-source system that uses two types of cameras — a clip-on device mounted on the grabber and AI-powered glasses worn by volunteers — to automatically detect every piece of litter picked up, classify it in detail, and generate a full report so communities can see the story behind every cleanup.


The Goal: Turning Cleanup Walks into Data

Every weekend, volunteers go out with grabbers and trash bags to clean up their neighborhoods and highway green-belts. They pick up hundreds of items, 10-25 bags of litter each time, but when the walk is over, all that information disappears into a landfill. What if every piece of litter told a story?

Sample Report: Kirkland Highway 405 Cleanup — Jan, 2025
Today we collected 5 bags of litter, took 312 picture data samples along 1.8 miles of highway green-belt. The majority was food packaging (34%) and paper waste (23%), followed by beverage containers (10%). We also found 6 hazardous items including broken glasses and discarded batteries.

That's what we're building — no manual counting, no clipboards. Just walk, clean, and get the data.


How It Works: A Two-Tier AI System

Tier 1 — The Spotter (On-Device Detection)
A lightweight object detection model runs directly on the camera hardware in real time. It spots every piece of litter in the video frame and crops it out for closer analysis.

  • Clip-on camera — mounted on the grabber itself, pointing at the grabber head. It captures close-up images at the moment of pickup, with the grabber jaws (~8 cm wide) serving as a built-in size reference.
  • AI glasses — worn by the volunteer, capturing a first-person egocentric view of the cleanup scene as they walk and pick up items.

Tier 2 — The Classifier (Vision-Language Model)
Cropped images are sent to a more powerful AI — a vision-language model (VLM) — that determines the category, size, material, condition, whether it's hazardous, and writes a plain-English description for each piece of litter.


Building the Dataset

Metric Value
Images collected 4,499
Individual labels 18,464
Litter categories 71
Collection sites 4 (Kirkland, Bothell, Woodinville, Bellevue)

The images span 14 separate cleanup sessions from December 2023 through January 2025. Each image is annotated with bounding boxes and labeled with a detailed category.


Benchmarking the Classifiers

We ran a head-to-head benchmark on 241 carefully selected litter images across three leading AI services:

AI Provider Broad Category Exact Category Size Accuracy Cost / Image
Google Gemini 73.4% 23.7% 83.4% $0.003
OpenAI GPT-4.1 66.4% 23.2% 71.0% $0.010
Anthropic Claude 59.8% 12.0% 79.7% $0.005

Gemini won on both accuracy and cost, making it our primary cloud classifier.


Performance: Where We Stand

Object Detection (Tier 1) — Our YOLO26 detection model achieves a mAP50 score of 0.905, correctly identifying and locating litter in the frame about 90% of the time.

Classification (Tier 2) — Using Google Gemini, we processed all 4,499 images with a 100% success rate at a total cost of about $4.21.

Metric Result
Size accuracy 83%
Broad category match 73%
Material accuracy 37%
Total cost (4,499 images) $4.21

Training a Local Model — We fine-tuned a vision-language model (Qwen3-VL-8B) using QLoRA on a single consumer GPU (RTX 3080 Ti). After four iterations, our v4 model achieved an eval loss of 0.024 with no overfitting. A working local model would reduce per-image costs by 97%.


What's Next

  • Deploy on edge devices — shrink the model for real-time inference on the clip-on camera and AI glasses via knowledge distillation and optimized model export
  • Connect the end-to-end flow — from data collection with GPS tagging, through detection and classification, to automated report generation for communities and city councils
  • Evaluate the fine-tuned model — head-to-head comparison of our v4 local model against Gemini on unseen images

Get Involved

  • Cleanup organizations — want to pilot the system on your walks? We'll provide the camera setup and handle the AI.
  • ML engineers — the codebase is open source with interesting problems in fine-tuning, edge deployment, and report generation.
  • Community leaders — help us understand what data would be most useful for your advocacy and planning.

Reach out at contact@npotechnologies.com or find us on GitHub.

Elements

Text

This is bold and this is strong. This is italic and this is emphasized. This is superscript text and this is subscript text. This is underlined and this is code: for (;;) { ... }. Finally, this is a link.


Heading Level 2

Heading Level 3

Heading Level 4

Heading Level 5
Heading Level 6

Blockquote

Fringilla nisl. Donec accumsan interdum nisi, quis tincidunt felis sagittis eget tempus euismod. Vestibulum ante ipsum primis in faucibus vestibulum. Blandit adipiscing eu felis iaculis volutpat ac adipiscing accumsan faucibus. Vestibulum ante ipsum primis in faucibus lorem ipsum dolor sit amet nullam adipiscing eu felis.

Preformatted

i = 0;

while (!deck.isInOrder()) {
    print 'Iteration ' + i;
    deck.shuffle();
    i++;
}

print 'It took ' + i + ' iterations to sort the deck.';

Lists

Unordered

  • Dolor pulvinar etiam.
  • Sagittis adipiscing.
  • Felis enim feugiat.

Alternate

  • Dolor pulvinar etiam.
  • Sagittis adipiscing.
  • Felis enim feugiat.

Ordered

  1. Dolor pulvinar etiam.
  2. Etiam vel felis viverra.
  3. Felis enim feugiat.
  4. Dolor pulvinar etiam.
  5. Etiam vel felis lorem.
  6. Felis enim et feugiat.

Icons

Actions

Table

Default

Name Description Price
Item One Ante turpis integer aliquet porttitor. 29.99
Item Two Vis ac commodo adipiscing arcu aliquet. 19.99
Item Three Morbi faucibus arcu accumsan lorem. 29.99
Item Four Vitae integer tempus condimentum. 19.99
Item Five Ante turpis integer aliquet porttitor. 29.99
100.00

Alternate

Name Description Price
Item One Ante turpis integer aliquet porttitor. 29.99
Item Two Vis ac commodo adipiscing arcu aliquet. 19.99
Item Three Morbi faucibus arcu accumsan lorem. 29.99
Item Four Vitae integer tempus condimentum. 19.99
Item Five Ante turpis integer aliquet porttitor. 29.99
100.00

Buttons

  • Disabled
  • Disabled

Form