Speculative Product · Information Architecture · AI-Assisted Build · 2026

Aurafact

A spatial archive for environmental sound. Built from a cultural thesis, developed with AI, and published as a fully functional product.

Visit Aurafact

Role

Lead Product Designer: Systems & IA

Type

Self-directed speculative product

Built With

Claude, Cursor, ChatGPT, p5.js, GSAP

Outcome

Live multi-page product hosted on GitHub, 13 archived specimens across 5 regions

Aurafact — hero screen showing animated waveform visualisation over a dark ground with the headline: Archiving the Acoustic Memory of Place

The Problem

We archive almost nothing we hear.

We live in a visually saturated world. Landscapes are mapped, photographed, and recorded with precision. But their acoustic identity, the density of birds before a wildfire, the rhythm of a tram line, the low hum of a harbour at dawn, disappears without record.

Existing approaches to sound archiving sit at two unhelpful extremes. Academic sound libraries are rich in data but absent of experience — digital graveyards that no one browses. Ambient sound apps are beautiful to inhabit but impossible to study, more like toys, not tools. Neither treats environmental audio as what it actually is: a primary historical artifact Sound is not absent. It is simply unkept.

Design question: How might we design an archive that treats environmental sound as a cultural artifact, searchable, interpretable, and usable across audiences?

Digital Graveyards

Academic sound archives, like The British Library Sound Archive, are rich in data, but are rarely experienced.

Aurafact sits between two worlds

A system that combines structure and experience. A research-grade archive that you actually want to inhabit.


Key Insight

Sound is not a file. It is an environmental event.

Most interfaces treat audio as something to play. Aurafact treats it as something that happened. Aurafact treats sound as a documented specimen, inseparable from where it was captured, how it was recorded, what was present in that moment.

This reframed the entire system because a recording doesn’t belong in a playlist, it belongs in an archive with context, structure, and provenance.

A place without its sound is only half remembered.

— Aurafact field notes, 2026
Aurafact Specimen Archive — searchable table showing specimen IDs, artifact names, locations, source types, subtypes, and frequency profiles with a detail panel open on the right
The Specimen Archive: Each recording documented as a traceable artifact, not a file.

AI Collaboration

Built with AI. Directed by design.

Aurafact was developed in collaboration with three AI tools, each used with a specific role not as shortcuts, but as extensions of the design process.

Claude + ChatGPT

Writing, structure, and rationale

Used to refine the manifesto, taxonomy, field notes, and case study language. Every output was reviewed and edited. The goal was precision, not automation.

Cursor + Claude

Front-end and interaction design

Used to write and iterate on the HTML, CSS, and JavaScript across all five pages. The p5.js waveform, GSAP transitions, archive table with split-panel detail view, and taxonomy colour system were all developed through directed prompting in Cursor, each output tested and refined against design intent.

ChatGPT + Perplexity

Research and taxonomy

Used to develop the three-axis classification system, the acoustic metadata schema and, the supporting research. All information was cross-checked before use. AI supported thinking, it did not replace it.

AI amplifies clarity or confusion. The quality of output followed the quality of the brief, and precision in prompting became a design skill.


Design Decisions

Four decisions that made sound navigable.

01

The interface begins in darkness

The homepage opens on a near-black field with a slow waveform animation, no imagery. No visual anchor. The absence of a visual anchor is a provocation, it asks the visitor to arrive differently than they would on any other site.

02

Dual typography system

Two typefaces defined two modes. Serif for narrative and interpretation, and monospace for data and structure. The shift is functional because it tells the user what they are reading. In an archive where interpretation and data coexist on the same page, typography becomes a navigational signal, not decoration.

03

A three-axis taxonomy

Classifying sound across three axes, source type, location, and frequency profile, required building a schema that works for both a researcher querying by biophony type and a visitor exploring by region. The taxonomy had to hold both use cases without flattening either.

04

Archive as database, not gallery

The default for audio online is a gallery or playlist, prioritising mood over meaning. A table was the only format that kept provenance, metadata, and specimen identity intact. The table earns its archival credibility precisely because it doesn't feel like a music app.

Visit the live product
Aurafact Field Guide — design principles panel on the left alongside acoustic studies photography grid on the right
The Field Guide: The system’s principles and acoustic studies, made visible.

Outcome

A live, multi-page product demonstrating a new model for sound archiving.

Aurafact is hosted on GitHub and live at jenniferyaya.github.io/aura-fact. The product includes four pages:Homepage, Archive, Field Guide, and Glossary. The archive contains 13 specimens across 5 regions, each with full metadata and environmental context.

The homepage features a p5.js waveform animation, a searchable and sortable archive table, and a split-panel detail view with playback and full metadata. Every interaction was built through directed AI collaboration and refined to match the product's tone.

The project demonstrated that when you work with AI, the clearer the design brief and then the better the output. The more specific the prompt, the less revision required. AI fluency, at its core, is design fluency.

Aurafact homepage — below-fold section showing acoustic preservation statement and recent additions table with specimen data
Homepage (below fold): The preservation manifesto alongside recent additions.

Reflection

What designing for sound revealed about interfaces

Designing for sound revealed how deeply visual most interface assumptions are. Building Aurafact required constructing a different representational vocabulary from scratch, taxonomy as sensory translation, metadata as environmental context, typography as content signal.

Working with AI throughout the build also changed how I think about the relationship between design thinking and tools. A prompt is a brief. A revision is a critique. Cursor and Claude don't know what good looks like in your product, you do. The designer's job doesn't disappear; it becomes more exact.

If the project continued, the next layer of work would include:

Longer term: using AI not just to build the archive, but to reconstruct lost sound environments.

Next Project

Clara: Clarity and collaboration for every journey