Published November 6, 2025 | Version v1

Preprint Open

Description

We present a deterministic, matrix-free framework for deep representation learning based on geometric invariants and analytic Koopman–tangent projections in a weighted log-prime Hilbert space rather than stochastic optimization. The method constructs a representation in a single, fully parallelizable pass through the data, without iterative optimization or gradient descent. The core of the framework is a Goldilocks (Gamma) measure (a multiplicative–additive equilibrium measure minimizing energy in the log-prime basis) that defines a weighted Hilbert space, a log-prime orthogonal basis that yields a diagonal *surrogate* under the weighted spectral basis, rendering per-mode updates independent, and a variational…

Similar Posts

Loading similar posts...

Keyboard Shortcuts

Navigation
Next / previous item
j/k
Open post
oorEnter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
gh
Interests
gi
Feeds
gf
Likes
gl
History
gy
Changelog
gc
Settings
gs
Browse
gb
Search
/
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc

Press ? anytime to show this help