Quantization, Attention Mechanisms, Batch Processing, KV Caching
When LLM Meets Time Series: Can LLMs Perform Multi-Step Time Series Reasoning and Inference
arxiv.orgยท22h
Yet Unnoticed in LSTM: Binary Tree Based Input Reordering, Weight Regularization, and Gate Nonlinearization
arxiv.orgยท22h
Solving Deepfakes with Traces, Frequency, and Attention!
pub.towardsai.netยท13h
Natural Latents: Latent Variables Stable Across Ontologies
lesswrong.comยท1h
SCOUT: Toward Sub-Quadratic Attention via Segment Compression for Optimized Utility in Transformers
arxiv.orgยท22h
Convolutional Denoising Autoencoders for Diagnostic Images
haydenramm.bearblog.devยท14h
Knowledge-integrated AutoEncoder Model
arxiv.orgยท22h
Loading...Loading more...