IntSeqBERT:通过模谱嵌入学习OEIS中的算术结构

📄 中文摘要

整数序列在OEIS中涵盖了从单数字常数到天文级别的阶乘和指数值,给标准的标记化模型带来了预测挑战,因为这些模型无法处理超出词汇表的值或利用周期性算术结构。研究提出了IntSeqBERT,这是一种双流Transformer编码器,专门用于OEIS中的掩码整数序列建模。每个序列元素沿两个互补轴进行编码:连续对数尺度的幅度嵌入和100个余数(模数$2$--$101$)的正弦/余弦模嵌入,通过FiLM融合。三个预测头(幅度回归、符号分类和100个模数的模预测)在274,705个OEIS序列上共同训练。在大规模模型(91.5M参数)下,IntSeqBERT展示了其在序列预测中的有效性。

📄 English Summary

IntSeqBERT: Learning Arithmetic Structure in OEIS via Modulo-Spectrum Embeddings

The research presents IntSeqBERT, a dual-stream Transformer encoder designed for masked integer-sequence modeling on the OEIS. Integer sequences in the OEIS range from single-digit constants to astronomical factorials and exponentials, posing challenges for standard tokenized models that cannot handle out-of-vocabulary values or exploit periodic arithmetic structures. Each sequence element is encoded along two complementary axes: a continuous log-scale magnitude embedding and sin/cos modulo embeddings for 100 residues (moduli 2 to 101), fused via FiLM. Three prediction heads—magnitude regression, sign classification, and modulo prediction for 100 moduli—are jointly trained on 274,705 OEIS sequences. At a large scale (91.5M parameters), IntSeqBERT demonstrates its effectiveness in sequence prediction.

Powered by Cloudflare Workers + Payload CMS + Claude 3.5

数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等