Lexsi Labs
Lexsi.ai
About Us
About Us
Lexsi Labs builds foundations for Safe Superintelligence, uniting alignment theory, interpretability science, and agentic autonomy for AI that's future ready.
Research Papers - Lexsi Labs
Research Papers - Lexsi Labs
Explore the latest Lexsi Labs research papers on Superintelligence, Alignment, and Trustworthy AI. Learn more about the Orion-MSP TFM and the TabTune library.
Contact Us - Lexsi Labs
Contact Us - Lexsi Labs
Connect with Lexsi Labs for strategic research partnerships, potential collaborations, or inquiries regarding Safe Superintelligence, AI Alignment, and Interpretability.
AGI Talks: A night with AI researchers (India AI Impact Summit) · Luma
AGI Talks: A night with AI researchers (India AI Impact Summit) · Luma
IndiaAI AGI Talks is a casual, after-hours get-together during AI Impact Summit for people who like their conversations high-bandwidth and their questions…
[2411.12643] DLBacktrace: A Model Agnostic Explainability for any Deep Learning Models
[2411.12643] DLBacktrace: A Model Agnostic Explainability for any Deep Learning Models
Abstract page for arXiv paper 2411.12643: DLBacktrace: A Model Agnostic Explainability for any Deep Learning Models
[2502.03014] xai_evals : A Framework for Evaluating Post-Hoc Local Explanation Methods
[2502.03014] xai_evals : A Framework for Evaluating Post-Hoc Local Explanation Methods
Abstract page for arXiv paper 2502.03014: xai_evals : A Framework for Evaluating Post-Hoc Local Explanation Methods
[2507.08330] Interpretability-Aware Pruning for Efficient Medical Image Analysis
[2507.08330] Interpretability-Aware Pruning for Efficient Medical Image Analysis
Abstract page for arXiv paper 2507.08330: Interpretability-Aware Pruning for Efficient Medical Image Analysis
[2509.08592] Interpretability as Alignment: Making Internal Understanding a Design Principle
[2509.08592] Interpretability as Alignment: Making Internal Understanding a Design Principle
Frontier AI systems require governance mechanisms that can verify internal alignment, not just behavioral compliance. Private governance mechanisms audits, certification, insurance, and procurement are emerging to complement public regulation, but they require technical substrates that generate verifiable causal evidence about model behavior. This paper argues that mechanistic interpretability provides this substrate. We frame interpretability not as post-hoc explanation but as a design constraint embedding auditability, provenance, and bounded transparency within model architectures. Integrating causal abstraction theory and empirical benchmarks such as MIB and LoBOX, we outline how interpretability-first models can underpin private assurance pipelines and role-calibrated transparency frameworks. This reframing situates interpretability as infrastructure for private AI governance bridging the gap between technical reliability and institutional accountability.
TabTune Library
TabTune Library
A Unified Library for Inference and Fine-Tuning Tabular Foundation Models
Orion-MSP: Multi-Scale Sparse Attention for Tabular In-Context Learning
Orion-MSP: Multi-Scale Sparse Attention for Tabular In-Context Learning
Abstract page for arXiv paper 2511.02818: Orion-MSP: Multi-Scale Sparse Attention for Tabular In-Context Learning
Orion-Bix: Bi-Axial Attention for Tabular In-Context Learning
Orion-Bix: Bi-Axial Attention for Tabular In-Context Learning
Abstract page for arXiv paper 2512.00181: Orion-Bix: Bi-Axial Attention for Tabular In-Context Learning
Exploring Fine-Tuning for Tabular Foundation Models
Exploring Fine-Tuning for Tabular Foundation Models
Tabular Foundation Models (TFMs) have recently shown strong in-context learning capabilities on structured data, achieving zero-shot performance comparable to traditional machine learning methods. We find that zero-shot TFMs already achieve strong performance, while the benefits of fine-tuning are highly model and data-dependent. Meta-learning and PEFT provide moderate gains under specific conditions, whereas full supervised fine-tuning (SFT) often reduces accuracy or calibration quality. This work presents the first comprehensive study of fine-tuning in TFMs across benchmarks including TALENT, OpenML-CC18, and TabZilla. We compare Zero-Shot, Meta-Learning, Supervised (SFT), and parameter-efficient (PEFT) approaches, analyzing how dataset factors such as imbalance, size, and dimensionality affect outcomes. Our findings cover performance, calibration, and fairness, offering practical guidelines on when fine-tuning is most beneficial and its limitations.
Beyond KL Divergence: Policy Optimization with Flexible Bregman Divergences for LLM Reasoning
Beyond KL Divergence: Policy Optimization with Flexible Bregman Divergences for LLM Reasoning
Abstract page for arXiv paper 2602.04380: Beyond KL Divergence: Policy Optimization with Flexible Bregman Divergences for LLM Reasoning
Beyond Uniform Credit: Causal Credit Assignment for Policy Optimization
Beyond Uniform Credit: Causal Credit Assignment for Policy Optimization
Abstract page for arXiv paper 2602.09331: Beyond Uniform Credit: Causal Credit Assignment for Policy Optimization
AlignTune Library
AlignTune Library
Comprehensive fine-tuning library for SFT and RL training with multi-backend support
C-ΔΘ: Circuit-Restricted Weight Arithmetic for Selective Refusal
C-ΔΘ: Circuit-Restricted Weight Arithmetic for Selective Refusal
Abstract page for arXiv paper 2602.04521: $C$-$ΔΘ$: Circuit-Restricted Weight Arithmetic for Selective Refusal
View on mobile
Explore other Linktrees
arianagrande
@arianagrande
morepurposepod
@morepurposepod
Heather Traska
@heathertraska
Ashe
@ashemusic
Ken Eurich
@keneurich
TMG Studios
@tmgstudios
zoltancomedy
@zoltancomedy
Sofia Tilbury
@sofiatilbury
Julian Sewell
@juliansewell
Alli Weatherly
@alliweatherly
previous
next
Discover more
See all
@gmsknstudio
@Steph_dancer
@Prnzsssn
@rebeldenim
@manitos_archive
@BlissBlissASMR
@Yazzeeen
@PAASSMakeADfference
@official128sports
@H3PPyGT
@somewacktatz
@pemdes_kandangan
@rayleen_pethouse
@designcv_
@tanasmaos
@cuteantel
@barriofinotpl
@ergchqho
@Sharkass
@celestearmilla
@Aamir_wani
@findingcrystal
@PPG_RenegadeRV
@Paulmona
@distiva
@kiarasarakos
@soupgirlinatub
@katbatcosplay
@Digitaldevas
@emocionamarketing
@pavilstudio
@ruangsenyum
@paincast
@gerencasino
@englishwithspeakayou
@milyolvi
@its_rickets
@QuerImovel
@cheekyfishing
@radiantwri
@tylerTartistry
@synergystrengthsys
@wafacoffeespace
@grasciela
@harajaltaif