Все публикации

#205 Chameleon: Plug-and-Play Compositional Reasoning with LLMs

ReAct: Synergizing Reasoning and Acting in Language Models

Video #203 GPTQ: Accurate Post-Training Quantization For Generative Pre-Trained Transformers

Video #202 MoE-LLaVA: Mixture of Experts for Large Vision-Language Models

Video #201: MultiTabQA: Generating Tabular Answers for Multi-Table QA

Video #200: PALO: A Polyglot Large Multimodal Model for 5B People

CALM: LLM Augmented LLMs

XTR: ConteXtualized Token Retriever

PLLaVA: Parameter-free LLaVA Extension from Images to Videos for Video Dense Captioning

YOCO: Decoder-Decoder Architectures for LLMs

KAN: Kolmogorov-Arnold Networks

LLMLingua: Compressing Prompts for Accelerated Inference of LLMs

LaMP: Personalization Benchmark for LLMs

LaMini-LM: A diverse herd of distilled models from large-scale instructions

Mamba: Linear-Time Sequence Modeling with Selective State Spaces

CircuitVQA: A VQA Dataset for Electrical Circuit Images

OAK: Enriching Doc Representations using Auxiliary Knowledge for XC

OpenAI's gpt4o

Apple's OpenELM models

Supporting Infinite Context Length using TransformerFAM

Infini attention and Infini Transformer

MEGALODON: Efficient LLM Pretraining and Inference with Unlimited Context Length

Microsoft VASA-1: Lifelike Audio Driven Talking Faces Generated in Real Time

welcome to shbcf.ru