BayJarvis Blogs
BayJarvis: all Blogs
paper
Faith and Fate: Limits of Transformers on Compositionality
- 2024-04-16
-->
Voyager: An Open-Ended Embodied Agent with Large Language Models
- 2024-04-13
-->
Reflexion: Language Agents with Verbal Reinforcement Learning
- 2024-04-13
-->
Scaling Laws for Fine-Grained Mixture of Experts
- 2024-04-06
-->
FrugalGPT: Making Large Language Models Affordable and Efficient
- 2024-04-04
-->
ROUTERBENCH: A Benchmark for Multi-LLM Routing System
- 2024-04-04
-->
Toy Models of Superposition
- 2024-04-03
-->
Cognitive Architectures for Language Agents
- 2024-04-01
-->
Retrieval-Augmented Generation for Large Language Models: A Survey
- 2024-03-31
-->
LLMLingua: Compressing Prompts for Accelerated Inference of Large Language Models
- 2024-03-26
-->
Efficient Memory Management for Large Language Model Serving with PagedAttention
- 2024-03-25
-->
Evolutionary Optimization of Model Merging Recipes
- 2024-03-24
-->
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
- 2024-03-21
-->
OpenMoE: An Early Effort on Open Mixture-of-Experts Language Models
- 2024-03-20
-->
Training Language Model Agents without Modifying Language Models
- 2024-03-19
-->
Characterizing Large Language Models Geometry for Toxicity Detection and Generation
- 2024-03-18
-->
MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
- 2024-03-17
-->
Scaling Laws for Forgetting When Fine-Tuning Large Language Models
- 2024-03-16
-->
Simple and Scalable Strategies to Continually Pre-train Large Language Models
- 2024-03-15
-->
A Rank Stabilization Scaling Factor for Fine-Tuning with LoRA
- 2024-03-14
-->
Decision Transformer: Reinforcement Learning via Sequence Modeling
- 2024-03-14
-->
In-Context Learning for Extreme Multi-Label Classification
- 2024-03-13
-->
PinnerFormer: Sequence Modeling for User Representation at Pinterest
- 2024-03-11
-->
BitNet: Scaling 1-bit Transformers for Large Language Models
- 2024-03-09
-->
Genie: Generative Interactive Environments
- 2024-02-28
-->
AMAGO: Scalable In-Context Reinforcement Learning for Adaptive Agents
- 2024-02-26
-->
Self-Discover: Large Language Models Self-Compose Reasoning Structures
- 2024-02-25
-->
Promptbreeder: Self-Referential Self-Improvement Via Prompt Evolution
- 2024-02-24
-->
A Decoder-Only Foundation Model for Time-Series Forecasting
- 2024-02-19
-->
Progressive Layered Extraction (PLE) for Multi-Task Learning in Personalized Recommendations
- 2024-02-14
-->
Hiformer: Heterogeneous Feature Interactions Learning with Transformers for Recommender Systems
- 2024-02-14
-->
MPNet: Masked and Permuted Pre-training for Language Understanding
- 2024-02-12
-->
Machine Unlearning for Image-to-Image Generative Models
- 2024-02-09
-->
Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models
- 2024-02-06
-->
Prompting Large Language Models With the Socratic Method
- 2024-01-05
-->
Multi-Agent Reasoning with Large Language Models for Effective Corporate Planning
- 2024-01-03
-->
Mamba: Linear-Time Sequence Modeling with Selective State Spaces
- 2023-12-30
-->
Beyond Human Data: Scaling Self-Training for Problem-Solving with Language Models
- 2023-12-25
-->
Mixture-of-Experts Meets Instruction Tuning: A Winning Combination for Large Language Models
- 2023-12-19
-->
Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer
- 2023-12-18
-->
Learning Factored Representations in a Deep Mixture of Experts
- 2023-12-15
-->
Diffusion Models for Reinforcement Learning: A Survey
- 2023-12-13
-->
Deep Reinforcement Learning from Human Preferences
- 2023-12-10
-->
Denoising Diffusion Probabilistic Models
- 2023-12-09
-->
ITRANSFORMER: Inverted Transformers Are Effective For Time Series Forecasting
- 2023-12-04
-->
Unraveling the Complexities of Multimodal AI: Insights from Visual Instruction Tuning
- 2023-11-30
-->
Orca 2: Teaching Small Language Models How to Reason
- 2023-11-29
-->
A Survey on Language Models for Code: from Statistical Models to AI-driven Code Mastery
- 2023-11-28
-->
Simplifying Transformer blocks: Innovations in Model Efficiency
- 2023-11-28
-->
Exploring the "System 2 Attention" in AI: Innovations and Variations
- 2023-11-27
-->
Let’s Verify Step by Step
- 2023-11-26
-->
Unraveling EcoAssistant: Autogen's Advancement in Economical and Precise Code-Driven Question Answering
- 2023-11-13
-->
AutoGen: Unleashing the Power of Multi-Agent Conversations in LLM Applications
- 2023-11-12
-->
MemGPT: Towards LLMS As Operating Systems
- 2023-11-11
-->
A Comprehensive Overview of LLM-Based Autonomous Agents
- 2023-11-10
-->
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
- 2023-11-05
-->
Branching Beyond PPO: How MCTS Sprouts Superior Text Generation
- 2023-11-05
-->
Constitutional AI - Training AI Systems to Be Helpful and Harmless Using AI Feedback
- 2023-11-04
-->
Representation Engineering: Unraveling the Top-Down Approach to AI Transparency
- 2023-11-02
-->
Prompting the Future: From Hard-Coded to Hard-Core Compiler Magic in DSPy
- 2023-10-31
-->
Cost-Effective Hyperparameter Tuning for LLMs on a Budget
- 2023-10-18
-->
Scaling Laws for Autoregressive Generative Modeling: A Review
- 2023-10-11
-->
From Draft to Target: Optimizing Language Model Decoding with Speculative Sampling
- 2023-09-04
-->
Revolutionizing Language Model Fine-Tuning: The Power of QLORA
- 2023-08-27
-->
Delving Deep into Low-Rank Updates with LoRA
- 2023-08-26
-->
The Nexus of AI and Human Intuition
- 2023-08-24
llm
In Brief: Welcome Google's Gemma - New Open LLM
- 2024-02-22
-->
Socratic Method Prompt Templates for LLM Interactions
- 2024-01-06
-->
Harnessing Zephyr's Breeze: DPO Training on Mistral-7B-GPTQ for Language Model Alignment
- 2023-11-09
-->
Unleashing Dual Power: Switching Seamlessly Between Zephyr & Mistral 7B Models in Multiple LLMs
- 2023-11-09
-->
Fine-tuning Zephyr 7B GPTQ with 4-Bit Quantization for Custom Data and Inference
- 2023-11-08
-->
Optimizing Llama 2: Harnessing the Power of Prompt, RAG, and Fine-Tuning
- 2023-11-04
-->
Building the Future of Instruction-Based Code Generation: An Exploration of Code Alpaca's LLaMA Models with Ludwig's Fine-Tuning QLORA Technique
- 2023-09-01
-->
From Big Servers to Your Laptop: Running Llama2, Dolly2, and More in Your Local Environment
- 2023-08-30
vision
Introduction to 3D Gaussian Splatting
- 2023-12-25
network-architecture
The Annotated S4: Understanding Structured State Spaces in Sequence Modeling
- 2023-12-22
reinforcement-learning
Cicero: Mastering the Art of Diplomacy through Advanced AI
- 2023-12-19
-->
Mastering Stability in PPO: Journey Beyond NaNs and Infs
- 2023-10-19
autonomous-agent
Implementing EcoAssistant: Leveraging AutoGen for Enhanced Code-driven Question Answering
- 2023-11-13