AI & ML interests

LLM

Recent Activity

Cqy2019Ā  updated a Space about 1 hour ago
OpenMOSS-Team/MOSS-TTSD-v1.0
Cqy2019Ā  updated a model about 6 hours ago
OpenMOSS-Team/MOVA-360p
View all activity

OpenMOSS-Team 's collections 17

ABC-Bench
Evaluating Agentic Backend Coding Capabilities in Real-World Development Scenarios
MOSS Embodied Planner
MHA2MLA-refactor
The MHA2MLA model published in the paper "Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-Based LLMs"
MHA2MLA
The MHA2MLA model published in the paper "Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-Based LLMs"
ABC-Bench
Evaluating Agentic Backend Coding Capabilities in Real-World Development Scenarios
MOSS Embodied Planner
MHA2MLA-refactor
The MHA2MLA model published in the paper "Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-Based LLMs"
MHA2MLA
The MHA2MLA model published in the paper "Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-Based LLMs"
OpenMOSS-Team (OpenMOSS)

AI & ML interests

LLM

Recent Activity

Cqy2019Ā  updated a Space about 1 hour ago
OpenMOSS-Team/MOSS-TTSD-v1.0
Cqy2019Ā  updated a model about 6 hours ago
OpenMOSS-Team/MOVA-360p
View all activity

OpenMOSS-Team 's collections 17

ABC-Bench
Evaluating Agentic Backend Coding Capabilities in Real-World Development Scenarios
MOSS Embodied Planner
MHA2MLA-refactor
The MHA2MLA model published in the paper "Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-Based LLMs"
MHA2MLA
The MHA2MLA model published in the paper "Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-Based LLMs"
ABC-Bench
Evaluating Agentic Backend Coding Capabilities in Real-World Development Scenarios
MOSS Embodied Planner
MHA2MLA-refactor
The MHA2MLA model published in the paper "Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-Based LLMs"
MHA2MLA
The MHA2MLA model published in the paper "Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-Based LLMs"