{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"spatten-llm","owner":"mit-han-lab","isFork":false,"description":"[HPCA'21] SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning","topicNames":["rtl","attention","hardware-acceleration","spinalhdl","llm-inference"],"topicsNotShown":0,"allTopics":["rtl","attention","hardware-acceleration","spinalhdl","llm-inference"],"primaryLanguage":{"name":"Scala","color":"#c22d40"},"pullRequestCount":0,"issueCount":1,"starsCount":48,"forksCount":3,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,8,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-03T21:28:22.876Z"}}],"repositoryCount":1,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}