{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"DNABERT_S","owner":"MAGICS-LAB","isFork":false,"description":"DNABERT_S: Learning Species-Aware DNA Embedding with Genome Foundation Models","topicNames":["dna","embedding","dna-embedding"],"topicsNotShown":0,"allTopics":["dna","embedding","dna-embedding"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":26,"forksCount":7,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,0,1,2,0,0,0,0,0,0,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-13T20:30:13.198Z"}},{"type":"Public","name":"OutEffHop","owner":"MAGICS-LAB","isFork":false,"description":"","topicNames":[],"topicsNotShown":0,"allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,10,0,0,0,0,0,0,0,0,3,0,0,0,0,2,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-09T21:23:32.636Z"}},{"type":"Public","name":"DNABERT_2","owner":"MAGICS-LAB","isFork":false,"description":"[ICLR 2024] DNABERT-2: Efficient Foundation Model and Benchmark for Multi-Species Genome","topicNames":["transcription-factors","genome","splice","dna","dataset","promoters","promoter","transcription-factor-binding","language-model","genome-analysis"],"topicsNotShown":6,"allTopics":["transcription-factors","genome","splice","dna","dataset","promoters","promoter","transcription-factor-binding","language-model","genome-analysis","promoter-analysis","transcription-factor-binding-site","covid","dna-processing","splice-site","dna-training"],"primaryLanguage":{"name":"Shell","color":"#89e051"},"pullRequestCount":3,"issueCount":34,"starsCount":184,"forksCount":41,"license":"Apache License 2.0","participation":[0,0,0,1,8,1,1,3,0,0,1,0,1,2,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-30T13:01:11.340Z"}},{"type":"Public","name":"UHop","owner":"MAGICS-LAB","isFork":false,"description":"Uniform Memory Retrieval with Larger Capacity for Modern Hopfield Models","topicNames":["hopfield-network","associative-memory","hopfield","memory-capacity","network-capacity","modern-hopfield-networks","modern-hopfield-model","sparse-hopfield","hopfield-model","generalized-sparse-hopfield"],"topicsNotShown":0,"allTopics":["hopfield-network","associative-memory","hopfield","memory-capacity","network-capacity","modern-hopfield-networks","modern-hopfield-model","sparse-hopfield","hopfield-model","generalized-sparse-hopfield"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,3,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-04T18:07:39.391Z"}},{"type":"Public","name":"BiSHop","owner":"MAGICS-LAB","isFork":false,"description":"BiSHop: Bi-Directional Cellular Learning for Tabular Data with Generalized Sparse Modern Hopfield Model","topicNames":["hopfield-network","tabular","hopfield","tabular-methods","tabular-model","modern-hopfield-networks","modern-hopfield-model","tabular-learn","generalized-sparse-modern-hopfield","sparse-hopfield"],"topicsNotShown":0,"allTopics":["hopfield-network","tabular","hopfield","tabular-methods","tabular-model","modern-hopfield-networks","modern-hopfield-model","tabular-learn","generalized-sparse-modern-hopfield","sparse-hopfield"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,0,1,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-04T16:35:59.761Z"}},{"type":"Public","name":"SparseModernHopfield","owner":"MAGICS-LAB","isFork":false,"description":"[NeurIPS 2023] On Sparse Modern Hopfield Model","topicNames":["transformer","attention","hopfield","modern-hopfield-networks","modern-hopfield-model"],"topicsNotShown":0,"allTopics":["transformer","attention","hopfield","modern-hopfield-networks","modern-hopfield-model"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":48,"forksCount":11,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,10,0,9,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-26T05:58:24.132Z"}},{"type":"Public","name":"STanHop","owner":"MAGICS-LAB","isFork":false,"description":"[ICLR 2024] STanHop: Sparse Tandem Hopfield Model for Memory-Enhanced Time Series Prediction","topicNames":["transformer","attention-mechanism","hopfield","time-series-prediction","time-series-forecasting","modern-hopfield-networks","modern-hopfield-model"],"topicsNotShown":0,"allTopics":["transformer","attention-mechanism","hopfield","time-series-prediction","time-series-forecasting","modern-hopfield-networks","modern-hopfield-model"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-18T23:22:42.029Z"}},{"type":"Public","name":"NonparametricHopfield","owner":"MAGICS-LAB","isFork":false,"description":"Nonparametric Modern Hopfield Models","topicNames":["efficient-transformers","efficient-attention","modern-hopfield-networks","modern-hopfield-model","efficient-hopfield-models","efficient-hopfield-networks"],"topicsNotShown":0,"allTopics":["efficient-transformers","efficient-attention","modern-hopfield-networks","modern-hopfield-model","efficient-hopfield-models","efficient-hopfield-networks"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":"Other","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-08T20:53:10.159Z"}},{"type":"Public","name":"DNABERT","owner":"MAGICS-LAB","isFork":true,"description":"[Bioinformatics] DNABERT: pre-trained Bidirectional Encoder Representations from Transformers model for DNA-language in genome","topicNames":["dna"],"topicsNotShown":0,"allTopics":["dna"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":150,"license":"Apache License 2.0","participation":[0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-26T20:04:37.361Z"}}],"repositoryCount":9,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}