{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"ChunkLlama","owner":"HKUNLP","isFork":false,"description":"[ICML'24] Data and code for our paper \"Training-Free Long-Context Scaling of Large Language Models\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":11,"starsCount":325,"forksCount":17,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-19T06:56:44.205Z"}},{"type":"Public","name":"diffusion-of-thoughts","owner":"HKUNLP","isFork":false,"description":"Code for the paper \"Diffusion of Thoughts: Chain-of-Thought Reasoning in Diffusion Language Models\"","allTopics":["machine-learning","natural-language-processing","text-generation","pytorch","diffusion-models","non-autoregressive","mathematical-reasoning","chain-of-thought-reasoning","diffusion-lm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":59,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-24T05:43:10.981Z"}},{"type":"Public","name":"RSA","owner":"HKUNLP","isFork":true,"description":"Retrieved Sequence Augmentation for Protein Representation Learning","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":41,"forksCount":4,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-01T13:46:27.461Z"}},{"type":"Public","name":"SymGen","owner":"HKUNLP","isFork":false,"description":"[EMNLP'23] Code for Generating Data for Symbolic Language with Large Language Models","allTopics":["python","bash","sql","prolog","data-generation","codex","mtop","symbolic-language"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":16,"forksCount":1,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-21T04:33:02.445Z"}},{"type":"Public","name":"multilingual-transfer","owner":"HKUNLP","isFork":false,"description":"Code for paper ”Language Versatilists vs. Specialists: An Empirical Revisiting on Multilingual Transfer Ability“","allTopics":["multilingual","bloom","transfer","llama","reasoning","llms"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":12,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-06-13T06:17:53.958Z"}},{"type":"Public","name":"GIMLET","owner":"HKUNLP","isFork":true,"description":"The code for GIMLET: A Unified Graph-Text Model for Instruction-Based Molecule Zero-Shot Learning","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":2,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-16T13:18:02.164Z"}},{"type":"Public","name":"icl-ceil","owner":"HKUNLP","isFork":false,"description":"[ICML 2023] Code for our paper “Compositional Exemplars for In-context Learning”.","allTopics":["retrieval","metric-learning","determinantal-point-processes","compositionality","in-context-learning","large-language-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":90,"forksCount":10,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-03-15T08:48:48.792Z"}},{"type":"Public","name":"efficient-attention","owner":"HKUNLP","isFork":false,"description":"[EVA ICLR'23; LARA ICML'22] Efficient attention mechanisms via control variates, random features, and importance sampling","allTopics":["machine-learning","machine-translation","python3","language-model","attention-mechanism","vision-transformer","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":78,"forksCount":4,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-03-07T08:47:28.627Z"}},{"type":"Public","name":"SunGen","owner":"HKUNLP","isFork":true,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":9,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-02-26T06:39:29.596Z"}},{"type":"Public","name":"HumanPrompt","owner":"HKUNLP","isFork":false,"description":"A framework for human-readable prompt-based method with large language models. Specially designed for researchers. (Deprecated, check out LangChain for better usage!)","allTopics":["natural-language-processing","large-language-models","prompt-engineering"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":126,"forksCount":8,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-02-25T10:24:05.496Z"}},{"type":"Public","name":"reparam-discrete-diffusion","owner":"HKUNLP","isFork":false,"description":"Reparameterized Discrete Diffusion Models for Text Generation","allTopics":["machine-learning","natural-language-processing","text-generation","python3","language-model","fairseq","diffusion-models","non-autoregressive","pytorch"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":4,"starsCount":90,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-02-14T01:44:44.667Z"}},{"type":"Public","name":"ProGen","owner":"HKUNLP","isFork":false,"description":"[EMNLP-2022 Findings] Code for paper “ProGen: Progressive Zero-shot Dataset Generation via In-context Feedback”.","allTopics":["data-generation","zero-shot-learning","influence-functions","in-context-learning","large-language-models"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":21,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-02-04T16:05:59.664Z"}},{"type":"Public","name":"ZeroGen","owner":"HKUNLP","isFork":true,"description":"[EMNLP 2022] Code for our paper “ZeroGen: Efficient Zero-shot Learning via Dataset Generation”.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":15,"forksCount":10,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-02-18T08:30:35.047Z"}}],"repositoryCount":13,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"HKUNLP repositories"}