{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"llama2d","owner":"Llama2D","isFork":false,"description":"2D Positional Embeddings for Webpage Structural Understanding 🦙👀","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":93,"forksCount":2,"license":"GNU General Public License v3.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-06T05:40:58.541Z"}},{"type":"Public","name":"llama-recipes","owner":"Llama2D","isFork":true,"description":"Examples and recipes for Llama 2 model","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":1727,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-08T02:54:30.226Z"}},{"type":"Public","name":"transformers","owner":"Llama2D","isFork":true,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":26450,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-04T23:53:57.904Z"}},{"type":"Public","name":"llama-finetuning","owner":"Llama2D","isFork":true,"description":"Guide for fine-tuning Llama/CodeLlama models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":77,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-21T06:50:16.980Z"}}],"repositoryCount":4,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Llama2D repositories"}