{"payload":{"pageCount":2,"repositories":[{"type":"Public","name":"llm-jp-data-contamination","owner":"llm-jp","isFork":false,"description":"The data contamination repository for the LLM-JP project. Forked from @eiei7\n","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,7,51,60,23,12,135,32,123,73,46,142,94,128,66,10,75,96],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-21T07:06:35.321Z"}},{"type":"Public","name":"awesome-japanese-llm","owner":"llm-jp","isFork":false,"description":"日本語LLMまとめ - Overview of Japanese LLMs","allTopics":["japanese","generative-model","japanese-language","language-models","language-model","generative-models","multimodal","vision-and-language","vision-language","foundation-models","large-language-models","llm","llms","generative-ai","large-language-model","vision-language-model","japanese-llm","japanese-language-model","llm-japanese"],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":0,"issueCount":5,"starsCount":957,"forksCount":29,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-21T03:51:29.719Z"}},{"type":"Public","name":"llm-jp-eval","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":2,"starsCount":93,"forksCount":34,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-20T14:07:12.896Z"}},{"type":"Public","name":"scripts","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Shell","color":"#89e051"},"pullRequestCount":9,"issueCount":4,"starsCount":1,"forksCount":1,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,8,3,6,3,4,5,2,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-20T11:40:59.960Z"}},{"type":"Public","name":"model-cards","owner":"llm-jp","isFork":false,"description":"Collections of LLM-jp model experiments","allTopics":[],"primaryLanguage":null,"pullRequestCount":3,"issueCount":0,"starsCount":1,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,5,3,0,0,1,1,0,0,0,3,0,0,0,4,1,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T08:23:49.375Z"}},{"type":"Public","name":"nii-sakura-cluster-configs","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Shell","color":"#89e051"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,4,22,6,8,7,3,0,1,13,2,7,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-11T10:41:01.561Z"}},{"type":"Public","name":"Megatron-LM","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-05T09:07:32.600Z"}},{"type":"Public","name":"llm-leaderboard","owner":"llm-jp","isFork":true,"description":"Project of llm evaluation to Japanese tasks","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":34,"license":null,"participation":[0,0,0,5,18,6,0,0,1,3,34,32,29,33,0,15,4,16,0,0,13,4,0,2,4,2,0,0,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,1,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-11T08:01:51.021Z"}},{"type":"Public","name":"FastChat","owner":"llm-jp","isFork":true,"description":"An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":4496,"license":"Apache License 2.0","participation":[7,13,20,12,8,14,11,6,23,7,7,13,5,39,4,11,9,22,7,11,20,4,3,5,5,6,4,7,7,0,3,3,0,0,1,0,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-09T08:53:35.878Z"}},{"type":"Public","name":"experiments","owner":"llm-jp","isFork":false,"description":"Issue-Only Pretrain Task Management Repository","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":26,"starsCount":0,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,15,3,3,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-08T02:22:12.579Z"}},{"type":"Public","name":"llm-jp-tokenizer","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Roff","color":"#ecdebe"},"pullRequestCount":0,"issueCount":1,"starsCount":26,"forksCount":7,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-07T04:20:07.265Z"}},{"type":"Public","name":"llmjp_topic_model","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-19T06:17:41.845Z"}},{"type":"Public","name":"modelwg","owner":"llm-jp","isFork":false,"description":"LLM-jp Model-WG Working Directory","allTopics":[],"primaryLanguage":{"name":"Shell","color":"#89e051"},"pullRequestCount":4,"issueCount":0,"starsCount":6,"forksCount":0,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-09T14:02:35.731Z"}},{"type":"Public","name":"DeepSpeed-MII","owner":"llm-jp","isFork":true,"description":"MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":174,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T03:22:54.023Z"}},{"type":"Public","name":"vllm","owner":"llm-jp","isFork":true,"description":"A high-throughput and memory-efficient inference and serving engine for LLMs","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":3978,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-21T08:11:02.218Z"}},{"type":"Public","name":"nii-geniac-gcp-lustre","owner":"llm-jp","isFork":false,"description":"GENIAC用のLustreクラスタ構築スクリプト","allTopics":[],"primaryLanguage":{"name":"Jinja","color":"#a52a22"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-20T07:28:03.958Z"}},{"type":"Public","name":"nii-geniac-hpc-toolkit","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"HCL","color":"#844FBA"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-13T13:47:41.661Z"}},{"type":"Public","name":"llm-jp-sft","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":45,"forksCount":14,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-13T04:17:38.028Z"}},{"type":"Public","name":"nii-geniac-gcp-maintenance","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Shell","color":"#89e051"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-07T15:18:47.830Z"}},{"type":"Public","name":"open-llm-leaderboard-local","owner":"llm-jp","isFork":false,"description":"Open LLM Leaderboard のローカル実行用スクリプト","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":3,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-24T05:45:11.649Z"}},{"type":"Public","name":"llm-jp-model-playground","owner":"llm-jp","isFork":false,"description":"Interactive application to verify multiple LLMs","allTopics":[],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":2,"issueCount":11,"starsCount":13,"forksCount":2,"license":"MIT No Attribution","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-20T00:14:53.781Z"}},{"type":"Public","name":"llm-jp-dpo","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-06T05:18:04.583Z"}},{"type":"Public","name":"llm-jp-corpus","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":40,"forksCount":4,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-02T03:07:50.309Z"}},{"type":"Public","name":"sakura-h100-setup","owner":"llm-jp","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-05T02:19:40.159Z"}},{"type":"Public","name":"Megatron-DeepSpeed","owner":"llm-jp","isFork":false,"description":"microsoft/Megatron-DeepSpeed のフォークです。","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":2,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-01T08:37:10.981Z"}},{"type":"Public","name":"japanese-llm-ranking","owner":"llm-jp","isFork":true,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":12,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-05T06:01:41.492Z"}},{"type":"Public","name":"gpt-neox","owner":"llm-jp","isFork":false,"description":"gpt-neoxのフォークです。mainは元に追従させています","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-09T10:12:07.642Z"}},{"type":"Public","name":"DeepSpeed","owner":"llm-jp","isFork":true,"description":"DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":4052,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-12T07:54:40.113Z"}},{"type":"Public","name":"llm-foundry","owner":"llm-jp","isFork":true,"description":"LLM training code for MosaicML foundation models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":523,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-07T10:43:03.325Z"}},{"type":"Public","name":"llm-jp-corpus-tools","owner":"llm-jp","isFork":false,"description":"Tools for llm-jp-corpus","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-07-11T20:02:16.869Z"}}],"repositoryCount":31,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"llm-jp repositories"}