Learning Video Representations from Large Language Models
📝We introduce LaViLa, a new approach to learning video-language representations by leveraging Large Language Models (LLMs).
https://github.com/facebookresearch/lavila
📝We introduce LaViLa, a new approach to learning video-language representations by leveraging Large Language Models (LLMs).
https://github.com/facebookresearch/lavila
GitHub
GitHub - facebookresearch/LaViLa: Code release for "Learning Video Representations from Large Language Models"
Code release for "Learning Video Representations from Large Language Models" - facebookresearch/LaViLa
EVA: Exploring the Limits of Masked Visual Representation Learning at Scale
📝We launch EVA, a vision-centric foundation model to explore the limits of visual representation at scale using only publicly accessible data.
https://github.com/baaivision/eva
📝We launch EVA, a vision-centric foundation model to explore the limits of visual representation at scale using only publicly accessible data.
https://github.com/baaivision/eva
GitHub
GitHub - baaivision/EVA: EVA Series: Visual Representation Fantasies from BAAI
EVA Series: Visual Representation Fantasies from BAAI - baaivision/EVA
ConvNeXt V2: Co-designing and Scaling ConvNets with Masked Autoencoders
📝This co-design of self-supervised learning techniques and architectural improvement results in a new model family called ConvNeXt V2, which significantly improves the performance of pure ConvNets on various recognition benchmarks, including ImageNet classification, COCO detection, and ADE20K segmentation.
https://github.com/facebookresearch/convnext-v2
📝This co-design of self-supervised learning techniques and architectural improvement results in a new model family called ConvNeXt V2, which significantly improves the performance of pure ConvNets on various recognition benchmarks, including ImageNet classification, COCO detection, and ADE20K segmentation.
https://github.com/facebookresearch/convnext-v2
GitHub
GitHub - facebookresearch/ConvNeXt-V2: Code release for ConvNeXt V2 model
Code release for ConvNeXt V2 model. Contribute to facebookresearch/ConvNeXt-V2 development by creating an account on GitHub.
Cramming: Training a Language Model on a Single GPU in One Day
📝Recent trends in language modeling have focused on increasing performance through scaling, and have resulted in an environment where training language models is out of reach for most researchers and practitioners.
https://github.com/jonasgeiping/cramming
📝Recent trends in language modeling have focused on increasing performance through scaling, and have resulted in an environment where training language models is out of reach for most researchers and practitioners.
https://github.com/jonasgeiping/cramming
GitHub
GitHub - JonasGeiping/cramming: Cramming the training of a (BERT-type) language model into limited compute.
Cramming the training of a (BERT-type) language model into limited compute. - JonasGeiping/cramming
Muse: Text-To-Image Generation via Masked Generative Transformers
📝Compared to pixel-space diffusion models, such as Imagen and DALL-E 2, Muse is significantly more efficient due to the use of discrete tokens and requiring fewer sampling iterations; compared to autoregressive models, such as Parti, Muse is more efficient due to the use of parallel decoding.
https://github.com/lucidrains/muse-pytorch
📝Compared to pixel-space diffusion models, such as Imagen and DALL-E 2, Muse is significantly more efficient due to the use of discrete tokens and requiring fewer sampling iterations; compared to autoregressive models, such as Parti, Muse is more efficient due to the use of parallel decoding.
https://github.com/lucidrains/muse-pytorch
GitHub
GitHub - lucidrains/muse-maskgit-pytorch: Implementation of Muse: Text-to-Image Generation via Masked Generative Transformers,…
Implementation of Muse: Text-to-Image Generation via Masked Generative Transformers, in Pytorch - GitHub - lucidrains/muse-maskgit-pytorch: Implementation of Muse: Text-to-Image Generation via Mask...
A Survey for In-context Learning
📝With the increasing ability of large language models (LLMs), in-context learning (ICL) has become a new paradigm for natural language processing (NLP), where LLMs make predictions only based on contexts augmented with a few training examples.
https://github.com/dqxiu/icl_paperlist
📝With the increasing ability of large language models (LLMs), in-context learning (ICL) has become a new paradigm for natural language processing (NLP), where LLMs make predictions only based on contexts augmented with a few training examples.
https://github.com/dqxiu/icl_paperlist
GitHub
GitHub - dqxiu/ICL_PaperList: Paper List for In-context Learning 🌷
Paper List for In-context Learning 🌷. Contribute to dqxiu/ICL_PaperList development by creating an account on GitHub.
Reasoning over Different Types of Knowledge Graphs: Static, Temporal and Multi-Modal
📝The early works in this domain mainly focus on static KGR and tend to directly apply general knowledge graph embedding models to the reasoning task.
https://github.com/liangke23/awesome-knowledge-graph-reasoning
📝The early works in this domain mainly focus on static KGR and tend to directly apply general knowledge graph embedding models to the reasoning task.
https://github.com/liangke23/awesome-knowledge-graph-reasoning
GitHub
GitHub - LIANGKE23/Awesome-Knowledge-Graph-Reasoning: AKGR: Awesome Knowledge Graph Reasoning is a collection of knowledge graph…
AKGR: Awesome Knowledge Graph Reasoning is a collection of knowledge graph reasoning works, including papers, codes and datasets - GitHub - LIANGKE23/Awesome-Knowledge-Graph-Reasoning: AKGR: Awesom...
BanditPAM: Almost Linear Time k-Medoids Clustering via Multi-Armed Bandits
📝In these experiments, we observe that BanditPAM returns the same results as state-of-the-art PAM-like algorithms up to 4x faster while performing up to 200x fewer distance computations.https://github.com/ThrunGroup/BanditPAM
📝In these experiments, we observe that BanditPAM returns the same results as state-of-the-art PAM-like algorithms up to 4x faster while performing up to 200x fewer distance computations.https://github.com/ThrunGroup/BanditPAM
GitHub
GitHub - motiwari/BanditPAM: BanditPAM C++ implementation and Python package
BanditPAM C++ implementation and Python package. Contribute to motiwari/BanditPAM development by creating an account on GitHub.
SegGPT: Segmenting Everything In Context
📝We unify various segmentation tasks into a generalist in-context learning framework that accommodates different kinds of segmentation data by transforming them into the same format of images.https://github.com/baaivision/painter
📝We unify various segmentation tasks into a generalist in-context learning framework that accommodates different kinds of segmentation data by transforming them into the same format of images.https://github.com/baaivision/painter
GitHub
GitHub - baaivision/Painter: Painter & SegGPT Series: Vision Foundation Models from BAAI
Painter & SegGPT Series: Vision Foundation Models from BAAI - GitHub - baaivision/Painter: Painter & SegGPT Series: Vision Foundation Models from BAAI
Instruction Tuning with GPT-4
📝Prior work has shown that finetuning large language models (LLMs) using machine-generated instruction-following data enables such models to achieve remarkable zero-shot capabilities on new tasks, and no human-written instructions are needed.https://github.com/Instruction-Tuning-with-GPT-4/GPT-4-LLM
📝Prior work has shown that finetuning large language models (LLMs) using machine-generated instruction-following data enables such models to achieve remarkable zero-shot capabilities on new tasks, and no human-written instructions are needed.https://github.com/Instruction-Tuning-with-GPT-4/GPT-4-LLM
GitHub
GitHub - Instruction-Tuning-with-GPT-4/GPT-4-LLM: Instruction Tuning with GPT-4
Instruction Tuning with GPT-4. Contribute to Instruction-Tuning-with-GPT-4/GPT-4-LLM development by creating an account on GitHub.
HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in HuggingFace
📝Solving complicated AI tasks with different domains and modalities is a key step toward advanced artificial intelligence.https://github.com/microsoft/JARVIS
📝Solving complicated AI tasks with different domains and modalities is a key step toward advanced artificial intelligence.https://github.com/microsoft/JARVIS
GitHub
GitHub - microsoft/JARVIS: JARVIS, a system to connect LLMs with ML community. Paper: https://arxiv.org/pdf/2303.17580.pdf
JARVIS, a system to connect LLMs with ML community. Paper: https://arxiv.org/pdf/2303.17580.pdf - microsoft/JARVIS
Segment Anything
📝We introduce the Segment Anything (SA) project: a new task, model, and dataset for image segmentation.https://github.com/facebookresearch/segment-anything
📝We introduce the Segment Anything (SA) project: a new task, model, and dataset for image segmentation.https://github.com/facebookresearch/segment-anything
GitHub
GitHub - facebookresearch/segment-anything: The repository provides code for running inference with the SegmentAnything Model (SAM)…
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model. -...
GPTQ: Accurate Post-Training Quantization for Generative Pre-trained Transformers
📝In this paper, we address this challenge, and propose GPTQ, a new one-shot weight quantization method based on approximate second-order information, that is both highly-accurate and highly-efficient.https://github.com/thudm/chatglm-6b
📝In this paper, we address this challenge, and propose GPTQ, a new one-shot weight quantization method based on approximate second-order information, that is both highly-accurate and highly-efficient.https://github.com/thudm/chatglm-6b
GitHub
GitHub - THUDM/ChatGLM-6B: ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型 - GitHub - THUDM/ChatGLM-6B: ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
SadTalker: Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation📝We present SadTalker, which generates 3D motion coefficients (head pose, expression) of the 3DMM from audio and implicitly modulates a novel 3D-aware face render for talking head generation.https://github.com/winfredy/sadtalker
Segment Everything Everywhere All at Once
📝https://github.com/ux-decoder/segment-everything-everywhere-all-at-once
Segment Everything Everywhere All at Once
📝https://github.com/ux-decoder/segment-everything-everywhere-all-at-once
GitHub
GitHub - OpenTalker/SadTalker: [CVPR 2023] SadTalker:Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single…
[CVPR 2023] SadTalker:Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation - OpenTalker/SadTalker
ImageReward: Learning and Evaluating Human Preferences for Text-to-Image Generation
📝https://github.com/thudm/imagereward
📝https://github.com/thudm/imagereward
GitHub
GitHub - THUDM/ImageReward: ImageReward: Learning and Evaluating Human Preferences for Text-to-image Generation
ImageReward: Learning and Evaluating Human Preferences for Text-to-image Generation - GitHub - THUDM/ImageReward: ImageReward: Learning and Evaluating Human Preferences for Text-to-image Generation
Self-Instruct: Aligning Language Model with Self Generated Instructions
📝https://github.com/databrickslabs/dolly
📝https://github.com/databrickslabs/dolly
GitHub
GitHub - databrickslabs/dolly: Databricks’ Dolly, a large language model trained on the Databricks Machine Learning Platform
Databricks’ Dolly, a large language model trained on the Databricks Machine Learning Platform - databrickslabs/dolly
CAMEL: Communicative Agents for "Mind" Exploration of Large Scale Language Model Society
📝https://github.com/lightaime/camel
📝https://github.com/lightaime/camel
GitHub
GitHub - camel-ai/camel: 🐫 CAMEL: The first and the best multi-agent framework. Finding the Scaling Law of Agents. https://www.camel…
🐫 CAMEL: The first and the best multi-agent framework. Finding the Scaling Law of Agents. https://www.camel-ai.org - camel-ai/camel
AGIEval: A Human-Centric Benchmark for Evaluating Foundation Models
📝https://github.com/microsoft/agieval
📝https://github.com/microsoft/agieval
GitHub
GitHub - microsoft/AGIEval
Contribute to microsoft/AGIEval development by creating an account on GitHub.
A Method for Animating Children's Drawings of the Human Figure
📝https://github.com/facebookresearch/AnimatedDrawings
📝https://github.com/facebookresearch/AnimatedDrawings
GitHub
GitHub - facebookresearch/AnimatedDrawings: Code to accompany "A Method for Animating Children's Drawings of the Human Figure"
Code to accompany "A Method for Animating Children's Drawings of the Human Figure" - facebookresearch/AnimatedDrawings
OpenAssistant Conversations -- Democratizing Large Language Model Alignment
📝https://github.com/laion-ai/open-assistant
📝https://github.com/laion-ai/open-assistant
GitHub
GitHub - LAION-AI/Open-Assistant: OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party…
OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so. - LAION-AI/Open-Assistant
HuaTuo: Tuning LLaMA Model with Chinese Medical Knowledge
📝https://github.com/scir-hi/huatuo-llama-med-chinese
📝https://github.com/scir-hi/huatuo-llama-med-chinese
GitHub
GitHub - SCIR-HI/Huatuo-Llama-Med-Chinese: Repo for BenTsao [original name: HuaTuo (华驼)], Instruction-tuning Large Language Models…
Repo for BenTsao [original name: HuaTuo (华驼)], Instruction-tuning Large Language Models with Chinese Medical Knowledge. 本草(原名:华驼)模型仓库,基于中文医学知识的大语言模型指令微调 - GitHub - SCIR-HI/Huatuo-Llama-Med-Chinese:...