This media is not supported in your browser
VIEW IN TELEGRAM
Adding TTT layers into a pre-trained Transformer enables generating a one-minute clip from text storyboards.
Videos, code & annotations released
#AI #VideoGeneration #MachineLearning #DeepLearning #Transformers #TTT #GenerativeAI
Please open Telegram to view this post
VIEW IN TELEGRAM
π3π₯°2
π The new HQ-SAM (High-Quality Segment Anything Model) has just been added to the Hugging Face Transformers library!
This is an enhanced version of the original SAM (Segment Anything Model) introduced by Meta in 2023. HQ-SAM significantly improves the segmentation of fine and detailed objects, while preserving all the powerful features of SAM β including prompt-based interaction, fast inference, and strong zero-shot performance. That means you can easily switch to HQ-SAM wherever you used SAM!
The improvements come from just a few additional learnable parameters. The authors collected a high-quality dataset with 44,000 fine-grained masks from various sources, and impressively trained the model in just 4 hours using 8 GPUs β all while keeping the core SAM weights frozen.
The newly introduced parameters include:
* A High-Quality Token
* A Global-Local Feature Fusion mechanism
This work was presented at NeurIPS 2023 and still holds state-of-the-art performance in zero-shot segmentation on the SGinW benchmark.
π Documentation: https://lnkd.in/e5iDT6Tf
π§ Model Access: https://lnkd.in/ehS6ZUyv
π» Source Code: https://lnkd.in/eg5qiKC2
#ArtificialIntelligence #ComputerVision #Transformers #Segmentation #DeepLearning #PretrainedModels #ResearchAndDevelopment #AdvancedModels #ImageAnalysis #HQ_SAM #SegmentAnything #SAMmodel #ZeroShotSegmentation #NeurIPS2023 #AIresearch #FoundationModels #OpenSourceAI #SOTA
πhttps://t.iss.one/DataScienceN
This is an enhanced version of the original SAM (Segment Anything Model) introduced by Meta in 2023. HQ-SAM significantly improves the segmentation of fine and detailed objects, while preserving all the powerful features of SAM β including prompt-based interaction, fast inference, and strong zero-shot performance. That means you can easily switch to HQ-SAM wherever you used SAM!
The improvements come from just a few additional learnable parameters. The authors collected a high-quality dataset with 44,000 fine-grained masks from various sources, and impressively trained the model in just 4 hours using 8 GPUs β all while keeping the core SAM weights frozen.
The newly introduced parameters include:
* A High-Quality Token
* A Global-Local Feature Fusion mechanism
This work was presented at NeurIPS 2023 and still holds state-of-the-art performance in zero-shot segmentation on the SGinW benchmark.
π Documentation: https://lnkd.in/e5iDT6Tf
π§ Model Access: https://lnkd.in/ehS6ZUyv
π» Source Code: https://lnkd.in/eg5qiKC2
#ArtificialIntelligence #ComputerVision #Transformers #Segmentation #DeepLearning #PretrainedModels #ResearchAndDevelopment #AdvancedModels #ImageAnalysis #HQ_SAM #SegmentAnything #SAMmodel #ZeroShotSegmentation #NeurIPS2023 #AIresearch #FoundationModels #OpenSourceAI #SOTA
πhttps://t.iss.one/DataScienceN
lnkd.in
LinkedIn
This link will take you to a page thatβs not on LinkedIn
β€2π2π₯1
π₯ Trending Repository: haystack
π Description: AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
π Repository URL: https://github.com/deepset-ai/haystack
π Website: https://haystack.deepset.ai
π Readme: https://github.com/deepset-ai/haystack#readme
π Statistics:
π Stars: 22.3K stars
π Watchers: 158
π΄ Forks: 2.3K forks
π» Programming Languages: Python - HTML
π·οΈ Related Topics:
==================================
π§ By: https://t.iss.one/DataScienceM
π Description: AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
π Repository URL: https://github.com/deepset-ai/haystack
π Website: https://haystack.deepset.ai
π Readme: https://github.com/deepset-ai/haystack#readme
π Statistics:
π Stars: 22.3K stars
π Watchers: 158
π΄ Forks: 2.3K forks
π» Programming Languages: Python - HTML
π·οΈ Related Topics:
#python #nlp #agent #machine_learning #information_retrieval #ai #transformers #orchestration #pytorch #gemini #question_answering #summarization #agents #semantic_search #rag #gpt_4 #large_language_models #llm #generative_ai #retrieval_augmented_generation
==================================
π§ By: https://t.iss.one/DataScienceM
π₯ Trending Repository: LLaMA-Factory
π Description: Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
π Repository URL: https://github.com/hiyouga/LLaMA-Factory
π Website: https://llamafactory.readthedocs.io
π Readme: https://github.com/hiyouga/LLaMA-Factory#readme
π Statistics:
π Stars: 61.3K stars
π Watchers: 295
π΄ Forks: 7.4K forks
π» Programming Languages: Python
π·οΈ Related Topics:
==================================
π§ By: https://t.iss.one/DataScienceM
π Description: Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
π Repository URL: https://github.com/hiyouga/LLaMA-Factory
π Website: https://llamafactory.readthedocs.io
π Readme: https://github.com/hiyouga/LLaMA-Factory#readme
π Statistics:
π Stars: 61.3K stars
π Watchers: 295
π΄ Forks: 7.4K forks
π» Programming Languages: Python
π·οΈ Related Topics:
#nlp #agent #ai #transformers #moe #llama #gpt #lora #quantization #gemma #fine_tuning #peft #large_language_models #llm #rlhf #instruction_tuning #qlora #qwen #deepseek #llama3
==================================
π§ By: https://t.iss.one/DataScienceM