Data Science | Machine Learning with Python for Researchers
32.7K subscribers
3.33K photos
126 videos
23 files
3.55K links
ads: @HusseinSheikho

The Data Science and Python channel is for researchers and advanced programmers

Buy ads: https://telega.io/c/dataScienceT
Download Telegram
Story-Adapter: A Training-free Iterative Framework for Long Story Visualization

Paper: https://arxiv.org/pdf/2410.06244v1.pdf

Code: https://github.com/jwmao1/story-adapter

https://t.iss.one/DataScienceT 📊
Please open Telegram to view this post
VIEW IN TELEGRAM
👍3
Into the Unknown Unknowns: Engaged Human Learning through Participation in Language Model Agent Conversations

Paper: https://arxiv.org/pdf/2408.15232v2.pdf

Code: https://github.com/stanford-oval/storm

https://t.iss.one/DataScienceT ⚠️
Please open Telegram to view this post
VIEW IN TELEGRAM
👍4
Dispider: Enabling Video LLMs with Active Real-Time Interaction via Disentangled Perception, Decision, and Reaction

Paper: https://arxiv.org/pdf/2501.03218v1.pdf

Code: https://github.com/mark12ding/dispider

Dataset: Video-MME

https://t.iss.one/DataScienceT 🎙
Please open Telegram to view this post
VIEW IN TELEGRAM
👍3
SHMT: Self-supervised Hierarchical Makeup Transfer via Latent Diffusion Models

Paper: https://arxiv.org/pdf/2412.11058v1.pdf

Code:
https://github.com/snowfallingplum/shmt
https://github.com/snowfallingplum/csd-mt

https://t.iss.one/DataScienceT 🐍#️⃣
Please open Telegram to view this post
VIEW IN TELEGRAM
AXIAL: Attention-based eXplainability for Interpretable Alzheimer's Localized Diagnosis using 2D CNNs on 3D MRI brain scans

Paper: https://arxiv.org/pdf/2407.02418v2.pdf

Code: https://github.com/GabrieleLozupone/AXIAL

Dataset: ADNI

https://t.iss.one/DataScienceT 🧠
Please open Telegram to view this post
VIEW IN TELEGRAM
3👍3
Parameter-Inverted Image Pyramid Networks for Visual Perception and Multimodal Understanding

🖥 Github: https://github.com/opengvlab/piip

📕 Paper: https://arxiv.org/abs/2501.07783v1

⭐️ Dataset: https://paperswithcode.com/dataset/gqa

https://t.iss.one/DataScienceT 🧠
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
👍41
FramePainter: Endowing Interactive Image Editing with Video Diffusion Priors

Paper: https://arxiv.org/pdf/2501.08225v1.pdf

Code: https://github.com/ybybzhang/framepainter

https://t.iss.one/DataScienceT ✈️
Please open Telegram to view this post
VIEW IN TELEGRAM
Stretching Each Dollar: Diffusion Training from Scratch on a Micro-Budget

Paper: https://arxiv.org/pdf/2407.15811v1.pdf

code: https://github.com/sonyresearch/micro_diffusion

Datasets: MS COCO

https://t.iss.one/DataScienceT 🧠
Please open Telegram to view this post
VIEW IN TELEGRAM
1
MiniRAG: Towards Extremely Simple Retrieval-Augmented Generation

Paper: https://arxiv.org/pdf/2501.06713v2.pdf

Code: https://github.com/hkuds/minirag

https://t.iss.one/DataScienceT 🧠
Please open Telegram to view this post
VIEW IN TELEGRAM
3👍2
Continual Forgetting for Pre-trained Vision Models (CVPR2024)

🖥 Github: https://github.com/bjzhb666/GS-LoRA

📕 Paper: https://arxiv.org/abs/2501.09705v1

🧠 Dataset: https://paperswithcode.com/dataset/coco

https://t.iss.one/DataScienceT 🧠
Please open Telegram to view this post
VIEW IN TELEGRAM
1👍1
Please open Telegram to view this post
VIEW IN TELEGRAM
👍1
UnCommon Objects in 3D

We introduce Uncommon Objects in 3D (uCO3D), a new object-centric dataset for 3D deep learning and 3D generative AI. uCO3D is the largest publicly-available collection of high-resolution videos of objects with 3D annotations that ensures full-360 coverage. uCO3D is significantly more diverse than MVImgNet and CO3Dv2, covering more than 1,000 object categories. It is also of higher quality, due to extensive quality checks of both the collected videos and the 3D annotations. Similar to analogous datasets, uCO3D contains annotations for 3D camera poses, depth maps and sparse point clouds. In addition, each object is equipped with a caption and a 3D Gaussian Splat reconstruction. We train several large 3D models on MVImgNet, CO3Dv2, and uCO3D and obtain superior results using the latter, showing that uCO3D is better for learning applications.

Paper: https://arxiv.org/pdf/2501.07574v1.pdf

Code: https://github.com/facebookresearch/uco3d

DataSet: MS COCO

https://t.iss.one/DataScienceT 🐻‍❄️
Please open Telegram to view this post
VIEW IN TELEGRAM
3👍1
The GAN is dead; long live the GAN! A Modern GAN Baseline

There is a widely-spread claim that GANs are difficult to train, and GAN architectures in the literature are littered with empirical tricks. We provide evidence against this claim and build a modern GAN baseline in a more principled manner. First, we derive a well-behaved regularized relativistic GAN loss that addresses issues of mode dropping and non-convergence that were previously tackled via a bag of ad-hoc tricks. We analyze our loss mathematically and prove that it admits local convergence guarantees, unlike most existing relativistic losses. Second, our new loss allows us to discard all ad-hoc tricks and replace outdated backbones used in common GANs with modern architectures. Using StyleGAN2 as an example, we present a roadmap of simplification and modernization that results in a new minimalist baseline -- R3GAN. Despite being simple, our approach surpasses StyleGAN2 on FFHQ, ImageNet, CIFAR, and Stacked MNIST datasets, and compares favorably against state-of-the-art GANs and diffusion models.

Paper: https://arxiv.org/pdf/2501.05441v1.pdf

Code: https://github.com/brownvc/r3gan

Dataset: CIFAR-10

https://t.iss.one/DataScienceT 😵‍💫
Please open Telegram to view this post
VIEW IN TELEGRAM
4👍2
Cold-Start Recommendation towards the Era of Large Language Models (LLMs): A Comprehensive Survey and Roadmap

Cold-start problem is one of the long-standing challenges in recommender systems, focusing on accurately modeling new or interaction-limited users or items to provide better recommendations. Due to the diversification of internet platforms and the exponential growth of users and items, the importance of cold-start recommendation (CSR) is becoming increasingly evident. At the same time, large language models (LLMs) have achieved tremendous success and possess strong capabilities in modeling user and item information, providing new potential for cold-start recommendations. However, the research community on CSR still lacks a comprehensive review and reflection in this field. Based on this, in this paper, we stand in the context of the era of large language models and provide a comprehensive review and discussion on the roadmap, related literature, and future directions of CSR. Specifically, we have conducted an exploration of the development path of how existing CSR utilizes information, from content features, graph relations, and domain information, to the world knowledge possessed by large language models, aiming to provide new insights for both the research and industrial communities on CSR. Related resources of cold-start recommendations are collected and continuously updated for the community in https://github.com/YuanchenBei/Awesome-Cold-Start-Recommendation.

Paper: https://arxiv.org/pdf/2501.01945v2.pdf

Code: https://github.com/yuanchenbei/awesome-cold-start-recommendation

https://t.iss.one/DataScienceT 🩷
Please open Telegram to view this post
VIEW IN TELEGRAM
👍1
Evolutionary Computation in the Era of Large Language Model: Survey and Roadmap

Large language models (LLMs) have not only revolutionized natural language processing but also extended their prowess to various domains, marking a significant stride towards artificial general intelligence. The interplay between LLMs and evolutionary algorithms (EAs), despite differing in objectives and methodologies, share a common pursuit of applicability in complex problems. Meanwhile, EA can provide an optimization framework for LLM's further enhancement under black-box settings, empowering LLM with flexible global search capacities. On the other hand, the abundant domain knowledge inherent in LLMs could enable EA to conduct more intelligent searches. Furthermore, the text processing and generative capabilities of LLMs would aid in deploying EAs across a wide range of tasks. Based on these complementary advantages, this paper provides a thorough review and a forward-looking roadmap, categorizing the reciprocal inspiration into two main avenues: LLM-enhanced EA and EA-enhanced #LLM. Some integrated synergy methods are further introduced to exemplify the complementarity between LLMs and EAs in diverse scenarios, including code generation, software engineering, neural architecture search, and various generation tasks. As the first comprehensive review focused on the EA research in the era of #LLMs, this paper provides a foundational stepping stone for understanding the collaborative potential of LLMs and EAs. The identified challenges and future directions offer guidance for researchers and practitioners to unlock the full potential of this innovative collaboration in propelling advancements in optimization and artificial intelligence.

Paper: https://arxiv.org/pdf/2401.10034v3.pdf

Code: https://github.com/wuxingyu-ai/llm4ec

https://t.iss.one/DataScienceT ⭐️
Please open Telegram to view this post
VIEW IN TELEGRAM
👍2
Forwarded from Python Courses
Unlock a treasure trove of knowledge with our exclusive paid channel! For just $2 a month, gain access to thousands of valuable resources, including essential books and premium courses from Coursera and Udemy. Plus, dive into exciting paid projects! Enjoy hassle-free automatic payments via Telegram. Join us today! Link
👍2