Data Science | Machine Learning with Python for Researchers
31.4K subscribers
1.54K photos
102 videos
22 files
1.81K links
Admin: @HusseinSheikho

The Data Science and Python channel is for researchers and advanced programmers

Buy ads: https://telega.io/c/dataScienceT
Download Telegram
🐼 PandaLM: ReProducible and Automated Language Model Assessment

Judge large language model, named PandaLM, which is trained to distinguish the superior model given several LLMs. PandaLM's focus extends beyond just the objective correctness of responses, which is the main focus of traditional evaluation datasets.

πŸ–₯ Github: https://github.com/weopenml/pandalm

πŸ“• Paper: https://arxiv.org/abs/2306.05087v1

πŸ”— Dataset: https://github.com/tatsu-lab/stanford_alpaca#data-release

https://t.iss.one/DataScienceT
❀‍πŸ”₯2
This media is not supported in your browser
VIEW IN TELEGRAM
πŸ“Ή Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding

LLaMA is working on empowering large language models with video and audio understanding capability.

πŸ–₯ Github: https://github.com/damo-nlp-sg/video-llama

πŸ“• Paper: https://arxiv.org/abs/2306.02858

⏩ Demo: https://huggingface.co/spaces/DAMO-NLP-SG/Video-LLaMA

πŸ“Œ Model: https://modelscope.cn/studios/damo/video-llama/summary

https://t.iss.one/DataScienceT
❀‍πŸ”₯3πŸ‘3πŸ†1
A list of the best Telegram channels related to data science, programming languages, and artificial intelligence.

Join Quickly:
https://t.iss.one/addlist/8_rRW2scgfRhOTc0
❀‍πŸ”₯3
πŸ”οΈ Large Language Model for Geoscience

We introduce K2 (7B), an open-source language model trained by firstly further pretraining LLaMA on collected and cleaned geoscience literature, including geoscience open-access papers and Wikipedia pages, and secondly fine-tuning with knowledge-intensive instruction tuning data (GeoSignal).

git clone https://github.com/davendw49/k2.git
cd k2
conda env create -f k2.yml
conda activate k2


πŸ–₯ Github: https://github.com/davendw49/k2

⭐️ Demo: https://huggingface.co/daven3/k2_fp_delta

πŸ“• Paper: https://arxiv.org/abs/2306.05064v1

πŸ”— Dataset: https://huggingface.co/datasets/daven3/geosignal

https://t.iss.one/DataScienceT
❀‍πŸ”₯4πŸ‘2
πŸ’² FinGPT: Open-Source Financial Large Language Models

Unlike proprietary models, FinGPT takes a data-centric approach, providing researchers and practitioners with accessible and transparent resources to develop their FinLLMs.

πŸ–₯ Github: https://github.com/ai4finance-foundation/fingpt

⭐️ FinNLP: https://github.com/ai4finance-foundation/finnlp

πŸ“• Paper: https://arxiv.org/abs/2306.06031v1

πŸ”— Project: https://ai4finance-foundation.github.io/FinNLP/

https://t.iss.one/DataScienceT
❀‍πŸ”₯4πŸ‘4❀1
You can now download and watch all paid data science courses for free by subscribing to our new channel

https://t.iss.one/udemy13
πŸ‘2❀‍πŸ”₯1
This media is not supported in your browser
VIEW IN TELEGRAM
πŸ§” 4DHumans: Reconstructing and Tracking Humans with Transformers

Fully "transformerized" version of a network for human mesh recovery.

πŸ–₯ Github: https://github.com/shubham-goel/4D-Humans

⭐️ Colab: https://colab.research.google.com/drive/1Ex4gE5v1bPR3evfhtG7sDHxQGsWwNwby?usp=sharing

πŸ“• Paper: https://arxiv.org/pdf/2305.20091.pdf

πŸ”— Project: https://shubham-goel.github.io/4dhumans/

https://t.iss.one/DataScienceT
❀‍πŸ”₯2
πŸ”₯ Scalable Diffusion Models with Transformers (DiT)

git clone https://github.com/facebookresearch/DiT.git

πŸ–₯ Github: https://github.com/facebookresearch/DiT

πŸ–₯ Colab: https://colab.research.google.com/github/facebookresearch/DiT/blob/main/run_DiT.ipynb

⭐️ Project: https://www.wpeebles.com/DiT

⏩ Paprer: https://arxiv.org/abs/2212.09748

βœ”οΈ Dataset: https://paperswithcode.com/dataset/imagenet

https://t.iss.one/DataScienceT
❀‍πŸ”₯3πŸ‘1
Galactic: Scaling End-to-End Reinforcement Learning for Rearrangement
at 100k Steps-Per-Second

πŸ–₯ Github: https://github.com/facebookresearch/galactic

⏩ Paper: https://arxiv.org/pdf/2306.07552v1.pdf

πŸ’¨ Dataset: https://paperswithcode.com/dataset/vizdoom

https://t.iss.one/DataScienceT
❀‍πŸ”₯4❀1
Macaw-LLM: Multi-Modal Language Modeling with Image, Audio, Video, and Text Integration

Macaw-LLM is a model of its kind, bringing together state-of-the-art models for processing visual, auditory, and textual information, namely CLIP, Whisper, and LLaMA.

πŸ–₯ Github: https://github.com/lyuchenyang/macaw-llm

⭐️ Model: https://tinyurl.com/yem9m4nf

πŸ“• Paper: https://tinyurl.com/4rsexudv

πŸ”— Dataset: https://github.com/lyuchenyang/Macaw-LLM/blob/main/data

https://t.iss.one/DataScienceT
πŸ‘4❀2❀‍πŸ”₯1
Semi-supervised learning made simple with self-supervised clustering [CVPR 2023]

πŸ–₯ Github: https://github.com/pietroastolfi/suave-daino

⏩ Paper: https://arxiv.org/pdf/2306.07483v1.pdf

πŸ’¨ Dataset: https://paperswithcode.com/dataset/imagenet

https://t.iss.one/DataScienceT
❀‍πŸ”₯2❀1πŸ‘1
This media is not supported in your browser
VIEW IN TELEGRAM
🌐 WizMap: Scalable Interactive Visualization for Exploring Large Machine Learning Embeddings

πŸ–₯ Github: https://github.com/poloclub/wizmap

⭐️ Colab: https://colab.research.google.com/drive/1GNdmBnc5UA7OYBZPtHu244eiAN-0IMZA?usp=sharing

πŸ“• Paper: https://arxiv.org/abs/2306.09328v1

πŸ”— Web demo: https://poloclub.github.io/wizmap.

https://t.iss.one/DataScienceT
❀1❀‍πŸ”₯1πŸ‘1πŸ†1
How do Transformers work?

All
the Transformer models mentioned above (GPT, BERT, BART, T5, etc.) have been trained as language models. This means they have been trained on large amounts of raw text in a self-supervised fashion. Self-supervised learning is a type of training in which the objective is automatically computed from the inputs of the model. That means that humans are not needed to label the data!

This type of model develops a statistical understanding of the language it has been trained on, but it’s not very useful for specific practical tasks. Because of this, the general pretrained model then goes through a process called transfer learning. During this process, the model is fine-tuned in a supervised way β€” that is, using human-annotated labels β€” on a given task

πŸ”— Read More

🌸 https://t.iss.one/DataScienceT
πŸ‘3❀2❀‍πŸ”₯2