DL in NLP links
@dlinnlp_links
1.06K
subscribers
5
photos
1
file
653
links
AI and DeepLearning news/articles links I use for
@dlinnlp
posts
Download Telegram
Join
DL in NLP links
1.06K subscribers
DL in NLP links
https://openreview.net/forum?id=n9xeGcI4Yg
OpenReview
The Consensus Game: Language Model Generation via Equilibrium Search
When applied to question answering and other text generation tasks, language models (LMs) may be queried generatively (by sampling answers from their output distribution) or discriminatively (by...
DL in NLP links
https://x.com/davidbau/status/1790218863101231459?s=12&t=QgBLS4SmhE8cqdYBmhrqJA
X (formerly Twitter)
David Bau (@ ICLR) (@davidbau) on X
Apply by May 26 if excited by
• High-throughput parallelism for **transparent** LLM inference
• Service autoscaling for public science*on NCSA HPC GPUs
• Robustness+monitoring for safe use
• A mission to unlock broad access to frontier AI research
http…
DL in NLP links
https://arxiv.org/abs/2405.09215
arXiv.org
Xmodel-VLM: A Simple Baseline for Multimodal Vision Language Model
We introduce Xmodel-VLM, a cutting-edge multimodal vision language model. It is designed for efficient deployment on consumer GPU servers. Our work directly confronts a pivotal industry issue by...
DL in NLP links
https://x.com/wenhuchen/status/1790597967319007564?s=12&t=QgBLS4SmhE8cqdYBmhrqJA
X (formerly Twitter)
Wenhu Chen (@WenhuChen) on X
Tired of MMLU? The current models already hit the ceiling? It's time to upgrade MMLU!
Introducing our new benchmark MMLU-Pro, a more robust and challenging massive multi-task language understanding benchmark with 12K questions.
What's New?
1. MMLU-Pro uses…
👍
1
DL in NLP links
https://arxiv.org/abs/2307.06440
arXiv.org
No Train No Gain: Revisiting Efficient Training Algorithms For...
The computation necessary for training Transformer-based language models has skyrocketed in recent years. This trend has motivated research on efficient training algorithms designed to improve...
DL in NLP links
https://x.com/rasbt/status/1794355934678122921?s=12&t=QgBLS4SmhE8cqdYBmhrqJA
X (formerly Twitter)
Sebastian Raschka (@rasbt) on X
It's always exciting when a new paper with a LoRA-like method for efficient LLM finetuning comes out. In "MoRA: High-Rank Updating for Parameter-Efficient Finetuning (https://t.co/ieQ3uxWpqM)," the authors take a related yet opposite approach to low-rank…
👍
1
DL in NLP links
https://arxiv.org/pdf/2405.15682
DL in NLP links
https://arxiv.org/abs/2312.07533
arXiv.org
VILA: On Pre-training for Visual Language Models
Visual language models (VLMs) rapidly progressed with the recent success of large language models. There have been growing efforts on visual instruction tuning to extend the LLM with visual...
DL in NLP links
https://x.com/tsitsulin_/status/1795991374308745627?s=12&t=QgBLS4SmhE8cqdYBmhrqJA
DL in NLP links
https://x.com/yampeleg/status/1796697170101379509?s=12&t=QgBLS4SmhE8cqdYBmhrqJA
😁
2
DL in NLP links
https://arxiv.org/abs/2405.16684#:~:text=Past%20work%20has%20established%20scaling,of%20a%20fixed%20compute%20budget
.
arXiv.org
gzip Predicts Data-dependent Scaling Laws
Past work has established scaling laws that predict the performance of a neural language model (LM) as a function of its parameter count and the number of tokens it's trained on, enabling optimal...
DL in NLP links
https://www.youtube.com/embed/jGOwBQOJA-4
YouTube
Exclusive Interview With Unitree Co-founder "Unitree G1 - Humanoid agent AI avatar"
#humanoide #robots #robotics #ai #unitree
DL in NLP links
https://x.com/keyonv/status/1803838591371555252?s=12&t=QgBLS4SmhE8cqdYBmhrqJA
DL in NLP links
https://huggingface.co/papers/2406.16260
huggingface.co
Paper page - Video-Infinity: Distributed Long Video Generation
Join the discussion on this paper page
DL in NLP links
https://x.com/imbue_ai/status/1805629542914211951?s=12&t=QgBLS4SmhE8cqdYBmhrqJA
DL in NLP links
https://situational-awareness.ai/
SITUATIONAL AWARENESS - The Decade Ahead
Introduction - SITUATIONAL AWARENESS: The Decade Ahead
Leopold Aschenbrenner, June 2024 You can see the future first in San Francisco. Over the past year, the talk of the town has shifted from $10 billion compute clusters to $100 billion clusters to trillion-dollar clusters. Every six months another zero is added…
👍
2
❤
1
🔥
1
🤡
1
DL in NLP links
https://x.com/iscienceluvr/status/1805431835914346933?s=12&t=QgBLS4SmhE8cqdYBmhrqJA
DL in NLP links
https://x.com/margs_li/status/1806788050904924611?s=12&t=QgBLS4SmhE8cqdYBmhrqJA
DL in NLP links
https://x.com/_chen_lu_/status/1806762271559016842?s=12&t=QgBLS4SmhE8cqdYBmhrqJA
DL in NLP links
https://x.com/christopher/status/1811406837675163998?s=12&t=QgBLS4SmhE8cqdYBmhrqJA
DL in NLP links
https://x.com/apaszke/status/1812897008031617493?s=12&t=QgBLS4SmhE8cqdYBmhrqJA