firecrawl
Turn entire websites into LLM-ready markdown or structured data. Scrape, crawl and extract with a single API.
Creator: Mendable
Stars βοΈ: 12.3k
Forked By: 861
GitHub Repo:
https://github.com/mendableai/firecrawl
β
https://t.iss.one/deep_learning_proj
Turn entire websites into LLM-ready markdown or structured data. Scrape, crawl and extract with a single API.
Creator: Mendable
Stars βοΈ: 12.3k
Forked By: 861
GitHub Repo:
https://github.com/mendableai/firecrawl
Please open Telegram to view this post
VIEW IN TELEGRAM
GitHub
GitHub - mendableai/firecrawl: π₯ Turn entire websites into LLM-ready markdown or structured data. Scrape, crawl and extract withβ¦
π₯ Turn entire websites into LLM-ready markdown or structured data. Scrape, crawl and extract with a single API. - mendableai/firecrawl
π2
MiniCPM-V
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
Creator: OpenBMB
Stars βοΈ: 11.4k
Forked By: 798
GitHub Repo:
https://github.com/OpenBMB/MiniCPM-V
ββββββββββββββ
Joinβ
https://t.iss.one/deep_learning_proj
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
Creator: OpenBMB
Stars βοΈ: 11.4k
Forked By: 798
GitHub Repo:
https://github.com/OpenBMB/MiniCPM-V
ββββββββββββββ
Join
Please open Telegram to view this post
VIEW IN TELEGRAM
GitHub
GitHub - OpenBMB/MiniCPM-o: MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone
MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone - OpenBMB/MiniCPM-o
LLM based Multi-Agent methods
π₯ Github: https://github.com/AgnostiqHQ/multi-agent-llm
π Paper: https://arxiv.org/abs/2409.12618v1
π€ Dataset: https://paperswithcode.com/dataset/hotpotqa
β
https://t.iss.one/deep_learning_proj
π€ Dataset: https://paperswithcode.com/dataset/hotpotqa
Please open Telegram to view this post
VIEW IN TELEGRAM
GitHub
GitHub - AgnostiqHQ/multi-agent-llm: Lean implementation of various multi-agent LLM methods, including Iteration of Thought (IoT)
Lean implementation of various multi-agent LLM methods, including Iteration of Thought (IoT) - AgnostiqHQ/multi-agent-llm
Please open Telegram to view this post
VIEW IN TELEGRAM
llama-stack
Model components of the Llama Stack APIs
Creator: Meta Llama
Stars βοΈ: 1.5k
Forked By: 137
https://github.com/meta-llama/llama-stack
β
https://t.iss.one/deep_learning_proj
Model components of the Llama Stack APIs
Creator: Meta Llama
Stars βοΈ: 1.5k
Forked By: 137
https://github.com/meta-llama/llama-stack
Please open Telegram to view this post
VIEW IN TELEGRAM
GitHub
GitHub - meta-llama/llama-stack: Composable building blocks to build Llama Apps
Composable building blocks to build Llama Apps. Contribute to meta-llama/llama-stack development by creating an account on GitHub.
Crawl 4 AI
Crawl4AI: Open-source LLM Friendly Web Crawler & Scrapper
Creator: UncleCode
Stars βοΈ: 8.6k
Forked By: 627
https://github.com/unclecode/crawl4ai
β
https://t.iss.one/deep_learning_proj
Crawl4AI: Open-source LLM Friendly Web Crawler & Scrapper
Creator: UncleCode
Stars βοΈ: 8.6k
Forked By: 627
https://github.com/unclecode/crawl4ai
Please open Telegram to view this post
VIEW IN TELEGRAM
GitHub
GitHub - unclecode/crawl4ai: ππ€ Crawl4AI: Open-source LLM Friendly Web Crawler & Scraper. Don't be shy, join here: https://disβ¦
ππ€ Crawl4AI: Open-source LLM Friendly Web Crawler & Scraper. Don't be shy, join here: https://discord.gg/jP8KfhDhyN - unclecode/crawl4ai
π₯ NVIDIA silently release a Llama 3.1 70B fine-tune that outperforms
GPT-4o and Claude Sonnet 3.5
Llama 3.1 Nemotron 70B Instruct a further RLHFed model on
huggingface
https://huggingface.co/collections/nvidia/llama-31-nemotron-70b-670e93cd366feea16abc13d8
β
https://t.iss.one/deep_learning_proj
GPT-4o and Claude Sonnet 3.5
Llama 3.1 Nemotron 70B Instruct a further RLHFed model on
huggingface
https://huggingface.co/collections/nvidia/llama-31-nemotron-70b-670e93cd366feea16abc13d8
Please open Telegram to view this post
VIEW IN TELEGRAM
Π ΡΠ΅ΠΌΠ΅ΠΉΡΡΠ²Π΅ 2 ΠΌΠΎΠ΄Π΅Π»ΠΈ:
# Clone repo
git clone https://github.com/Zyphra/transformers_zamba2.git
cd transformers_zamba2
# Install the repository & accelerate:
pip install -e .
pip install accelerate
# Inference:
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
tokenizer = AutoTokenizer.from_pretrained("Zyphra/Zamba2-2.7B-instruct")
model = AutoModelForCausalLM.from_pretrained("Zyphra/Zamba2-2.7B-instruct", device_map="cuda", torch_dtype=torch.bfloat16)
user_turn_1 = "user_prompt1."
assistant_turn_1 = "assistant_prompt."
user_turn_2 = "user_prompt2."
sample = [{'role': 'user', 'content': user_turn_1}, {'role': 'assistant', 'content': assistant_turn_1}, {'role': 'user', 'content': user_turn_2}]
chat_sample = tokenizer.apply_chat_template(sample, tokenize=False)
input_ids = tokenizer(chat_sample, return_tensors='pt', add_special_tokens=False).to("cuda")
outputs = model.generate(**input_ids, max_new_tokens=150, return_dict_in_generate=False, output_scores=False, use_cache=True, num_beams=1, do_sample=False)
print((tokenizer.decode(outputs[0])))
https://t.iss.one/deep_learning_proj
Please open Telegram to view this post
VIEW IN TELEGRAM
π2
https://t.iss.one/deep_learning_proj
Please open Telegram to view this post
VIEW IN TELEGRAM
π3
LLM-based agents for Software Engineering
"Large Language Model-Based Agents for Software Engineering: A Survey".
https://github.com/FudanSELab/Agent4SE-Paper-List.
https://t.iss.one/deep_learning_proj
"Large Language Model-Based Agents for Software Engineering: A Survey".
https://github.com/FudanSELab/Agent4SE-Paper-List.
https://t.iss.one/deep_learning_proj
Welcome to Ollama's Prompt Engineering Interactive Tutorial
π Github
https://t.iss.one/deep_learning_proj
π Github
https://t.iss.one/deep_learning_proj
π3
Forwarded from Machine learning books and papers
@Machine_learn
Please open Telegram to view this post
VIEW IN TELEGRAM
Forwarded from Machine learning books and papers
NVIDIA BioNeMo2 Framework is a set of tools, libraries, and models for computational drug discovery and design.
@Machine_learn
Please open Telegram to view this post
VIEW IN TELEGRAM
π2
Forwarded from Machine learning books and papers
Large Language Models Course: Learn by Doing LLM Projects
π₯ Github: https://github.com/peremartra/Large-Language-Model-Notebooks-Course
π Paper: https://doi.org/10.31219/osf.io/qgxea
@Machine_learn
@Machine_learn
Please open Telegram to view this post
VIEW IN TELEGRAM
Forwarded from Machine learning books and papers
Foundations of Large Language Models (1).pdf
1.9 MB
Foundations of Large Language Models
π Table of Contents:
β Pre-training
β Generative Models
β Prompting
β Alignment
Tong Xiao and Jingbo Zhu
January 17, 2025
π Download from arXiv.
@Machine_learn
π Table of Contents:
β Pre-training
β Generative Models
β Prompting
β Alignment
Tong Xiao and Jingbo Zhu
January 17, 2025
π Download from arXiv.
@Machine_learn
π1
Evolutionary Computation in the Era of Large Language Model: Survey and Roadmap
Large language models (LLMs) have not only revolutionized natural language processing but also extended their prowess to various domains, marking a significant stride towards artificial general intelligence. The interplay between LLMs and evolutionary algorithms (EAs), despite differing in objectives and methodologies, share a common pursuit of applicability in complex problems. Meanwhile, EA can provide an optimization framework for LLM's further enhancement under black-box settings, empowering LLM with flexible global search capacities. On the other hand, the abundant domain knowledge inherent in LLMs could enable EA to conduct more intelligent searches. Furthermore, the text processing and generative capabilities of LLMs would aid in deploying EAs across a wide range of tasks. Based on these complementary advantages, this paper provides a thorough review and a forward-looking roadmap, categorizing the reciprocal inspiration into two main avenues: LLM-enhanced EA and EA-enhanced #LLM. Some integrated synergy methods are further introduced to exemplify the complementarity between LLMs and EAs in diverse scenarios, including code generation, software engineering, neural architecture search, and various generation tasks. As the first comprehensive review focused on the EA research in the era of #LLMs, this paper provides a foundational stepping stone for understanding the collaborative potential of LLMs and EAs. The identified challenges and future directions offer guidance for researchers and practitioners to unlock the full potential of this innovative collaboration in propelling advancements in optimization and artificial intelligence.
Paper: https://arxiv.org/pdf/2401.10034v3.pdf
Code: https://github.com/wuxingyu-ai/llm4ec
https://t.iss.one/deep_learning_proj
Large language models (LLMs) have not only revolutionized natural language processing but also extended their prowess to various domains, marking a significant stride towards artificial general intelligence. The interplay between LLMs and evolutionary algorithms (EAs), despite differing in objectives and methodologies, share a common pursuit of applicability in complex problems. Meanwhile, EA can provide an optimization framework for LLM's further enhancement under black-box settings, empowering LLM with flexible global search capacities. On the other hand, the abundant domain knowledge inherent in LLMs could enable EA to conduct more intelligent searches. Furthermore, the text processing and generative capabilities of LLMs would aid in deploying EAs across a wide range of tasks. Based on these complementary advantages, this paper provides a thorough review and a forward-looking roadmap, categorizing the reciprocal inspiration into two main avenues: LLM-enhanced EA and EA-enhanced #LLM. Some integrated synergy methods are further introduced to exemplify the complementarity between LLMs and EAs in diverse scenarios, including code generation, software engineering, neural architecture search, and various generation tasks. As the first comprehensive review focused on the EA research in the era of #LLMs, this paper provides a foundational stepping stone for understanding the collaborative potential of LLMs and EAs. The identified challenges and future directions offer guidance for researchers and practitioners to unlock the full potential of this innovative collaboration in propelling advancements in optimization and artificial intelligence.
Paper: https://arxiv.org/pdf/2401.10034v3.pdf
Code: https://github.com/wuxingyu-ai/llm4ec
https://t.iss.one/deep_learning_proj
Forwarded from Machine learning books and papers
ChatGPT Cheat Sheet for Business (2025).pdf
8 MB
π3