Data Analytics
28.2K subscribers
1.24K photos
30 videos
44 files
1.08K links
Dive into the world of Data Analytics โ€“ uncover insights, explore trends, and master data-driven decision making.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
๐Ÿ”ฅ2026 New IT Certification Prep Kit โ€“ Free!

SPOTO cover: #Python #AI #Cisco #PMI #Fortinet #AWS #Azure #Excel #CompTIA #ITIL #Cloud + more

โœ… Grab yours free kit now:
โ€ข Free Courses (Python, Excel, Cyber Security, Cisco, SQL, ITIL, PMP, AWS)
๐Ÿ‘‰ https://bit.ly/3Ogtn3i
โ€ข IT Certs E-book
๐Ÿ‘‰ https://bit.ly/41KZlru
โ€ข IT Exams Skill Test
๐Ÿ‘‰ https://bit.ly/4ve6ZbC
โ€ข Free AI Materials & Support Tools
๐Ÿ‘‰ https://bit.ly/4vagTuw
โ€ข Free Cloud Study Guide
๐Ÿ‘‰ https://bit.ly/4c3BZCh

๐Ÿ’ฌ Need exam help? Contact admin: wa.link/w6cems

โœ… Join our IT community: get free study materials, exam tips & peer support
https://chat.whatsapp.com/BiazIVo5RxfKENBv10F444
โค2
Follow the Machine Learning with Python channel on WhatsApp: https://whatsapp.com/channel/0029VaC7Weq29753hpcggW2A
โค1
๐Ÿงน ๐—ฆ๐˜๐—ฒ๐—ฝ๐˜€ ๐˜๐—ผ ๐—–๐—น๐—ฒ๐—ฎ๐—ป ๐—ฎ ๐——๐—ฎ๐˜๐—ฎ๐˜€๐—ฒ๐˜ ๐—ถ๐—ป ๐—ฃ๐˜†๐˜๐—ต๐—ผ๐—ป

https://t.iss.one/DataAnalyticsX
โค6๐Ÿ‘2
๐Ÿš€ Sber has released two open-source MoE models: GigaChat-3.1 Ultra and Lightning

Both code and weights are available under the MIT license on HuggingFace.

๐Ÿ‘‰ Key details:

โ€ข Trained from scratch (not a finetune) on proprietary data and infrastructure
โ€ข Mixture-of-Experts (MoE) architecture

Models:

๐Ÿง  GigaChat-3.1 Ultra
โ€ข 702B MoE model for high-performance environments
โ€ข Outperforms DeepSeek-V3-0324 and Qwen3-235B on math and reasoning benchmarks
โ€ข Supports FP8 training and MTP

โšก๏ธ GigaChat-3.1 Lightning
โ€ข 10B model (1.8B active parameters)
โ€ข Outperforms Qwen3-4B and Gemma-3-4B on Sber benchmarks
โ€ข Efficient local inference
โ€ข Up to 256k context

Engineering highlights:

โ€ข Custom metric to detect and reduce generation loops
โ€ข DPO training moved to native FP8
โ€ข Improvements in post-training pipeline
โ€ข Identified and fixed a critical issue affecting evaluation quality

๐ŸŒ Trained on 14 languages (optimized for English and Russian)

Use cases:

โ€ข chatbots
โ€ข AI assistants
โ€ข copilots
โ€ข internal ML systems

Sber provides a solid open foundation for developers to build production-ready AI systems with lower infrastructure costs.
โค2
โœ”๏ธ 10 Books to Understand How Large Language Models Function (2026)

1. Deep Learning
https://deeplearningbook.org
The definitive reference for neural networks, covering backpropagation, architectures, and foundational concepts.

2. Artificial Intelligence: A Modern Approach
https://aima.cs.berkeley.edu
A fundamental perspective on artificial intelligence as a comprehensive system.

3. Speech and Language Processing
https://web.stanford.edu/~jurafsky/slp3/
An in-depth examination of natural language processing, transformers, and linguistics.

4. Machine Learning: A Probabilistic Perspective
https://probml.github.io/pml-book/
An exploration of probabilities, statistics, and the theoretical foundations of machine learning.

5. Understanding Deep Learning
https://udlbook.github.io/udlbook/
A contemporary explanation of deep learning principles with strong intuitive insights.

6. Designing Machine Learning Systems
https://oreilly.com/library/view/designing-machine-learning/9781098107956/
Strategies for deploying models into production environments.

7. Generative Deep Learning
https://github.com/3p5ilon/ML-books/blob/main/generative-deep-learning-teaching-machines-to-paint-write-compose-and-play.pdf
Practical applications of generative models and transformer architectures.

8. Natural Language Processing with Transformers
https://dokumen.pub/natural-language-processing-with-transformers-revised-edition-1098136799-9781098136796-9781098103248.html
Methodologies for constructing natural language processing systems based on transformers.

9. Machine Learning Engineering
https://mlebook.com
Principles of machine learning engineering and operational deployment.

10. The Hundred-Page Machine Learning Book
https://themlbook.com
A highly concentrated foundational overview without extraneous detail. ๐Ÿ“š๐Ÿค–
โค2
This channels is for Programmers, Coders, Software Engineers.

0๏ธโƒฃ Python
1๏ธโƒฃ Data Science
2๏ธโƒฃ Machine Learning
3๏ธโƒฃ Data Visualization
4๏ธโƒฃ Artificial Intelligence
5๏ธโƒฃ Data Analysis
6๏ธโƒฃ Statistics
7๏ธโƒฃ Deep Learning
8๏ธโƒฃ programming Languages

โœ… https://t.iss.one/addlist/8_rRW2scgfRhOTc0

โœ… https://t.iss.one/Codeprogrammer
Please open Telegram to view this post
VIEW IN TELEGRAM
โค2๐Ÿ‘1
๐Ÿ“ 12 Essential Articles for Data Scientists

๐Ÿท Article: Seq2Seq Learning with NN
https://arxiv.org/pdf/1409.3215
An introduction to Seq2Seq models, which serve as the foundation for machine translation utilizing deep learning.

๐Ÿท Article: GANs
https://arxiv.org/pdf/1406.2661
An introduction to Generative Adversarial Networks (GANs) and the concept of generating synthetic data. This forms the basis for creating images and videos with artificial intelligence.

๐Ÿท Article: Attention is All You Need
https://arxiv.org/pdf/1706.03762
This paper was revolutionary in natural language processing. It introduced the Transformer architecture, which underlies GPT, BERT, and contemporary intelligent language models.

๐Ÿท Article: Deep Residual Learning
https://arxiv.org/pdf/1512.03385
This work introduced the ResNet model, enabling neural networks to achieve greater depth and accuracy without compromising the learning process.

๐Ÿท Article: Batch Normalization
https://arxiv.org/pdf/1502.03167
This paper introduced a technique that facilitates faster and more stable training of neural networks.

๐Ÿท Article: Dropout
https://jmlr.org/papers/volume15/srivastava14a/srivastava14a.pdf
A straightforward method designed to prevent overfitting in neural networks.

๐Ÿท Article: ImageNet Classification with DCNN
https://proceedings.neurips.cc/paper_files/paper/2012/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf
The first successful application of a deep neural network for image recognition.

๐Ÿท Article: Support-Vector Machines
https://link.springer.com/content/pdf/10.1007/BF00994018.pdf
This seminal work introduced the Support Vector Machine (SVM) algorithm, a widely utilized method for data classification.

๐Ÿท Article: A Few Useful Things to Know About ML
https://homes.cs.washington.edu/~pedro/papers/cacm12.pdf
A comprehensive collection of practical and empirical insights regarding machine learning.

๐Ÿท Article: Gradient Boosting Machine
https://www.cse.iitb.ac.in/~soumen/readings/papers/Friedman1999GreedyFuncApprox.pdf
This paper introduced the "Gradient Boosting" method, which serves as the foundation for many modern machine learning models, including XGBoost and LightGBM.

๐Ÿท Article: Latent Dirichlet Allocation
https://jmlr.org/papers/volume3/blei03a/blei03a.pdf
This work introduced a model for text analysis capable of identifying the topics discussed within an article.

๐Ÿท Article: Random Forests
https://www.stat.berkeley.edu/~breiman/randomforest2001.pdf
This paper introduced the "Random Forest" algorithm, a powerful machine learning method that aggregates multiple models to achieve enhanced accuracy.

https://t.iss.one/CodeProgrammer ๐ŸŒŸ
Please open Telegram to view this post
VIEW IN TELEGRAM
โค3
๐Ÿš€ LLM Architectures ๐Ÿง 
Transformer architectures may look similar, but they solve very different problems once data starts flowing through them. ๐Ÿ”„

The four main Transformer families in simple terms. ๐Ÿ“š

๐Ÿ‘‰ Decoder-only models like GPT and LLaMA generate text one token at a time. Each new token looks only at previous tokens. This makes them great for chat, code generation, and text completion. ๐Ÿ’ฌ๐Ÿ’ป

๐Ÿ‘‰ Encoder-only models like BERT and RoBERTa focus on understanding text. Every token sees the full sentence at once. These models are used for classification, search, and extracting meaning rather than generating text. ๐Ÿ”๐Ÿ“–

๐Ÿ‘‰ Encoder-decoder models like T5 and BART first understand the input, then generate an output. This setup is common for translation, summarization, and question answering. ๐ŸŒ๐Ÿ“

๐Ÿ‘‰ Mixture of Experts (MoE) models like Mixtral and GLaM scale smarter, not harder. A router sends tokens to a small set of expert networks, allowing very large models to run efficiently. โšก๏ธ๐Ÿค–

Example:
Summarizing a document ๐Ÿ“„
- Decoder-only generates fluent text โœ๏ธ
- Encoder-only ranks important sentences ๐Ÿท
- Encoder-decoder produces a clean summary ๐Ÿงน
- MoE scales the process with lower compute cost ๐Ÿ’ฐ

Choosing the right Transformer matters more than choosing the largest one. โš–๏ธโœจ

https://t.iss.one/DataAnalyticsX ๐Ÿ”ฐ
Please open Telegram to view this post
VIEW IN TELEGRAM
โค17
๐€๐ณ๐ฎ๐ซ๐ž_๐ƒ๐š๐ญ๐š_๐„๐ง๐ ๐ข๐ง๐ž๐ž๐ซ.pdf
10.2 MB
Everyone wants to become a ๐ƒ๐š๐ญ๐š ๐„๐ง๐ ๐ข๐ง๐ž๐ž๐ซโ€ฆ ๐Ÿ“Š But very few follow a structured path. ๐Ÿ›ค

They keep learning random tools, watching endless tutorials and still feel unprepared. ๐Ÿคฏ

Meanwhile, some people are quietly transitioning into roles like:
๐Ÿ’ผ Azure Data Engineer
๐Ÿ’ผ Data Architect
๐Ÿ’ผ Senior Data Engineer

What are they doing differently? ๐Ÿค”
Theyโ€™re not doing more.
Theyโ€™re doing the right things consistently. โœจ

Hereโ€™s whatโ€™s working for them:
โœ”๏ธ A step-by-step Azure Data Engineering roadmap ๐Ÿ—บ
โœ”๏ธ Mastering SQL & Python (not just basics) ๐Ÿ’ป
โœ”๏ธ Hands-on with Azure tools (ADF, Synapse, Data Lake) โ˜๏ธ
โœ”๏ธ Building real-world, portfolio-ready projects ๐Ÿ—
โœ”๏ธ Preparing specifically for interviews ๐ŸŽฏ
โœ”๏ธ Learning with a focused community ๐Ÿค
Please open Telegram to view this post
VIEW IN TELEGRAM
โค6
Please open Telegram to view this post
VIEW IN TELEGRAM
โค2๐Ÿ”ฅ1
๐Ÿš€ Thrilled to announce a major milestone in our collective upskilling journey! ๐ŸŒŸ

I am incredibly excited to share a curated ecosystem of high-impact resources focused on Machine Learning and Artificial Intelligence. By consolidating a comprehensive library of PDFsโ€”from foundational onboarding to advanced strategic insightsโ€”into a single, unified repository, we are effectively eliminating search friction and accelerating our learning velocity. ๐Ÿ“šโœจ

This initiative represents a powerful opportunity to align our technical growth with future-ready priorities, ensuring we are always ahead of the curve. ๐Ÿ’ก๐Ÿ”—

โ›“๏ธ Unlock your potential here:
https://github.com/Ramakm/AI-ML-Book-References

#MachineLearning #AI #ContinuousLearning #GrowthMindset #TechCommunity #OpenSource
โค2
This bot will help you get a course that's available for free for a limited time so you can register before others.

Benefit from it
t.iss.one/UdemySybot
Tired of losing nights? With $100 start, Mr Pastore EA can target $500/day automatically - try the 7-day demo and breathe easy. Join now: Start

#ad ๐Ÿ“ข InsideAd
Please open Telegram to view this post
VIEW IN TELEGRAM
LLM Engineering Roadmap (2026 Practical Guide) ๐Ÿ—บโœจ

If your goal is to build real LLM apps (not just prompts), follow this order. ๐Ÿš€

1๏ธโƒฃ Python + APIs ๐Ÿ๐Ÿ”Œ

Youโ€™ll spend most of your time wiring systems.

Learn:
โ†’ functions, classes
โ†’ working with APIs (requests, JSON)
โ†’ async basics
โ†’ environment variables

Resources
โ†’ Python for Everybody
https://lnkd.in/gUqkvnGG
โ†’ Introduction to Python
https://lnkd.in/g7xfYJVZ
โ†’ MLTUT Python Basics Course
https://lnkd.in/gCqfyCGZ

2๏ธโƒฃ Text Basics (NLP) ๐Ÿ“๐Ÿง 

You donโ€™t need heavy theory, just the essentials.

Learn:
โ†’ tokenization
โ†’ text cleaning
โ†’ similarity (cosine)
โ†’ basic embeddings idea

Resources
โ†’ Natural Language Processing Specialization
https://lnkd.in/gz_xmqD9
โ†’ NLP in Python
https://lnkd.in/gnpcJxhz

3๏ธโƒฃ Transformers (Whatโ€™s happening behind the API) ๐Ÿค–๐Ÿ”

Enough to not treat it like a black box.

Learn:
โ†’ tokens, context window
โ†’ attention (high level)
โ†’ why embeddings work
โ†’ limits of LLMs

Resources
โ†’ Generative AI with Large Language Models
https://lnkd.in/gk3PPtyf
โ†’ Hugging Face Transformers Course
https://lnkd.in/ggSR5JNb

4๏ธโƒฃ Prompting (Make outputs reliable) ๐Ÿ’ฌ๐ŸŽฏ

Treat prompts like code.

Learn:
โ†’ few-shot examples
โ†’ structured outputs (JSON)
โ†’ system vs user instructions
โ†’ simple evals (does it break?)

Resources
โ†’ Prompt Engineering for ChatGPT
https://lnkd.in/gyg4EiJS
โ†’ Prompt Engineering with LLMs
https://lnkd.in/gn67Mxga

5๏ธโƒฃ Embeddings + Vector DBs ๐Ÿ“Š๐Ÿ—„

This is how you add your data.

Learn:
โ†’ embedding generation
โ†’ similarity search
โ†’ indexing
Tools:
โ†’ FAISS
โ†’ Pinecone
โ†’ Chroma

Resources
โ†’ Working with Embeddings
https://lnkd.in/gnngPW4E
โ†’ Vector Databases & Semantic Search
https://lnkd.in/gP2HdMmD

6๏ธโƒฃ RAG Pipelines ๐Ÿ”—๐Ÿ”„

Most useful apps use this pattern.

Learn:
โ†’ chunking documents
โ†’ retrieval + ranking
โ†’ prompt + context design
โ†’ basic evaluation

Resources
โ†’ Generative AI for Software Development
https://lnkd.in/g3uduecv
โ†’ Build RAG Apps with LangChain
https://lnkd.in/ggXJjgDN

7๏ธโƒฃ Build Real Applications ๐Ÿ› ๐Ÿ’ป

Keep them small and usable.

Build:
โ†’ document Q&A (PDF โ†’ answers)
โ†’ internal knowledge bot
โ†’ code assistant (repo Q&A)
โ†’ support chatbot

Tools:
โ†’ LangChain
โ†’ LlamaIndex
โ†’ OpenAI APIs

Resources
โ†’ Build LLM Apps with LangChain & Python
https://lnkd.in/g6xXVX_8
โ†’ LLM Applications
https://lnkd.in/gzs8_SRk

8๏ธโƒฃ Deployment ๐Ÿšขโ˜๏ธ

Make it usable by others.

Learn:
โ†’ FastAPI endpoints
โ†’ streaming responses
โ†’ caching (reduce cost)
โ†’ logging + monitoring

Tools:
โ†’ FastAPI
โ†’ Docker
โ†’ AWS / GCP

Resources
โ†’Machine Learning Engineering for Production (MLOps)
https://lnkd.in/gCMtYSk5
โ†’ MLOps Fundamentals
https://lnkd.in/g8TGrUzT

https://t.iss.one/DataAnalyticsX โœ…
Please open Telegram to view this post
VIEW IN TELEGRAM
โค2
Today, the public mint for Lobsters on TON goes live on Getgems ๐Ÿฆž

This is not just another NFT drop.
In my view, Lobsters is one of the first truly cohesive products at the intersection of blockchain, NFTs, and AI.

Here, the NFT is not just an image and not just a collectible.
Each Lobster is an NFT with a built-in AI agent inside: a digital character with its own soul, on-chain biography, persistent memory, and a unified identity across Telegram, Mini App, Claude, and API.

So you are not just getting an asset in your wallet.
You are getting an AI-native digital character that can interact, remember, and stay consistent across different interfaces.

What makes this especially interesting is the timing.

In the recent video Pavel Durov shared in his post about agentic bots in Telegram, the lobster imagery was right there. Against that backdrop, Lobsters does not feel like a random mint โ€” it feels like a very precise fit for the new narrative:

Telegram-native agents + TON infrastructure + NFT ownership layer + AI utility

Put simply, this is one of the first real attempts to turn an NFT from โ€œjust an imageโ€ into a digital agent.

Public mint: today, 16:00
Price: 50 TON

๐Ÿ‘‰ Mint your Lobster on Getgems ๐Ÿฆž๐Ÿฆž๐Ÿฆž
โค2