Kuldeep Singh Sidhu's picture
6 3

Kuldeep Singh Sidhu

singhsidhukuldeep

AI & ML interests

πŸ˜ƒ TOP 3 on HuggingFace for posts πŸ€— Seeking contributors for a completely open-source πŸš€ Data Science platform! singhsidhukuldeep.github.io

Recent Activity

posted an update 2 days ago
Breaking News: LinkedIn's Content Search Engine Gets a Powerful Semantic Upgrade! Excited to share insights about LinkedIn's innovative approach to content search, recently detailed in a groundbreaking paper by their Mountain View team. This advancement represents a significant shift from traditional keyword-based search to semantic understanding. >> Technical Architecture The new search engine employs a sophisticated two-layer architecture: Retrieval Layer - Token Based Retriever (TBR) for exact keyword matching - Embedding Based Retriever (EBR) using a two-tower model with multilingual-e5 embeddings - Pre-computed post embeddings stored in a dedicated embedding store for efficient retrieval Multi-Stage Ranking - L1 Stage: Initial filtering using a lightweight model - L2 Stage: Advanced ranking with complex features including: - Query-post semantic matching - Author reputation analysis - User engagement metrics - Content freshness evaluation >> Performance Improvements The system has achieved remarkable results: - 10%+ improvement in both on-topic rate and long-dwell metrics - Enhanced ability to handle complex natural language queries - Significant boost in sitewide engagement This advancement enables LinkedIn to better serve complex queries like "how to ask for a raise?" while maintaining high performance at scale. The system intelligently balances between exact keyword matching and semantic understanding, ensuring optimal results for both navigational and conceptual searches. What impresses me most is how the team solved the scale challenge - processing billions of posts efficiently using pre-computed embeddings and approximate nearest neighbor search. This is enterprise-scale AI at its finest.
posted an update 5 days ago
Excited to share a groundbreaking development in recommendation systems - Legommenders, a comprehensive content-based recommendation library that revolutionizes how we approach personalized content delivery. >> Key Innovations End-to-End Training The library enables joint training of content encoders alongside behavior and interaction modules, making it the first of its kind to offer truly integrated content understanding in recommendation pipelines. Massive Scale - Supports creation and analysis of over 1,000 distinct models - Compatible with 15 diverse datasets - Features 15 content operators, 8 behavior operators, and 9 click predictors Advanced LLM Integration Legommenders pioneers LLM integration in two crucial ways: - As feature encoders for enhanced content understanding - As data generators for high-quality training data augmentation Superior Architecture The system comprises four core components: - Dataset processor for unified data handling - Content operator for embedding generation - Behavior operator for user sequence fusion - Click predictor for probability calculations Performance Optimization The library introduces an innovative caching pipeline that achieves up to 50x speedup in evaluation compared to traditional approaches. Developed by researchers from The Hong Kong Polytechnic University, this open-source project represents a significant leap forward in recommendation system technology. For those interested in content-based recommendation systems, this is a must-explore tool. The library is available on GitHub for implementation and experimentation.
View all activity

Organizations

MLX Community's profile picture Social Post Explorers's profile picture C4AI Community's profile picture

singhsidhukuldeep's activity

posted an update 2 days ago
view post
Post
971
Breaking News: LinkedIn's Content Search Engine Gets a Powerful Semantic Upgrade!

Excited to share insights about LinkedIn's innovative approach to content search, recently detailed in a groundbreaking paper by their Mountain View team. This advancement represents a significant shift from traditional keyword-based search to semantic understanding.

>> Technical Architecture

The new search engine employs a sophisticated two-layer architecture:

Retrieval Layer
- Token Based Retriever (TBR) for exact keyword matching
- Embedding Based Retriever (EBR) using a two-tower model with multilingual-e5 embeddings
- Pre-computed post embeddings stored in a dedicated embedding store for efficient retrieval

Multi-Stage Ranking
- L1 Stage: Initial filtering using a lightweight model
- L2 Stage: Advanced ranking with complex features including:
- Query-post semantic matching
- Author reputation analysis
- User engagement metrics
- Content freshness evaluation

>> Performance Improvements

The system has achieved remarkable results:
- 10%+ improvement in both on-topic rate and long-dwell metrics
- Enhanced ability to handle complex natural language queries
- Significant boost in sitewide engagement

This advancement enables LinkedIn to better serve complex queries like "how to ask for a raise?" while maintaining high performance at scale. The system intelligently balances between exact keyword matching and semantic understanding, ensuring optimal results for both navigational and conceptual searches.

What impresses me most is how the team solved the scale challenge - processing billions of posts efficiently using pre-computed embeddings and approximate nearest neighbor search. This is enterprise-scale AI at its finest.
posted an update 4 days ago
view post
Post
500
Just read a fascinating survey paper on Query Optimization in Large Language Models by researchers at Tencent's Machine Learning Platform Department.

The paper deep dives into how we can enhance LLMs' ability to understand and answer complex queries, particularly in Retrieval-Augmented Generation (RAG) systems. Here's what caught my attention:

>> Key Technical Innovations

Core Operations:
- Query Expansion: Both internal (using LLM's knowledge) and external (web/knowledge base) expansion
- Query Disambiguation: Handling ambiguous queries through intent clarification
- Query Decomposition: Breaking complex queries into manageable sub-queries
- Query Abstraction: Stepping back to understand high-level principles

Under the Hood:
The system employs sophisticated techniques like GENREAD for contextual document generation, Query2Doc for pseudo-document creation, and FLARE's iterative anticipation mechanism for enhanced retrieval.

>> Real-World Applications

The framework addresses critical challenges in:
- Domain-specific tasks
- Knowledge-intensive operations
- Multi-hop reasoning
- Complex information retrieval

What's particularly impressive is how this approach significantly reduces hallucinations in LLMs while maintaining cost-effectiveness. The researchers have meticulously categorized query difficulties into four types, ranging from single-piece explicit evidence to multiple-piece implicit evidence requirements
posted an update 5 days ago
view post
Post
607
Excited to share a groundbreaking development in recommendation systems - Legommenders, a comprehensive content-based recommendation library that revolutionizes how we approach personalized content delivery.

>> Key Innovations

End-to-End Training
The library enables joint training of content encoders alongside behavior and interaction modules, making it the first of its kind to offer truly integrated content understanding in recommendation pipelines.

Massive Scale
- Supports creation and analysis of over 1,000 distinct models
- Compatible with 15 diverse datasets
- Features 15 content operators, 8 behavior operators, and 9 click predictors

Advanced LLM Integration
Legommenders pioneers LLM integration in two crucial ways:
- As feature encoders for enhanced content understanding
- As data generators for high-quality training data augmentation

Superior Architecture
The system comprises four core components:
- Dataset processor for unified data handling
- Content operator for embedding generation
- Behavior operator for user sequence fusion
- Click predictor for probability calculations

Performance Optimization
The library introduces an innovative caching pipeline that achieves up to 50x speedup in evaluation compared to traditional approaches.

Developed by researchers from The Hong Kong Polytechnic University, this open-source project represents a significant leap forward in recommendation system technology.

For those interested in content-based recommendation systems, this is a must-explore tool. The library is available on GitHub for implementation and experimentation.
posted an update 7 days ago
view post
Post
1754
Groundbreaking Survey on Large Language Models in Recommendation Systems!

Just read a comprehensive survey that maps out how LLMs are revolutionizing recommender systems. The authors have meticulously categorized existing approaches into two major paradigms:

Discriminative LLMs for Recommendation:
- Leverages BERT-like models for understanding user-item interactions
- Uses fine-tuning and prompt tuning to adapt pre-trained models
- Excels at tasks like user representation learning and ranking

Generative LLMs for Recommendation:
- Employs GPT-style models to directly generate recommendations
- Implements innovative techniques like in-context learning and zero-shot recommendation
- Supports natural language interaction and explanation generation

Key Technical Insights:
- Novel taxonomy of modeling paradigms: LLM Embeddings + RS, LLM Tokens + RS, and LLM as RS
- Integration methods spanning from simple prompting to sophisticated instruction tuning
- Hybrid approaches combining collaborative filtering with LLM capabilities
- Advanced prompt engineering techniques for controlled recommendation generation

Critical Challenges Identified:
- Position and popularity bias in LLM recommendations
- Limited context length affecting user history processing
- Need for better evaluation metrics for generative recommendations
- Controlled output generation and personalization challenges

This work opens exciting possibilities for next-gen recommendation systems while highlighting crucial areas for future research.
  • 1 reply
Β·
posted an update 10 days ago
view post
Post
1426
Groundbreaking Research Alert: Correctness β‰  Faithfulness in RAG Systems

Fascinating new research from L3S Research Center, University of Amsterdam, and TU Delft reveals a critical insight into Retrieval Augmented Generation (RAG) systems. The study exposes that up to 57% of citations in RAG systems could be unfaithful, despite being technically correct.

>> Key Technical Insights:

Post-rationalization Problem
The researchers discovered that RAG systems often engage in "post-rationalization" - where models first generate answers from their parametric memory and then search for supporting evidence afterward. This means that while citations may be correct, they don't reflect the actual reasoning process.

Experimental Design
The team used Command-R+ (104B parameters) with 4-bit quantization on NVIDIA A100 GPU, testing on the NaturalQuestions dataset. They employed BM25 for initial retrieval and ColBERT v2 for reranking.

Attribution Framework
The research introduces a comprehensive framework for evaluating RAG systems across multiple dimensions:
- Citation Correctness: Whether cited documents support the claims
- Citation Faithfulness: Whether citations reflect actual model reasoning
- Citation Appropriateness: Relevance and meaningfulness of citations
- Citation Comprehensiveness: Coverage of key points

Under the Hood
The system processes involve:
1. Document relevance prediction
2. Citation prediction
3. Answer generation without citations
4. Answer generation with citations

This work fundamentally challenges our understanding of RAG systems and highlights the need for more robust evaluation metrics in AI systems that claim to provide verifiable information.
  • 2 replies
Β·
posted an update 13 days ago
view post
Post
3390
Exciting breakthrough in e-commerce recommendation systems!
Walmart Global Tech researchers have developed a novel Triple Modality Fusion (TMF) framework that revolutionizes how we make product recommendations.

>> Key Innovation
The framework ingeniously combines three distinct data types:
- Visual data to capture product aesthetics and context
- Textual information for detailed product features
- Graph data to understand complex user-item relationships

>> Technical Architecture
The system leverages a Large Language Model (Llama2-7B) as its backbone and introduces several sophisticated components:

Modality Fusion Module
- All-Modality Self-Attention (AMSA) for unified representation
- Cross-Modality Attention (CMA) mechanism for deep feature integration
- Custom FFN adapters to align different modality embeddings

Advanced Training Strategy
- Curriculum learning approach with three complexity levels
- Parameter-Efficient Fine-Tuning using LoRA
- Special token system for behavior and item representation

>> Real-World Impact
The results are remarkable:
- 38.25% improvement in Electronics recommendations
- 43.09% boost in Sports category accuracy
- Significantly higher human evaluation scores compared to traditional methods

Currently deployed in Walmart's production environment, this research demonstrates how combining multiple data modalities with advanced LLM architectures can dramatically improve recommendation accuracy and user satisfaction.
  • 2 replies
Β·
posted an update 14 days ago
view post
Post
3096
Groundbreaking Research Alert: Rethinking RAG with Cache-Augmented Generation (CAG)

Researchers from National Chengchi University and Academia Sinica have introduced a paradigm-shifting approach that challenges the conventional wisdom of Retrieval-Augmented Generation (RAG).

Instead of the traditional retrieve-then-generate pipeline, their innovative Cache-Augmented Generation (CAG) framework preloads documents and precomputes key-value caches, eliminating the need for real-time retrieval during inference.

Technical Deep Dive:
- CAG preloads external knowledge and precomputes KV caches, storing them for future use
- The system processes documents only once, regardless of subsequent query volume
- During inference, it loads the precomputed cache alongside user queries, enabling rapid response generation
- The cache reset mechanism allows efficient handling of multiple inference sessions through strategic token truncation

Performance Highlights:
- Achieved superior BERTScore metrics compared to both sparse and dense retrieval RAG systems
- Demonstrated up to 40x faster generation times compared to traditional approaches
- Particularly effective with both SQuAD and HotPotQA datasets, showing robust performance across different knowledge tasks

Why This Matters:
The approach significantly reduces system complexity, eliminates retrieval latency, and mitigates common RAG pipeline errors. As LLMs continue evolving with expanded context windows, this methodology becomes increasingly relevant for knowledge-intensive applications.
posted an update 18 days ago
view post
Post
1617
Excited to share insights from Walmart's groundbreaking semantic search system that revolutionizes e-commerce product discovery!

The team at Walmart Global Technology(the team that I am a part of 😬) has developed a hybrid retrieval system that combines traditional inverted index search with neural embedding-based search to tackle the challenging problem of tail queries in e-commerce.

Key Technical Highlights:

β€’ The system uses a two-tower BERT architecture where one tower processes queries and another processes product information, generating dense vector representations for semantic matching.

β€’ Product information is enriched by combining titles with key attributes like category, brand, color, and gender using special prefix tokens to help the model distinguish different attribute types.

β€’ The neural model leverages DistilBERT with 6 layers and projects the 768-dimensional embeddings down to 256 dimensions using a linear layer, achieving optimal performance while reducing storage and computation costs.

β€’ To improve model training, they implemented innovative negative sampling techniques combining product category matching and token overlap filtering to identify challenging negative examples.

Production Implementation Details:

β€’ The system uses a managed ANN (Approximate Nearest Neighbor) service to enable fast retrieval, achieving 99% recall@20 with just 13ms latency.

β€’ Query embeddings are cached with preset TTL (Time-To-Live) to reduce latency and costs in production.

β€’ The model is exported to ONNX format and served in Java, with custom optimizations like fixed input shapes and GPU acceleration using NVIDIA T4 processors.

Results:
The system showed significant improvements in both offline metrics and live experiments, with:
- +2.84% improvement in NDCG@10 for human evaluation
- +0.54% lift in Add-to-Cart rates in live A/B testing

This is a fantastic example of how modern NLP techniques can be successfully deployed at scale to solve real-world e-
  • 1 reply
Β·
posted an update 20 days ago
view post
Post
2103
Groundbreaking Research Alert: Revolutionizing Document Ranking with Long-Context LLMs

Researchers from Renmin University of China and Baidu Inc . have introduced a novel approach to document ranking that challenges conventional sliding window methods. Their work demonstrates how long-context Large Language Models can process up to 100 documents simultaneously, achieving superior performance while reducing API costs by 50%.

Key Technical Innovations:
- Full ranking strategy enables processing all passages in a single inference
- Multi-pass sliding window approach for comprehensive listwise label construction
- Importance-aware learning objective that prioritizes top-ranked passage IDs
- Support for context lengths up to 128k tokens using models like LLaMA 3.1-8B-Instruct

Performance Highlights:
- 2.2 point improvement in NDCG@10 metrics
- 29.3% reduction in latency compared to traditional methods
- Significant API cost savings through elimination of redundant passage processing

Under the hood, the system leverages advanced long-context LLMs to perform global interactions among passages, enabling more nuanced relevance assessment. The architecture incorporates a novel importance-aware loss function that assigns differential weights based on passage ranking positions.

The research team's implementation demonstrated remarkable versatility across multiple datasets, including TREC DL and BEIR benchmarks. Their fine-tuned model, RankMistral, showcases the practical viability of full ranking approaches in production environments.

This advancement marks a significant step forward in information retrieval systems, offering both improved accuracy and computational efficiency. The implications for search engines and content recommendation systems are substantial.
posted an update 25 days ago
view post
Post
2188
Exciting News in AI: JinaAI Releases JINA-CLIP-v2!

The team at Jina AI has just released a groundbreaking multilingual multimodal embedding model that's pushing the boundaries of text-image understanding. Here's why this is a big deal:

πŸš€ Technical Highlights:
- Dual encoder architecture combining a 561M parameter Jina XLM-RoBERTa text encoder and a 304M parameter EVA02-L14 vision encoder
- Supports 89 languages with 8,192 token context length
- Processes images up to 512Γ—512 pixels with 14Γ—14 patch size
- Implements FlashAttention2 for text and xFormers for vision processing
- Uses Matryoshka Representation Learning for efficient vector storage

⚑️ Under The Hood:
- Multi-stage training process with progressive resolution scaling (224β†’384β†’512)
- Contrastive learning using InfoNCE loss in both directions
- Trained on massive multilingual dataset including 400M English and 400M multilingual image-caption pairs
- Incorporates specialized datasets for document understanding, scientific graphs, and infographics
- Uses hard negative mining with 7 negatives per positive sample

πŸ“Š Performance:
- Outperforms previous models on visual document retrieval (52.65% nDCG@5)
- Achieves 89.73% image-to-text and 79.09% text-to-image retrieval on CLIP benchmark
- Strong multilingual performance across 30 languages
- Maintains performance even with 75% dimension reduction (256D vs 1024D)

🎯 Key Innovation:
The model solves the long-standing challenge of unifying text-only and multi-modal retrieval systems while adding robust multilingual support. Perfect for building cross-lingual visual search systems!

Kudos to the research team at Jina AI for this impressive advancement in multimodal AI!
posted an update 26 days ago
view post
Post
1279
Fascinating insights from @Pinterest 's latest research on improving feature interactions in recommendation systems!

Pinterest's engineering team has tackled a critical challenge in their Homefeed ranking system that serves 500M+ monthly active users. Here's what makes their approach remarkable:

>> Technical Deep Dive

Architecture Overview
β€’ The ranking model combines dense features, sparse features, and embedding features to represent users, Pins, and context
β€’ Sparse features are processed using learnable embeddings with size based on feature cardinality
β€’ User sequence embeddings are generated using a transformer architecture processing past engagements

Feature Processing Pipeline
β€’ Dense features undergo normalization for numerical stability
β€’ Sparse and embedding features receive L2 normalization
β€’ All features are concatenated into a single feature embedding

Key Innovations
β€’ Implemented parallel MaskNet layers with 3 blocks
β€’ Used projection ratio of 2.0 and output dimension of 512
β€’ Stacked 4 DCNv2 layers on top for higher-order interactions

Performance Improvements
β€’ Achieved +1.42% increase in Homefeed Save Volume
β€’ Boosted Overall Time Spent by +0.39%
β€’ Maintained memory consumption increase to just 5%

>> Industry Constraints Addressed

Memory Management
β€’ Optimized for 60% GPU memory utilization
β€’ Prevented OOM errors while maintaining batch size efficiency

Latency Optimization
β€’ Removed input-output concatenation before MLP
β€’ Reduced hidden layer sizes in MLP
β€’ Achieved zero latency increase while improving performance

System Stability
β€’ Ensured reproducible results across retraining
β€’ Maintained model stability across different data distributions
β€’ Successfully deployed in production environment

This work brilliantly demonstrates how to balance academic innovations with real-world industrial constraints. Kudos to the Pinterest team!
posted an update 28 days ago
view post
Post
3643
Exciting breakthrough in AI: @Meta 's new Byte Latent Transformer (BLT) revolutionizes language models by eliminating tokenization!

The BLT architecture introduces a groundbreaking approach that processes raw bytes instead of tokens, achieving state-of-the-art performance while being more efficient and robust. Here's what makes it special:

>> Key Innovations
Dynamic Patching: BLT groups bytes into variable-sized patches based on entropy, allocating more compute power where the data is more complex. This results in up to 50% fewer FLOPs during inference compared to traditional token-based models.

Three-Component Architecture:
β€’ Lightweight Local Encoder that converts bytes to patch representations
β€’ Powerful Global Latent Transformer that processes patches
β€’ Local Decoder that converts patches back to bytes

>> Technical Advantages
β€’ Matches performance of Llama 3 at 8B parameters while being more efficient
β€’ Superior handling of non-English languages and rare character sequences
β€’ Remarkable 99.9% accuracy on spelling tasks
β€’ Better scaling properties than token-based models

>> Under the Hood
The system uses an entropy model to determine patch boundaries, cross-attention mechanisms for information flow, and hash n-gram embeddings for improved representation. The architecture allows simultaneous scaling of both patch and model size while maintaining fixed inference costs.

This is a game-changer for multilingual AI and could reshape how we build future language models. Excited to see how this technology evolves!
Β·
posted an update about 1 month ago
view post
Post
1246
Groundbreaking Research Alert: The 'H' in HNSW Stands for "Hubs", Not "Hierarchy"!

Fascinating new research reveals that the hierarchical structure in the popular HNSW (Hierarchical Navigable Small World) algorithm - widely used for vector similarity search - may be unnecessary for high-dimensional data.

πŸ”¬ Key Technical Findings:

β€’ The hierarchical layers in HNSW can be completely removed for vectors with dimensionality > 32, with no performance loss

β€’ Memory savings of up to 38% achieved by removing the hierarchy

β€’ Performance remains identical in both median and tail latency cases across 13 benchmark datasets

πŸ› οΈ Under The Hood:
The researchers discovered that "hub highways" naturally form in high-dimensional spaces. These hubs are well-connected nodes that are frequently traversed during searches, effectively replacing the need for explicit hierarchical layers.

The hub structure works because:
β€’ A small subset of nodes appear disproportionately in nearest neighbor lists
β€’ These hub nodes form highly connected subgraphs
β€’ Queries naturally traverse through these hubs early in the search process
β€’ The hubs efficiently connect distant regions of the graph

πŸ’‘ Industry Impact:
This finding has major implications for vector databases and similarity search systems. Companies can significantly reduce memory usage while maintaining performance by implementing flat navigable small world graphs instead of hierarchical ones.

πŸš€ What's Next:
The researchers have released FlatNav, an open-source implementation of their flat navigable small world approach, enabling immediate practical applications of these findings.
posted an update about 1 month ago
view post
Post
469
Fascinating new research alert! Just read a groundbreaking paper on understanding Retrieval-Augmented Generation (RAG) systems and their performance factors.

Key insights from this comprehensive study:

>> Architecture Deep Dive
The researchers analyzed RAG systems across 6 datasets (3 code-related, 3 QA-focused) using multiple LLMs. Their investigation revealed critical insights into four key design factors:

Document Types Impact:
β€’ Oracle documents (ground truth) aren't always optimal
β€’ Distracting documents significantly degrade performance
β€’ Surprisingly, irrelevant documents boost code generation by up to 15.6%

Retrieval Precision:
β€’ Performance varies dramatically by task
β€’ QA tasks need 20-100% retrieval recall
β€’ Perfect retrieval still fails up to 12% of the time on previously correct instances

Document Selection:
β€’ More documents β‰  better results
β€’ Adding documents can cause errors on previously correct samples
β€’ Performance degradation increases ~1% per 5 additional documents in code tasks

Prompt Engineering:
β€’ Most advanced prompting techniques underperform simple zero-shot prompts
β€’ Technique effectiveness varies significantly across models and tasks
β€’ Complex prompts excel at difficult problems but struggle with simple ones

>> Technical Implementation
The study utilized:
β€’ Multiple retrievers including BM25, dense retrievers, and specialized models
β€’ Comprehensive corpus of 70,956 unique API documents
β€’ Over 200,000 API calls and 1,000+ GPU hours of computation
β€’ Sophisticated evaluation metrics tracking both correctness and system confidence

πŸ’‘ Key takeaway: RAG system optimization requires careful balancing of multiple factors - there's no one-size-fits-all solution.
  • 1 reply
Β·
posted an update about 1 month ago
view post
Post
1816
Exciting new research alert! πŸš€ A groundbreaking paper titled "Understanding LLM Embeddings for Regression" has just been released, and it's a game-changer for anyone working with large language models (LLMs) and regression tasks.

Key findings:

1. LLM embeddings outperform traditional feature engineering in high-dimensional regression tasks.

2. LLM embeddings preserve Lipschitz continuity over feature space, enabling better regression performance.

3. Surprisingly, factors like model size and language understanding don't always improve regression outcomes.

Technical details:

The researchers used both T5 and Gemini model families to benchmark embedding-based regression. They employed a key-value JSON format for string representations and used average-pooling to aggregate Transformer outputs.

The study introduced a novel metric called Normalized Lipschitz Factor Distribution (NLFD) to analyze embedding continuity. This metric showed a high inverse relationship between the skewedness of the NLFD and regression performance.

Interestingly, the paper reveals that applying forward passes of pre-trained models doesn't always significantly improve regression performance for certain tasks. In some cases, using only vocabulary embeddings without a forward pass yielded comparable results.

The research also demonstrated that LLM embeddings are dimensionally robust, maintaining strong performance even with high-dimensional data where traditional representations falter.

This work opens up exciting possibilities for using LLM embeddings in various regression tasks, particularly those with high degrees of freedom. It's a must-read for anyone working on machine learning, natural language processing, or data science!
posted an update about 1 month ago
view post
Post
2091
Exciting breakthrough in E-commerce Recommendation Systems!

Just read a fascinating paper from @eBay 's research team on "LLM-PKG" - a novel approach that combines Large Language Models with Product Knowledge Graphs for explainable recommendations.

Here's what makes it groundbreaking:

>> Technical Architecture
- The system uses a two-module approach: offline construction and online serving
- LLM generates initial product relationships and rationales, which are transformed into RDF triplets (Subject, Predicate, Object) to build the knowledge graph
- The system employs rigorous validation using LLM-based scoring (1-10 scale) to evaluate recommendation quality and prune low-quality nodes (score < 6)

>> Under the Hood
- Product mapping uses BERT embeddings and KNN indexing for semantic matching between LLM recommendations and actual inventory
- The system caches graph triplets in key-value databases for lightning-fast retrieval during online serving
- Supports both item-centric and user-centric recommendation scenarios

>> Real-World Impact
The A/B testing results are impressive:
- 5.19% increase in clicks
- 7.59% boost in transactions
- 8.56% growth in Gross Merchandise Bought
- 10.84% increase in ad revenue

This is a game-changer for e-commerce platforms looking to provide transparent, explainable recommendations while maintaining high performance at scale.
posted an update about 1 month ago
view post
Post
1263
Exciting breakthrough in AI Recommendation Systems! Just read a fascinating paper from Meta AI and UW-Madison researchers on unifying generative and dense retrieval methods for recommendations.

The team introduced LIGER (LeveragIng dense retrieval for GEnerative Retrieval), a novel hybrid approach that combines the best of both worlds:

Key Technical Innovations:
- Integrates semantic ID-based generative retrieval with dense embedding methods
- Uses a T5 encoder-decoder architecture with 6 layers, 6 attention heads, and 128-dim embeddings
- Processes item attributes through sentence-T5-XXL for text representations
- Employs a dual-objective training approach combining cosine similarity and next-token prediction
- Implements beam search with size K for candidate generation
- Features an RQ-VAE with 3-layer MLP for semantic ID generation

Performance Highlights:
- Significantly outperforms traditional methods on cold-start recommendations
- Achieves state-of-the-art results on major benchmark datasets (Amazon Beauty, Sports, Toys, Steam)
- Reduces computational complexity from O(N) to O(tK) where t is semantic ID count
- Maintains minimal storage requirements while improving recommendation quality

The most impressive part? LIGER effectively solves the cold-start problem that has long plagued recommendation systems while maintaining computational efficiency.

This could be a game-changer for e-commerce platforms and content recommendation systems!

What are your thoughts on hybrid recommendation approaches?
posted an update about 1 month ago
view post
Post
290
Exciting breakthrough in Search Engine Technology! Just read a fascinating paper on "Best Practices for Distilling Large Language Models into BERT for Web Search Ranking" from @TencentGlobal

Game-Changing Innovation: DisRanker
A novel distillation pipeline that combines the power of Large Language Models with BERT's efficiency for web search ranking - now deployed in commercial search engines!

Key Technical Highlights:
β€’ Implements domain-specific Continued Pre-Training using clickstream data, treating queries as inputs to generate clicked titles and summaries
β€’ Uses an end-of-sequence token to represent query-document pairs during supervised fine-tuning
β€’ Employs hybrid Point-MSE and Margin-MSE loss for knowledge distillation, optimizing both absolute scores and relative rankings

Under the Hood:
- The system first pre-trains on massive clickstream data (59M+ query-document pairs)
- Transfers ranking expertise from a 7B parameter LLM to a compact BERT model
- Reduces inference latency from ~100ms to just 10ms while maintaining performance
- Achieves significant improvements:
β€’ +0.47% PageCTR
β€’ +0.58% UserCTR
β€’ +1.2% Dwell Time

Real-World Impact:
Successfully integrated into production search systems as of February 2024, demonstrating that academic research can translate into practical industry solutions

What are your thoughts on this breakthrough?
posted an update about 1 month ago
view post
Post
387
Exciting Research Alert: Revolutionizing Recommendation Systems with PSL (Pairwise Softmax Loss)!

I just read a fascinating paper that introduces PSL - a groundbreaking approach to improve recommendation systems. Here's why this matters:

>> Key Innovations

Core Concept: PSL reimagines the traditional Softmax Loss by viewing it through a pairwise perspective, addressing two critical limitations of current systems:
- The loose connection between Softmax Loss and ranking metrics like DCG
- High sensitivity to false negative instances

Technical Implementation:
- Replaces exponential functions with alternative activation functions (Tanh, Atan, ReLU)
- Reformulates loss calculation from a pairwise perspective
- Integrates Distributionally Robust Optimization (DRO) principles

>> Real-World Impact

Enhanced Performance:
- Tighter surrogate for ranking metrics
- Better balance in data contribution weights
- Improved robustness against false negatives
- Superior handling of out-of-distribution scenarios

Practical Applications:
- E-commerce recommendations
- Content discovery systems
- Personalized service platforms

>> Implementation Benefits

The beauty of PSL lies in its simplicity - it requires minimal code modifications while delivering significant improvements in:
- Recommendation accuracy
- System robustness
- Training stability
- Distribution shift handling

This research opens new possibilities for building more reliable and accurate recommendation systems. The code is available on GitHub for those interested in implementation.

What are your thoughts on this approach? Have you encountered similar challenges in recommendation systems?
posted an update about 2 months ago
view post
Post
1315
Exciting breakthrough in Document AI! Researchers from UNC Chapel Hill and Bloomberg have developed M3DocRAG, a revolutionary framework for multi-modal document understanding.

The innovation lies in its ability to handle complex document scenarios that traditional systems struggle with:
- Process 40,000+ pages across 3,000+ documents
- Answer questions requiring information from multiple pages
- Understand visual elements like charts, tables, and figures
- Support both closed-domain (single document) and open-domain (multiple documents) queries

Under the hood, M3DocRAG operates through three sophisticated stages:

>> Document Embedding:
- Converts PDF pages to RGB images
- Uses ColPali to project both text queries and page images into a shared embedding space
- Creates dense visual embeddings for each page while maintaining visual information integrity

>> Page Retrieval:
- Employs MaxSim scoring to compute relevance between queries and pages
- Implements inverted file indexing (IVFFlat) for efficient search
- Reduces retrieval latency from 20s to under 2s when searching 40K+ pages
- Supports approximate nearest neighbor search via Faiss

>> Question Answering:
- Leverages Qwen2-VL 7B as the multi-modal language model
- Processes retrieved pages through a visual encoder
- Generates answers considering both textual and visual context

The results are impressive:
- State-of-the-art performance on MP-DocVQA benchmark
- Superior handling of non-text evidence compared to text-only systems
- Significantly better performance on multi-hop reasoning tasks

This is a game-changer for industries dealing with large document volumesβ€”finance, healthcare, and legal sectors can now process documents more efficiently while preserving crucial visual context.
Β·