Connect with us

Hi, what are you looking for?

AI Generative

LinkedIn Reveals LLM-Based Feed Overhaul, Boosts Content Relevance by 30x with GPUs

LinkedIn overhauls its Feed with LLMs and GPUs, boosting content relevance by 30x and driving a 121% return on ad spend for marketers.

LinkedIn announced this week a significant overhaul of its Feed recommendation system, utilizing large language models (LLMs) and cutting-edge GPU technology to enhance content ranking for its global user base of over 1.3 billion professionals. The detailed engineering blog, authored by Hristo Danchev and published on March 12, 2026, provides the first granular insights into how the platform curates content, a critical factor as LinkedIn continues to dominate B2B paid media, accounting for 41% of total budgets in 2025, according to Dreamdata.

The previous architecture of LinkedIn’s Feed was described as “heterogeneous,” relying on multiple separate systems that operated independently. This fragmented approach produced a variety of content but involved substantial maintenance costs. Each system maintained its own infrastructure and optimization logic, making holistic tuning complex and inefficient. The new unified retrieval pipeline aims to streamline this process by generating embeddings through LLMs, which better capture the semantic proximity of posts and member profiles, enhancing relevance in content presentation.

One key improvement highlighted in the announcement is the system’s ability to handle “cold-start” scenarios, wherein new members join the platform with minimal data. Traditional methods could only identify basic correlations based on profile information. In contrast, the newly deployed model, trained on extensive pre-existing data, can infer deeper interests relevant to users, suggesting that an electrical engineer may have latent interests in renewable energy, even without past engagement.

LinkedIn’s engineers faced challenges in processing raw engagement metrics, as numerical features were treated as arbitrary text tokens, leading to a lack of correlation between item popularity and relevance. The solution involved percentile bucketing, which transformed raw counts into percentile ranks that provided a clearer signal of a post’s engagement level. This adjustment resulted in a 30-fold improvement in correlation between popularity and embedding similarity, significantly enhancing relevance in content retrieval.

The system employs a dual encoder architecture where a shared LLM processes both member and item prompts, generating embeddings that are evaluated through cosine similarity. Training methodologies included using both easy and hard negative examples to refine the model’s ability to distinguish between nearly relevant and genuinely valuable content. This nuanced approach improved the system’s recall rate, underlining the importance of incorporating a member’s interaction history to enhance performance.

At the ranking stage, LinkedIn’s Generative Recommender (GR) model processes over 1,000 historical interactions, utilizing a transformer architecture with causal attention. This design respects the temporal flow of user engagement, allowing the model to recognize and adapt to shifts in a member’s interests over time. By employing a technique called late fusion, the model integrates static context features without inflating computational costs, ensuring efficient processing.

Serving this sophisticated model at scale poses distinct engineering challenges. Traditional ranking models relied on CPUs, but the LLM architecture necessitates high-bandwidth memory available only on GPUs. LinkedIn’s innovative solutions include a disaggregated architecture that separately handles feature processing and model inference, along with a custom Flash Attention variant that enhances performance even further, achieving sub-50ms retrieval latency across millions of indexed posts.

The implications of these advancements extend beyond improved organic reach. The same ranking logic that governs organic content also influences LinkedIn’s sponsored placements, helping to drive a substantial return on ad spend of 121% in 2025. Marketers may find the new system increasingly receptive to content targeting professionals in adjacent or emerging fields, as the model’s understanding of latent interests could yield new opportunities for engagement.

As LinkedIn’s role in B2B media continues to grow—now accounting for the largest share of budget allocations—it becomes crucial for marketers to grasp how the Feed prioritizes content. The shift towards LLM-based reasoning signifies a departure from traditional keyword competition, allowing posts about “data security” to connect with broader themes such as regulatory compliance and operational risk. This expanded competitive landscape requires content strategists to adapt their approaches significantly.

LinkedIn’s ongoing commitment to responsible AI practices is also notable. The platform emphasizes regular audits of its models to ensure fair competition among creators, while the system’s design deliberately excludes demographic attributes, focusing instead on professional signals and engagement patterns. This transparency is crucial as the platform continues to evolve and influence B2B marketing strategies.

Staff
Written By

The AiPressa Staff team brings you comprehensive coverage of the artificial intelligence industry, including breaking news, research developments, business trends, and policy updates. Our mission is to keep you informed about the rapidly evolving world of AI technology.

You May Also Like

Top Stories

Pentagon halts Anthropic's AI contracts over surveillance and lethal weapons concerns, igniting a legal battle that could redefine military tech governance.

AI Marketing

Webflow acquires Vidoso, a marketing content startup, to enhance platform integration, signaling a shift towards a comprehensive AI-driven marketing solution.

AI Technology

NVIDIA's stock at $180.25 is overvalued by 35.5% compared to its intrinsic value of $133.06, amid surging AI chip demand and regulatory challenges.

AI Generative

Google researchers enhance large language models' accuracy to 81% using a novel Bayesian teaching method for improved probabilistic reasoning in user interactions

Top Stories

Amazon secures a federal court ruling blocking Perplexity AI's Comet shopping assistant from accessing its site, raising critical data security concerns in AI commerce.

AI Cybersecurity

Cloudflare warns that by 2026, AI will enable cybercriminals to automate attacks at scale, threatening corporate security with sophisticated tactics like deepfakes and SaaS...

AI Generative

Researchers unveil P-EAGLE, boosting LLM inference speeds by up to 1.69x on NVIDIA B200, revolutionizing token generation efficiency.

Top Stories

Amazon enforces new oversight rules for AI coding tools after outages led to 6.3 million lost orders, signaling a major shift in tech risk...

© 2025 AIPressa · Part of Buzzora Media · All rights reserved. This website provides general news and educational content for informational purposes only. While we strive for accuracy, we do not guarantee the completeness or reliability of the information presented. The content should not be considered professional advice of any kind. Readers are encouraged to verify facts and consult appropriate experts when needed. We are not responsible for any loss or inconvenience resulting from the use of information on this site. Some images used on this website are generated with artificial intelligence and are illustrative in nature. They may not accurately represent the products, people, or events described in the articles.