Activity
Mon
Wed
Fri
Sun
Oct
Nov
Dec
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
What is this?
Less
More

Owned by Guerin

Burstiness and Perplexity

234 members • Free

Master AI use cases from legal & the supply chain to digital marketing & SEO. Agents, analysis, content creation--Burstiness & Perplexity from NovCog

Memberships

AI Money Lab

27.7k members • Free

Turboware - Skunk.Tech

24 members • Free

Ai Automation Vault

12.3k members • Free

AI Automation Society

143.8k members • Free

CribOps

54 members • $39/m

AI Marketeers

245 members • $40/m

Skoolers

174.8k members • Free

AI Automation Agency Ninjas

17.4k members • Free

45 contributions to Burstiness and Perplexity
DeepSeek-R1 and the Emergence of Reasoning via Reinforcement Learning
This document synthesizes findings on DeepSeek-R1, a Large Language Model (LLM) whose reasoning abilities have been significantly enhanced through a novel application of pure Reinforcement Learning (RL). The core thesis is that LLMs possess substantial latent reasoning potential that can be unlocked without extensive human-annotated reasoning trajectories. By providing hard reasoning questions, a reliable verifier (reward signal), and sufficient computational resources, the model can self-evolve sophisticated problem-solving strategies. The initial model, DeepSeek-R1-Zero, was trained using RL on the DeepSeek-V3 Base model, bypassing conventional supervised fine-tuning. It achieved superior performance on verifiable tasks in mathematics, coding, and STEM fields, notably improving its score on the AIME 2024 benchmark from 15.6% to 77.9%. This process led to the emergence of advanced reasoning patterns such as self-reflection, verification, and dynamic strategy adaptation. The final model, DeepSeek-R1, builds upon this foundation through a multi-stage pipeline that integrates RL with supervised fine-tuning and rejection sampling. This approach preserves the advanced reasoning of its predecessor while aligning the model with human preferences, improving instruction-following, readability, and general capabilities. The project highlights significant limitations, including challenges in structured output, token efficiency, and the risk of "reward hacking" in domains without rule-based verifiers. The models, data samples, and distilled smaller versions have been made publicly available to advance research in AI reasoning. Core Thesis: Incentivizing Reasoning with Pure Reinforcement Learning The central argument is that the reasoning capabilities of LLMs can be substantially incentivized through a pure Reinforcement Learning framework, obviating the need for human-labelled reasoning paths. Traditional methods, such as Chain-of-Thought (CoT) prompting or supervised learning on human demonstrations, are effective but have key limitations:
3
0
The "Traffic Apocalypse" for News Websites Driven by AI
Subject: Analysis of the significant decline in web traffic to major news websites in July 2025, primarily attributed to Google's AI Overviews feature and its broader implications for the media industry. Sources: Excerpts from "News Websites' Traffic Decline: The AI Impact" (Novel Cognition and Aibrandintelligence.com research, compiling data from Press Gazette, SimilarWeb, Columbia Journalism Review, Pew Research Center, and others). Executive Summary The digital news industry is experiencing a profound crisis, dubbed a "traffic apocalypse," with a widespread and significant decline in web traffic across major U.S. news websites. In July 2025, 46 of the top 50 U.S. news sites saw year-over-year traffic declines, with some publishers losing up to 50% of their audience compared to July 2024. The primary driver of this decline is Google's AI Overviews feature, launched in May 2024, which has drastically increased "zero-click" searches and reduced organic traffic to news sites. This shift is disrupting traditional publishing business models, leading to legal challenges and a pivot towards reader revenue, while raising concerns about the future of quality journalism. Key Themes and Most Important Ideas/Facts 1. Widespread and Severe Traffic Declines Across Major News Sites - Prevalence: "46 of the top 50 news sites experienced year-over-year traffic declines" in July 2025 compared to July 2024. - Magnitude: Some publishers "losing as much as 50% of their audience." - Major Publishers Hit Hard:Forbes: -50% (63 million visits) - Daily Mail: -44% (76.8 million visits) - NBC News: -42% (74.4 million visits) - HuffPost: -42% (38.5 million visits) - Washington Post: -40% (69.4 million visits) - CNN: -34% (323 million visits) - Fox News: -25% (249 million visits) - Limited Exceptions: Only three websites saw double-digit year-over-year growth: Substack (+47%), India Times (+46%), and Newsbreak (+24%). 2. Google AI Overviews as the Primary Driver of Decline
0
0
Understanding and Mitigating AI Hallucinations
This briefing document summarizes the core insights from the provided sources regarding the phenomenon of AI hallucinations, their underlying causes, and proposed solutions. 1. The Nature of AI Hallucinations AI hallucinations are defined as instances where large language models (LLMs) "confidently make things up," producing "plausible yet incorrect statements instead of admitting uncertainty." This differs fundamentally from human perceptual hallucinations. The problem is not necessarily about making models smarter or training them on more data; rather, it stems from the way AI models are currently trained and evaluated. Key Facts: - LLMs often provide "overconfident, plausible falsehoods," which "diminish their utility." - Examples include generating incorrect birthdates or dissertation titles for known individuals, even when explicitly asked to respond "only if known." - Hallucinations can be "intrinsic" (contradicting the user's prompt, e.g., miscounting letters in a word) or "extrinsic" (contradicting training data or external reality). Quote: "Language models are known to produce overconfident, plausible falsehoods, which diminish their utility. This error mode is known as 'hallucination,' though it differs fundamentally from the human perceptual experience." – why-language-models-hallucinate.pdf 2. Root Causes: Training and Evaluation Incentives The core argument across both sources is that AI models hallucinate because the current training and evaluation paradigms inadvertently reward guessing over honesty. Main Themes: - "Terrible Test-Takers": LLMs are "essentially training AI to be terrible test-takers who guess instead of admitting uncertainty." - Binary Scoring: Most benchmarks operate like "multiple-choice exams" with "binary 0-1 scheme[s]" where "1 point for a correct answer and none for blanks or IDKs." This incentivizes guessing, as "leaving an answer blank guarantees failure but guessing gives you a 1-in-365 chance of nailing someone's birthday." - Vicious Cycle: This leads to models learning to "bluff," generating "confident-sounding nonsense rather than admit uncertainty." As models become more capable, they continue to hallucinate because "that's what scores best on tests." - Statistical Origins (Pretraining): Hallucinations "originate simply as errors in binary classification." Even with error-free training data, the statistical objectives minimized during pretraining can lead to errors. This is due to factors like: - Arbitrary Facts: When there's no learnable pattern in data (e.g., specific birthdays), models are likely to hallucinate, with the hallucination rate being at least the "fraction of training facts that appear once." - Poor Models: The model architecture itself may be insufficient to represent the concept well (e.g., trigram models struggling with longer dependencies) or may not be a good fit even if expressive enough. - Computational Hardness: Problems that are computationally intractable for even superhuman AI will lead to errors if the model attempts to solve them rather than defer. - Distribution Shift (OOD Prompts): Prompts that differ significantly from training data can induce errors. - GIGO (Garbage In, Garbage Out): Training corpora often contain factual errors, which base models can replicate. - Persistence (Post-Training): Despite efforts to reduce hallucinations during post-training (e.g., RLHF), they persist because "guessing when unsure maximizes expected score under a binary 0-1 scheme." Existing primary evaluations "overwhelmingly penalize uncertainty."
2
0
Say What?? AI Inference
AI model inference is the process where a trained artificial intelligence model uses its knowledge to make predictions or draw conclusions from new, previously unseen data. It is the operational phase of AI, where the model applies what it learned during training to real-world situations.[1][4][5][7] The AI Lifecycle: Training vs. Inference The lifecycle of a machine learning model consists of two main phases: training and inference.[4][8] - Training This is the learning phase. An AI model is "trained" by processing vast amounts of labeled data to learn to recognize patterns, relationships, and features within that data. For example, a model designed to identify spam emails is fed millions of emails that are already labeled as "spam" or "not spam". This process builds the model's knowledge base, which is stored as parameters or "weights".[3][6][9][4] - Inference This is the application phase. Once trained, the model is deployed to perform its designated task on live, real-world data it has never encountered before. It uses its stored knowledge to "infer" or deduce an outcome. For example, when a new email arrives, the trained spam model analyzes its content and characteristics to predict whether it is spam. This is considered the "moment of truth" for an AI model.[6][8][9][3] Think of training as a student studying for an exam by reviewing course materials, while inference is the student taking the exam and applying that knowledge to answer new questions.[3] Examples of AI Inference AI inference is at the core of many modern technologies: - Generative AI Large language models (LLMs) like ChatGPT use inference to predict the next most likely word in a sequence, allowing them to generate coherent sentences and paragraphs.[3] - Autonomous Vehicles A self-driving car uses inference to identify a stop sign on a road it has never driven on by recognizing patterns learned from millions of images of stop signs during training.[1] - Facial Recognition A model trained on millions of facial images can infer an individual's identity in a new photo by identifying features like eye color and nose shape.[2] - Fraud Detection Banks use AI to analyze credit card transactions in real-time and infer whether a transaction is likely fraudulent based on learned patterns.[3]
0
0
Feeling out of depth?
Try this explainer prompt: {What the fuck does this mean? Explain it in non-jargon, while defining terms. You can do this parenthetically— placing the term definition in ( ) while walking thru all the details.}
2
0
1-10 of 45
Guerin Green
4
63points to level up
@guerin-green-9848
Novel Cognition, Burstiness and Perplexity. Former print newspaperman, public opinion & market research and general arbiter of trouble, great & small.

Active 4h ago
Joined Jan 20, 2025
Colorado
Powered by