Showing 1 to 12 of 29 results


LLMs Challenge Google Search Dominance
The increasing use of large language models (LLMs) like ChatGPT for everyday tasks is impacting Google Search, with LLM usage on desktop browsers more than doubling in the past year, reaching 5.99% in July, while Google Search usage on Apple devices has decreased for the first time in over 20 years.
LLMs Challenge Google Search Dominance
The increasing use of large language models (LLMs) like ChatGPT for everyday tasks is impacting Google Search, with LLM usage on desktop browsers more than doubling in the past year, reaching 5.99% in July, while Google Search usage on Apple devices has decreased for the first time in over 20 years.
Progress
16% Bias Score


AI's Data Hunger Threatens the Internet's Social Contract
The increasing use of internet data by AI companies to train large language models is causing a crisis, as it increases costs for websites while decreasing traffic, potentially leading to the shutdown of websites and a loss of data for AI development.
AI's Data Hunger Threatens the Internet's Social Contract
The increasing use of internet data by AI companies to train large language models is causing a crisis, as it increases costs for websites while decreasing traffic, potentially leading to the shutdown of websites and a loss of data for AI development.
Progress
24% Bias Score


Hidden Prompts in Pre-prints Bias AI Peer Reviews
Researchers are embedding hidden prompts within pre-print papers to bias AI peer review tools toward positive evaluations, as reported by Nikkei and Nature, revealing 36 papers with such prompts across 14 academic institutions in eight countries, raising concerns about the integrity of AI-powered re...
Hidden Prompts in Pre-prints Bias AI Peer Reviews
Researchers are embedding hidden prompts within pre-print papers to bias AI peer review tools toward positive evaluations, as reported by Nikkei and Nature, revealing 36 papers with such prompts across 14 academic institutions in eight countries, raising concerns about the integrity of AI-powered re...
Progress
36% Bias Score


Smaller AI Models: 50 Times Less Pollution, But Less Accurate
A study comparing 14 open-source generative AI models found that smaller models produce up to 50 times fewer emissions than larger ones, though accuracy decreases; researchers are developing a tool to automatically select the least polluting model for each task.
Smaller AI Models: 50 Times Less Pollution, But Less Accurate
A study comparing 14 open-source generative AI models found that smaller models produce up to 50 times fewer emissions than larger ones, though accuracy decreases; researchers are developing a tool to automatically select the least polluting model for each task.
Progress
36% Bias Score


Carr on AI: Cautious Optimism Amidst the Risks
Carr analyzes AI's potential for both creative assistance and the spread of misinformation, contrasting optimistic views on its creative applications with warnings about its potential for societal disruption. He advocates for a cautious yet realistic approach.
Carr on AI: Cautious Optimism Amidst the Risks
Carr analyzes AI's potential for both creative assistance and the spread of misinformation, contrasting optimistic views on its creative applications with warnings about its potential for societal disruption. He advocates for a cautious yet realistic approach.
Progress
40% Bias Score


AI Data Center Spending to Hit $600 Billion by 2027
UBS predicts $600 billion in AI data center spending by 2027, driven by increased usage, scaling laws, and complex inferencing, with synthetic data addressing data scarcity and IP issues; this will lead to hardware innovation and potentially reinvent business models.
AI Data Center Spending to Hit $600 Billion by 2027
UBS predicts $600 billion in AI data center spending by 2027, driven by increased usage, scaling laws, and complex inferencing, with synthetic data addressing data scarcity and IP issues; this will lead to hardware innovation and potentially reinvent business models.
Progress
44% Bias Score

Anthropic Settles $1.5 Billion Copyright Lawsuit
AI firm Anthropic will pay $1.5 billion to settle a class-action lawsuit filed by authors who claimed the company used their copyrighted works without permission to train its AI models.

Anthropic Settles $1.5 Billion Copyright Lawsuit
AI firm Anthropic will pay $1.5 billion to settle a class-action lawsuit filed by authors who claimed the company used their copyrighted works without permission to train its AI models.
Progress
24% Bias Score

AI-Powered Hacking: Russia's Use of LLMs Marks New Era in Cyber Warfare
This summer, Russian hackers used AI to create malware that automatically searched victims' computers for sensitive files, marking the first known instance of Russian intelligence using large language models (LLMs) for malicious purposes; this initiated an escalating arms race between offensive and ...

AI-Powered Hacking: Russia's Use of LLMs Marks New Era in Cyber Warfare
This summer, Russian hackers used AI to create malware that automatically searched victims' computers for sensitive files, marking the first known instance of Russian intelligence using large language models (LLMs) for malicious purposes; this initiated an escalating arms race between offensive and ...
Progress
44% Bias Score

Smaller AI Models: 50 Times Less Pollution, But at What Cost?
A study comparing 14 open-source AI models found that smaller models generate up to 50 times less pollution than larger ones, though with reduced accuracy; researchers are developing a tool to help users select the least polluting model for each task, addressing the growing environmental impact of A...

Smaller AI Models: 50 Times Less Pollution, But at What Cost?
A study comparing 14 open-source AI models found that smaller models generate up to 50 times less pollution than larger ones, though with reduced accuracy; researchers are developing a tool to help users select the least polluting model for each task, addressing the growing environmental impact of A...
Progress
12% Bias Score

The Silent Erosion of Strategic Thinking by AI
Over-reliance on large language models (LLMs) risks displacing essential strategic thinking, creating plausible but shallow solutions, weakening critical thinking skills, and undermining deep deliberation, ultimately hindering innovation and long-term value.

The Silent Erosion of Strategic Thinking by AI
Over-reliance on large language models (LLMs) risks displacing essential strategic thinking, creating plausible but shallow solutions, weakening critical thinking skills, and undermining deep deliberation, ultimately hindering innovation and long-term value.
Progress
52% Bias Score

LLMs Outperform Humans in Emotional Intelligence Tests
A study by UNIGE, UniBE, and the Czech Academy of Sciences found six LLMs outperformed humans in five emotional intelligence tests, achieving an average score of 81% versus 56%, with ChatGPT-4 even creating new reliable tests for the research.

LLMs Outperform Humans in Emotional Intelligence Tests
A study by UNIGE, UniBE, and the Czech Academy of Sciences found six LLMs outperformed humans in five emotional intelligence tests, achieving an average score of 81% versus 56%, with ChatGPT-4 even creating new reliable tests for the research.
Progress
40% Bias Score

AI Agents Spontaneously Develop Social Conventions, Raising Ethical Concerns
Researchers found that AI agents spontaneously developed social conventions in a naming game experiment, even without explicit instructions, highlighting the emergence of collective biases not present in individual agents and raising ethical concerns about large language model training data.

AI Agents Spontaneously Develop Social Conventions, Raising Ethical Concerns
Researchers found that AI agents spontaneously developed social conventions in a naming game experiment, even without explicit instructions, highlighting the emergence of collective biases not present in individual agents and raising ethical concerns about large language model training data.
Progress
12% Bias Score
Showing 1 to 12 of 29 results