Author: advancedainews
Image: iStock/BeeBright A bipartisan report, recently issued by the US Select Committee on the Chinese Communist Party (CCP), accuses DeepSeek of a series of subversive, illegal, and immoral practices. Moreover, the tech giant NVIDIA is also catching the ire of US government officials for supplying DeepSeek with the chips needed to create the AI models. Investigating DeepSeek The report, titled “DeepSeek Unmasked: Exposing the CCP’s Latest Tool for Spying, Stealing, and Subverting U.S. Export Control Restrictions,” was published in April 2025. It levies numerous accusations against DeepSeek, including: Actively suppressing more than 85% of responses that are related to human…
Startup Deep Cogito Inc. launched today with a series of language models that it claims can outperform comparably sized open-source alternatives. According to TechCrunch, the company was founded last June by former Google LLC staffers Drishan Arora and Dhruv Malhotra. Arora worked as a senior software engineer at the search giant. Malhotra, in turn, was a product manager at the Google DeepMind machine learning lab. The duo have raised an undisclosed amount of funding from South Park Commons. Deep Cogito’s lineup of open-source language models is known as the Cogito v1 series. The algorithms are available in five sizes ranging…
On 26 March, Suzhou Pudu Co-Intelligence Technology Company, a joint venture between Xi’an Jiaotong-Liverpool University (XJTLU) and Baidu Group, was launched as China’s first AI-focused joint venture co-founded by Baidu and a university. SUZHOU, China, March 31, 2025 /PRNewswire/ — On 26 March, Suzhou Pudu Co-Intelligence Technology Company, a joint venture between Xi’an Jiaotong-Liverpool University (XJTLU) and Baidu Group, was launched as China’s first AI-focused joint venture co-founded by Baidu and a university. AI+education: Redefining learning Through AI-driven innovation, Pudu Co-Intelligence seeks to transform the whole education value chain, empower industrial evolution and cultivate localised service ecosystems. The company will…
NVIDIA’s Wade Vinson during his keynote at Data Center World 2025. Image: Drew Robb/TechnologyAdvice NVIDIA kicked off the Data Center World 2025 event this week in Washington, D.C., with a bold vision for the future of AI infrastructure. In his keynote, Wade Vinson, NVIDIA’s chief data center engineer, introduced the concept of AI-scale data centers; these massive, energy-efficient facilities would meet the soaring demand of accelerated computing. NVIDIA envisions sprawling “AI factories” powered by Blackwell GPUs and DGX SuperPODs, supported by advanced cooling and power systems by Vertiv and Schneider Electric. “There is no doubt that AI factories are a…
A high-profile legal case has unearthed a trove of internal Meta communications, and one particular document has caught the eye of some AI researchers.This reveals new insights into how models are built and could influence who gets to share in the spoils of this new technology.Buried in these court filings is a description of how Meta researchers used a process called ablation to identify which data helped improve the company’s Llama AI models.Ablation is a medical technique that purposely destroys tissue to improve things like brain function. In AI, it involves removing parts of a system to study how those…
Anthropic will roll out its Claude artificial intelligence application to U.S. government agencies under a partnership with Palantir Technologies (PLTR, Financials), the companies announced. Claude for Enterprise will be made available through Palantir’s FedStart program, allowing federal agencies to access the AI tool under Federal Risk and Authorization Management Program High and Department of Defense Impact Level 5 security standards. The deployment is expected in the coming months. FedStart helps companies meet government compliance standards quickly by running products within Palantir’s accredited environment. Anthropic said the partnership will allow federal workers to use Claude for writing, data analysis and complex…
“We’ve been really pushing on ‘thinking,’” says Jack Rae, a principal research scientist at DeepMind. Such models, which are built to work through problems logically and spend more time arriving at an answer, rose to prominence earlier this year with the launch of the DeepSeek R1 model. They’re attractive to AI companies because they can make an existing model better by training it to approach a problem pragmatically. That way, the companies can avoid having to build a new model from scratch. When the AI model dedicates more time (and energy) to a query, it costs more to run. Leaderboards…
Illustration of AI models racing on a trackAI-generated image via DALL·E On the heels of the recent GPT-4.1 announcement, OpenAI has unveiled a whole set of new, more powerful models. As someone who relies on AI daily for everything from research questions to finding local sports information, I couldn’t wait to put these models through their paces. Below are my first impressions, along with some practical comparisons. Before diving in, an important caveat: AI performance is non-deterministic and highly use-case specific. In simpler terms, Your Mileage May Vary. Don’t take this (or any other) article as the final word—instead, test…
Call centers have long been front and center in providing customer service and support — in fact, IBIS World estimates that there are 36,787 customer care centers employing 305,000 individuals in the United States.Despite the widespread proliferation of call centers, there is no denying that not all centers operate at the same level of efficiency. This can have a direct impact on the customer experience and influence the public’s perception of the business. However, the introduction of AI is having a noteworthy impact in this space as well.As Han Butler, president and co-founder of ROI CX Solutions, a company that…
[Submitted on 30 May 2024 (v1), last revised 18 Apr 2025 (this version, v3)] View a PDF of the paper titled Is In-Context Learning Sufficient for Instruction Following in LLMs?, by Hao Zhao and 3 other authors View PDF HTML (experimental) Abstract:In-context learning (ICL) allows LLMs to learn from examples without changing their weights: this is a particularly promising capability for long-context LLMs that can potentially learn from many examples. Recently, Lin et al. (2024) proposed URIAL, a method using only three in-context examples to align base LLMs, achieving non-trivial instruction following performance. In this work, we show that, while…