The Most/Recent Articles

Showing posts with label nvidia. Show all posts
Showing posts with label nvidia. Show all posts

Is Google About to Take on NVidia? Popular AI Startup Anthropic May Switch to Google AI Chips in a Multi-Billion Dollar Deal...


Anthropic is in talks with Google about multi-billion dollar deal for cloud computing services that would see the popular AI startup using Google's tensor processing units, a move that could signal Google's desire to move in to a space currently dominated by NVidia.

Video Courtesy of Bloomberg Tech

NVIDIA Ships Out First Batch of $3999 AI Supercomputers...

Nvidia spark

Nvidia’s long-teased, developer-centric mini-PC is finally leaving preorders and hitting shelves: the DGX Spark goes on sale this week (online at Nvidia and through select retailers such as Micro Center) with a street price that landed around $3,999 in early listings. 

Think compact workstation, not consumer desktop. The Spark packs Nvidia’s new GB10 Grace Blackwell “superchip” — a 20-core Arm-based Grace CPU tightly paired with a Blackwell GPU — into a palm-sized chassis delivering about a petaflop of FP4 AI throughput. It ships with 128 GB of unified LPDDR5x system memory and up to 4 TB NVMe storage, and it’s preconfigured with Nvidia’s AI stack so you can jump into training and fine-tuning mid-sized models locally. Those are not marketing-only numbers: Nvidia positions the Spark for local experimentation on models up to ~200B parameters, and two Sparks linked together can be used for even larger (Nvidia cites ~405B parameter) workloads. 

Under the hood it’s Linux first: DGX Spark runs DGX OS, Nvidia’s Ubuntu-based distro tuned for the Grace/Blackwell stack and preloaded with CUDA, frameworks, and the company’s NIM/Blueprint toolsets — in short, a developer environment that’s meant to feel familiar to anyone who’s spent time on Linux-based model development. That linux/ARM orientation also signals this isn’t optimized as a plug-and-play Windows gaming box; it’s built to be a compact node in an AI workflow. 

Why this matters for the Valley (and who will buy it)

Nvidia is selling the Spark as a way to bring datacenter-class AI tooling to labs, startups, and university benches without immediately routing everything to cloud instances. For teams iterating on model architectures, RLHF loops, or multimodal prototypes, being able to run large-parameter models locally — with 128 GB of coherent memory and GB10’s integrated memory architecture — cuts friction on experiments and iteration cycles. It also enables fast prototyping of models that can later scale to larger DGX setups or cloud clusters. 

Practically: expect early adopters to be small AI teams that value low-latency development cycles, research labs wanting local reproducibility, and edge-oriented startups that prefer on-prem inference for privacy or cost reasons. For generalists and gamers, the Spark’s ARM/Linux DNA and software focus make it a niche purchase. (Enthusiasts will still tinker, but this is not marketed as a consumer GPU box.) 

The ecosystem angle

Nvidia isn’t going it alone: OEMs including Acer, Asus, Dell, Gigabyte, HP, Lenovo, MSI and others are shipping their own DGX Spark variants and the larger DGX Station desktop tower — the Station uses the beefier GB300/Grace Blackwell Ultra silicon and targets heavier local training workloads. That OEM breadth makes Spark part of a broader push to make DGX software + silicon a platform developers can buy from many vendors. 

Networking and scale matter here: Spark includes high-speed ConnectX networking (and QSFP/200G options) so two Sparks can cooperate as a small cluster for models larger than what a single unit can handle — a practical way to prototype distributed inference without immediately renting a rack. 

Caveats and hard truths

Software compatibility. The Spark’s Arm-centric platform and DGX OS make the CUDA/tooling story smooth for supported stacks, but expect some extra work for niche toolchains or Windows-first workflows. If your pipelines assume x86 Windows tooling, factor in integration time. 

Thermals & real-world throughput. A petaflop of FP4 in a tiny chassis is impressive, but sustained training on huge models still favors larger systems (and racks) with beefier cooling and power budgets. The Spark is best framed as a development node and prototyping workhorse. 

Pricing vs cloud. At ~$3,999 per node (retail listings), teams need to weigh capital expenditure against cloud flexibility — Spark is most compelling when local iteration speed, data privacy, or long-term TCO favor owning hardware. 

Watch how quickly third-party software (e.g., Docker Model Runner, popular MLOps stacks, and smaller OSS frameworks) certify Spark and DGX OS workflows; that will determine the friction for real-world adoption. Docker has already flagged support, which is a positive sign for quick onboarding. 

Nvidia’s wider silicon roadmap: there are signals (and comments from Nvidia leadership) that similar GB10/N1 designs could make their way into more consumer-facing devices down the line, and MediaTek collaboration threads hint at broader ARM partnerships — keep an eye on where Nvidia pushes ARM into the mainstream PC market. 

Final Thought

Nvidia’s DGX Spark is a tidy, ambitious product: it distills a lot of datacenter capability into a desktop footprint with a clear audience in mind — developers iterating on large models, labs that need local reproducibility, and startups that want a deterministic development environment. It’s not a replacement for scale-out clusters, but it’s a meaningful step toward decentralizing serious AI development outside the data center — provided your team is ready for Linux/ARM toolchains and the upfront hardware buy.

-----------
Author: Trevor Kingsley
Tech News CITY /New York Newsroom

Alibaba's New AI Chip: China Sends it's Corporate Goliath to Take Another Swing at Nvidia's Market Domination...

Alibaba VS Nvidia GPU chips

Alibaba has entered the competitive AI chip sector with a new homegrown processor, creating significant buzz in the industry. This development has already impacted the market, causing NVIDIA's stock to drop over 3%, while Alibaba’s shares surged by 12%.

The Facts Behind the Chip

Recent reports indicate that Alibaba is testing a new AI chip specifically designed for AI inference. 

Unlike Alibaba's earlier chips, which were produced by Taiwan's TSMC, this new processor is being manufactured domestically by a Chinese company. This shift highlights a commitment to local production. The chip is expected to be more versatile than previous models, capable of handling a wider range of AI tasks.

The Timing: A Strategic Move

Alibaba's decision to develop this chip is not just a casual venture; it is a strategic response to geopolitical tensions and trade restrictions that have made it challenging for Chinese companies to access NVIDIA's advanced technology.

With U.S. restrictions limiting access to NVIDIA's high-end chips, Alibaba is taking the initiative to develop its own solutions. The company has committed to investing at least 380 billion Chinese yuan (approximately $53.1 billion) in AI development over the next three years, signaling its serious intent.

Strategic Focus: Internal Use

Rather than selling the chip commercially, Alibaba plans to use it exclusively for its cloud services, allowing customers to rent computing power rather than purchase hardware. This approach leverages Alibaba's existing cloud infrastructure, which has already demonstrated impressive growth, with a 26% year-over-year increase and consistent triple-digit growth in AI-related product revenue.

Technical Details: What We Still Don’t Know

While the announcement is exciting, specific performance details remain unclear. Questions about how this chip compares to NVIDIA's offerings—such as speed and efficiency—are still unanswered. Additionally, the timeline for its market readiness is uncertain, as Alibaba has a history of taking time to launch new products.

The Bigger Picture: A Shift in Tech Independence

This development reflects a broader trend of Chinese tech companies striving for independence from American technology. Alibaba's chip initiative is part of a larger strategy to create a self-sufficient technological ecosystem. While financial investment is crucial, building competitive semiconductors also requires advanced technical expertise and long-term partnerships.

Looking Ahead

In the short term, Alibaba may remain cautious about releasing performance metrics until they are confident in the chip's capabilities. If the chip performs well, Alibaba could expand its internal use and potentially license the technology to other Chinese companies. In the long term, this could either mark a significant advancement for China's semiconductor industry or serve as a costly learning experience.

The Nvidia Wildcard

There's one chip we know even less about than Alibaba's - and that's Nvidia's next chip, code named 'Rubin' we talked about here.  At least according to rumors, it may double the performance of their newest, publicly available chips. Considering it's unlikely Alibaba has been able to match Nivdia's current performance, doubling that would leave any competitor in the dust.  

In any other circumstance this would sound far-fetched, but when it comes to GPU's Nvidia has such a head start and is credited with inventing a large portion of how these chips function, when it comes to development their advantage can't be dismissed. 

Conclusion

Regardless of the outcome, Alibaba's new chip signifies a determined effort by Chinese tech firms to shape their own technological future. As the AI chip competition continues, the stakes are high, with significant implications for both domestic and global markets. The world will be watching closely to see how this unfolds. What are your thoughts? Will Alibaba's efforts succeed, or is NVIDIA's position too strong to challenge? Only time will tell.
_________________

Author: Ross Davis
Silicon Valley Newsroom | Tech News CITY