Tether has introduced a cross-platform framework that reduces the cost and hardware requirements of AI model training, enabling advanced LLMs to be fine-tuned onTether has introduced a cross-platform framework that reduces the cost and hardware requirements of AI model training, enabling advanced LLMs to be fine-tuned on

Tether Launches Cross-Platform BitNet LoRA Framework Enabling Billion-Parameter AI Training And Inference On Consumer Devices

2026/03/18 15:00
4 min read
For feedback or concerns regarding this content, please contact us at crypto.news@mexc.com
Tether Launches Cross-Platform BitNet LoRA Framework Enabling Billion-Parameter AI Training And Inference On Consumer Devices

USDT stablecoin issuer Tether announced the launch of what it describes as the first cross-platform LoRA fine-tuning framework designed for Microsoft BitNet models, which are based on 1-bit large language model architecture. The capability is integrated into its QVAC Fabric system and is reported to significantly reduce both memory usage and computational demands. According to the company, this development enables large-scale language models, including those with billions of parameters, to be fine-tuned using widely available consumer hardware such as laptops, standard graphics processing units, and modern smartphones.

The development and maintenance of artificial intelligence systems have traditionally required enterprise-grade hardware, particularly specialized NVIDIA infrastructure or cloud-based environments. These requirements have contributed to high operational costs, limiting access to advanced AI development primarily to large organizations with substantial financial resources and access to specialized computing systems.

Tether stated that its QVAC Fabric large language model, enhanced by the newly introduced BitNet-based framework, addresses these limitations by supporting cross-platform LoRA fine-tuning and accelerating inference across a range of heterogeneous consumer GPUs. These include hardware from Intel, AMD, and Apple Silicon, among others. As a result, users are able to train and customize AI models directly on commonly available consumer devices rather than relying on centralized infrastructure.

The company reported that its engineering team has successfully demonstrated BitNet fine-tuning on mobile graphics processing units for the first time, including platforms such as Adreno, Mali, and Apple Bionic GPUs. Internal testing indicated that a 125 million-parameter BitNet model could be fine-tuned in approximately ten minutes on a Samsung S25 device equipped with an Adreno GPU using a biomedical dataset consisting of roughly 300 documents, or about 18,000 tokens. For a 1 billion-parameter model, the same dataset required approximately one hour and eighteen minutes on the Samsung S25 and one hour and forty-five minutes on an iPhone 16. The company also reported that it was able to extend testing to models as large as 13 billion parameters on the iPhone 16 under maximum device capacity conditions.

Advancements In Edge-Based AI Training And Performance Optimization

Further findings suggest that the framework can support fine-tuning of models up to twice the size of comparable non-BitNet models operating under Q4 quantization on edge devices. This outcome is attributed to the reduced memory footprint associated with the BitNet architecture.

In addition to improvements in training, the framework also demonstrates enhanced inference performance. Tests conducted on mobile devices indicated that BitNet models perform substantially faster when executed on GPUs, with processing speeds ranging from two to eleven times higher than CPU-based execution. These results indicate that mobile GPUs are increasingly capable of handling workloads that previously required specialized hardware or data center-level resources.

The system also shows notable gains in memory efficiency. Benchmark data suggests that a BitNet-1B model using TQ1_0 configuration requires up to 77.8 percent less VRAM compared to a 16-bit Gemma-3-1B model and 65.6 percent less than a 16-bit Qwen3-0.6B model during both inference and LoRA fine-tuning processes. These reductions provide additional capacity for running larger models and enabling personalization features on hardware that would previously have been considered insufficient.

Tether further indicated that the framework introduces LoRA fine-tuning capabilities for 1-bit large language models on non-NVIDIA hardware for the first time, extending compatibility to AMD, Intel, Apple Silicon, and mobile GPU platforms. By reducing reliance on specialized infrastructure and cloud services, the approach allows sensitive data to remain stored locally on user devices. The company noted that this efficiency may also support the development of federated learning systems, in which models can be trained collaboratively across distributed devices while maintaining data privacy and minimizing dependence on centralized systems.

The post Tether Launches Cross-Platform BitNet LoRA Framework Enabling Billion-Parameter AI Training And Inference On Consumer Devices appeared first on Metaverse Post.

Market Opportunity
CROSS Logo
CROSS Price(CROSS)
$0.06814
$0.06814$0.06814
-0.64%
USD
CROSS (CROSS) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact crypto.news@mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Let insiders trade – Blockworks

Let insiders trade – Blockworks

The post Let insiders trade – Blockworks appeared on BitcoinEthereumNews.com. This is a segment from The Breakdown newsletter. To read more editions, subscribe ​​“The most valuable commodity I know of is information.” — Gordon Gekko, Wall Street Ten months ago, FBI agents raided Shayne Coplan’s Manhattan apartment, ostensibly in search of evidence that the prediction market he founded, Polymarket, had illegally allowed US residents to place bets on the US election. Two weeks ago, the CFTC gave Polymarket the green light to allow those very same US residents to place bets on whatever they like. This is quite the turn of events — and it’s not just about elections or politics. With its US government seal of approval in hand, Polymarket is reportedly raising capital at a valuation of $9 billion — a reflection of the growing belief that prediction markets will be used for much more than betting on elections once every four years. Instead, proponents say prediction markets can provide a real service to the world by providing it with better information about nearly everything. I think they might, too — but only if insiders are free to participate. Yesterday, for example, Polymarket announced new betting markets on company earnings reports, with a promise that it would improve the information that investors have to work with.  Instead of waiting three months to find out how a company is faring, investors could simply watch the odds on Polymarket.  If the probability of an earnings beat is rising, for example, investors would know at a glance that things are going well. But that will only happen if enough of the people betting actually know how things are going. Relying on the wisdom of crowds to magically discern how a business is doing won’t add much incremental knowledge to the world; everyone’s guesses are unlikely to average out to the truth. If…
Share
BitcoinEthereumNews2025/09/18 05:16
The Linux Foundation has been awarded $12.5 million to address low-quality security reports generated by AI.

The Linux Foundation has been awarded $12.5 million to address low-quality security reports generated by AI.

PANews reported on March 18 that the Linux Foundation 's Alpha-Omega project and OpenSSF have launched a new initiative, receiving a total of $ 12.5 million in
Share
PANews2026/03/18 17:11
Finastra Strengthens AI Capabilities with New Center of Excellence and Leadership Appointment

Finastra Strengthens AI Capabilities with New Center of Excellence and Leadership Appointment

Company Expands Hiring in Atlanta and India Artificial intelligence is creating new opportunities across the financial services industry, helping institutions improve
Share
Globalfintechseries2026/03/18 16:23