Large language models (LLMs) like BERT and GPT are driving major advances in artificial intelligence, but their size and complexity typically require powerful servers and cloud infrastructure. Running these models directly on devices—without relying on external computation—has remained a difficult technical challenge....
Read more
JUPITER became the world's fourth fastest supercomputer when it debuted last month. Though housed in Germany at the Jülich Supercomputing Center (JSC), Georgia Tech played a supporting role in helping the system land on the latest TOP500 list....
Read more
AI cloud infrastructure gets faster and greener: NPU core improves inference performance by over 60%
The latest generative AI models such as OpenAI's ChatGPT-4 and Google's Gemini 2.5 require not only high memory bandwidth but also large memory capacity. This is why generative AI cloud operating companies like Microsoft and Google purchase hundreds of thousands of NVIDIA GPUs....
Read more
Seoul National University College of Engineering announced that a research team has developed a new hardware security technology based on commercially available 3D NAND flash memory (V-NAND flash memory)....
Read more
Researchers at NYU Tandon School of Engineering have created VeriGen, the first specialized artificial intelligence model successfully trained to generate Verilog code, the programming language that describes how a chip's circuitry functions....
Read more
Artificial intelligence is considered to be computationally and energy-intensive—a challenge for the Internet of Things (IoT), where small, embedded sensors have to make do with limited computing power, little memory and small batteries....
Read more
When it comes to storing images, DNA strands could be a sustainable, stable alternative to hard drives. Researchers at EPFL are developing a new image compression standard designed specifically for this emerging technology....
Read more
BingoCGN, a scalable and efficient graph neural network accelerator that enables inference of real-time, large-scale graphs through graph partitioning, has been developed by researchers at the Institute of Science Tokyo, Japan. This breakthrough framework utilizes an innovative cross-partition message quantization technique and a novel training algorithm to significantly reduce memory...
Read more
A novel power supply technology for 3D-integrated chips has been developed by employing a three-dimensionally stacked computing architecture consisting of processing units placed directly above stacks of dynamic random access memory....
Read more
Researchers at the University of Massachusetts Amherst have pushed forward the development of computer vision with new, silicon-based hardware that can both capture and process visual data in the analog domain. Their work, described in the journal Nature Communications, could ultimately add to large-scale, data-intensive and latency-sensitive computer vision tasks....
Read more
The promise of a new type of computer chip that could reshape the future of artificial intelligence and be more environmentally friendly is explored in a technology review paper published by UC Riverside engineers in the journal Device....
Read more
MIT researchers have designed a compact, low-power receiver for 5G-compatible smart devices that is about 30 times more resilient to a certain type of interference than some traditional wireless receivers....
Read more
The human brain is more powerful and energy-efficient than any computer. Scientists are imitating the way it works to produce better computer chips and help deal with the growing amounts of data generated every day....
Read more
As more connected devices demand an increasing amount of bandwidth for tasks like teleworking and cloud computing, it will become extremely challenging to manage the finite amount of wireless spectrum available for all users to share....
Read more
A team of engineers, AI specialists and chip design researchers at the Chinese Academy of Sciences has designed, built and tested what they are describing as the first AI-based chip design system. The group has published a paper describing their system, called QiMeng, on the arXiv preprint server....
Read more