Postdoctorate Viet Anh Trinh led a project within Strand 1 to develop a novel neural network architecture that can both recognize and generate speech. He has since moved on from iSAT to a role at ...
Imagine unlocking the full potential of a massive language model, tailoring it to your unique needs without breaking the bank or requiring a supercomputer. Sounds impossible? It’s not. Thanks to ...
Fine-tuning large language models (LLMs) might sound like a task reserved for tech wizards with endless resources, but the reality is far more approachable—and surprisingly exciting. If you’ve ever ...
A popular strategy for engaging with generative AI chatbots is to start with a well-crafted prompt. In fact, prompt engineering is an emerging skill for those pursuing career advancement in this age ...
ChemELLM, a 70-billion-parameter LLM tailored for chemical engineering, outperforms leading LLMs (e.g., Deepseek-R1) on ChemEBench across 101 tasks, trained on ChemEData’s 19 billion pretraining and 1 ...
Join the event trusted by enterprise leaders for nearly two decades. VB Transform brings together the people building real enterprise AI strategy. Learn more Microsoft is a major backer and partner of ...
Have you ever watched someone step off a boat, and it immediately started leaning to one side or even capsizing because their weight was keeping it balanced? The same thing can happen in companies.
Chinese scientists recently made history after fine-tuning a billion-parameter artificial intelligence large model on their independently developed quantum computer named Origin Wukong. According to a ...