Large Language Models up to 4x Faster on RTX With TensorRT-LLM for Windowsblogs.nvidia.comexternal-linkcross-posted to: technews@radiation.party ijeff ( @ijeff@lemdro.id ) M AI Stuff@lemdro.idEnglish • 1 year ago message-square1fedilinkarrow-up18
arrow-up18external-linkLarge Language Models up to 4x Faster on RTX With TensorRT-LLM for Windowsblogs.nvidia.com ijeff ( @ijeff@lemdro.id ) M AI Stuff@lemdro.idEnglish • 1 year ago message-square1fedilinkcross-posted to: technews@radiation.party
minus-square ijeff ( @ijeff@lemdro.id ) OPlinkfedilinkEnglish1•edit-21 year agoTheir inference prowess has been keeping me on Nvidia. Really wish AMD would step up its development in this area.
Their inference prowess has been keeping me on Nvidia. Really wish AMD would step up its development in this area.