It also has a 1 million token input context window! That would roughly cover the entirety of J.R.R. Tolkien's "The Lord of the Rings: The Fellowship of the Ring", or approximately 400 pages of text.


The 12B model responds almost instantly and is excellent for good-quality, rapid example work.
The 24B model takes about 30 seconds to respond, but it has deep, obscure, nuanced knowledge of the world. I would have to spend 5 times more to do the same with NVidia GPUs.
Another benefit of using the "Dolphin" is that it is uncensored, which gives me direct answers to my questions without trying to "protect me" from facts like "Tiananmen Square protests of 1980", or any other enforced ideology.
No comments:
Post a Comment
Please be polite.