If I regurgiate something I read in copyrighted book without proper license that also would be theft, no distinction there.
I'm not distributing my brain, at least same (but probably more restrictive) should apply to models - training is okay, but using and distributing should be limited by copyright
It would not be illegal based on fair use (though you have to be careful there also), but if you try to regurgiate large portions of the book then it would be. And we do know that models regurgiate training material verbatim (Copilot)
Wait until you hear about frankenmodels. You rip parts of one model (often attention heads) and transplant them in another and somehow that produces coherent results! Witchcraft
With, but it's still bonkers that it works so well
>Also is there a practical motivation for creating them?
You could get in-between model sizes (like 20b instead of 13b or 34b). Before better quantization it was useful for inference (if you are unlucky with vram size), but now I see this being useful only for training because you can't train on quants
Personally, I have found that Mistral 7B (with its native 8K context, and decent results stretched out even more) is performing much better than llama 13B tunes for storytelling, where that long context is really important.
And I think the optimized backends should implement that sliding 16k context soon...
Anyway, point is a huge context really helps certain types of queries, and VRAM usage is reasonable with a 7B model.
You could go even lower with smaller quantization if necessary. I personally wouldn't use anything smaller than 7B and Mistral already pushing it in coherence. Overall it depends on your use case, not everyone needs smart models, or large context that sometimes takes half of required memory.
Codellama is also surprisingly good even for non-coding tasks
"Any loan or lease of defense articles to the Government of Ukraine under paragraph (1) shall be subject to all applicable laws concerning the return of and reimbursement and repayment for defense articles loan or leased to foreign governments."
Can you provide a source for your statement that these were donations?
Most aid takes the form of cash for humanitarian support, ~disposing of~ donating old hardware to Ukraine, or a loan for the express intent of buying a modern system.
The vast majority of the "billions" given to Ukraine is made up of stuff that we actively are trying to throw away.
IIRC temperature is proportional to velocity squared so increase in simulation speed will result in squared increase in temperature. But even if we account for that we are not dealing with ideal gas in real world, there bound to be collisions that break molecules apart or even start nuclear fusion at certain speeds
Sure! It's just a classification of events that each person has.
Testible prediction? Socially close people will have similar classifications as opposed to randomly chosen ones (I will leave defining metric space over possible classifications as an exercise to the reader)
I'm not distributing my brain, at least same (but probably more restrictive) should apply to models - training is okay, but using and distributing should be limited by copyright