Letting a language model do the work of thinking is like building a house and using a circular saw to put nails in.
It will do it but you should not trust the results.
It is not Google. It can, will, and has made up facts as long as it fits the format expected
Not at the very least proof reading and fact checking the output is beyond lazy and a terrible use of a tool. Using it to create the end product instead of as a tool to use in creation of an end product are two very different things.
Why do you mind that?
Have you seen that legal brief?
No. Communicate please and we can have a real conversation.
Letting a language model do the work of thinking is like building a house and using a circular saw to put nails in. It will do it but you should not trust the results.
It is not Google. It can, will, and has made up facts as long as it fits the format expected
Not at the very least proof reading and fact checking the output is beyond lazy and a terrible use of a tool. Using it to create the end product instead of as a tool to use in creation of an end product are two very different things.