Y'all heard about ChatGPT yet? AI instantly generates question answers, entire essays etc.

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,448
Reputation
8,519
Daps
160,190
Wrote a grant for a non profit if approved I get a cut, should be 10k or more

I'm starting to think bard is better than gpt. What you think @bnew ?
thats dope :obama:

don't handicap yourself into using only one AI model. they all have their strengths and weaknesses. I use multiple and compare the responses.
PROTIP: get a clipboard manager to be more productive

I found that chatgpt and bing which uses a modified version are great for coding.

Bard can explain a lot of of code too as well as break it down in many ways, I really like their multiple drafts options and they added the ability to make the response more casual in some cases, longer or more professional.

Bard can also transcribe text from images which i think is only available to chatgpt-4 subscribers.

bard hallucinates a lot and omit can code but it's also great for summarizing tasks in any kind of way I want it told to me.
 

IIVI

Superstar
Joined
Mar 11, 2022
Messages
11,556
Reputation
2,744
Daps
39,246
Reppin
Los Angeles


I built a software career barely visiting StackOverflow, really no need to use it anymore.

If anybody has tried the official Code Interpreter plugin yet, shyt is :wow:

Literally drag and drop in your .zipped code repository and ask your questions. It's only in beta but incredible.
 
Last edited:

Guvnor

Da Speculative Spectacle®
Joined
Jan 17, 2017
Messages
23,151
Reputation
4,665
Daps
33,336
Reppin
BKLYN
I'm starting grad school, do you guys think ChatGPT can be used to do research for my essays from scholarly sources? Can it highlight key findings and even write the sources in APA format? Is there a better software specifically for researchers?

Also is it free?
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,448
Reputation
8,519
Daps
160,190
I'm starting grad school, do you guys think ChatGPT can be used to do research for my essays from scholarly sources? Can it highlight key findings and even write the sources in APA format? Is there a better software specifically for researchers?

Also is it free?

users on reddit have been mentioning using it for similar purposes. go for it but don't just copy paste the response. modify it or add personality or tone to the output using chatgpt to evade ai detectors.. the A.I detectors are B.S anyway but you don't want the hassle of that accusation anyway.
 

thaKEAF

#grizzlies #titans
Supporter
Joined
May 1, 2012
Messages
38,026
Reputation
9,062
Daps
111,147
Reppin
Memphis
How did you go about that? Did you just copy and paste it in there then ask it to help or what was the process? I'm starting grad school and working full time so I'm gonna need some help.

I asked it to create resume bullet points for specific jobs I’ve worked and to tailor it to the field I’m trying to move to.
 

Guvnor

Da Speculative Spectacle®
Joined
Jan 17, 2017
Messages
23,151
Reputation
4,665
Daps
33,336
Reppin
BKLYN
users on reddit have been mentioning using it for similar purposes. go for it but don't just copy paste the response. modify it or add personality or tone to the output using chatgpt to evade ai detectors.. the A.I detectors are B.S anyway but you don't want the hassle of that accusation anyway.
Yeah but I don't even want to use to write the essay just gather my sources for me.

I asked it to create resume bullet points for specific jobs I’ve worked and to tailor it to the field I’m trying to move to.
Salute I'm going to give that a try.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,448
Reputation
8,519
Daps
160,190
Yeah but I don't even want to use to write the essay just gather my sources for me.


Salute I'm going to give that a try.
any "sources" you gather should have to be individually confirmed since chatgpt hallucinates a lot. Bing chat actually links to things it cites.
 

Guvnor

Da Speculative Spectacle®
Joined
Jan 17, 2017
Messages
23,151
Reputation
4,665
Daps
33,336
Reppin
BKLYN
any "sources" you gather should have to be individually confirmed since chatgpt hallucinates a lot. Bing chat actually links to things it cites.
Really hallucinates how? So in other words I'd have to verify and read through them myself?
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,448
Reputation
8,519
Daps
160,190
Really hallucinates how? So in other words I'd have to verify and read through them myself?

yeah you don't want to end up like this guy.


Here’s What Happens When Your Lawyer Uses ChatGPT​

A lawyer representing a man who sued an airline relied on artificial intelligence to help prepare a court filing. It did not go well.


28airline-suit1-superJumbo.jpg

As an Avianca flight approached Kennedy International Airport in New York, a serving cart collision began a legal saga, prompting the question: Is artificial intelligence so smart?Credit...Nicolas Economou/NurPhoto, via Getty Images


By Benjamin Weiser
May 27, 2023

The lawsuit began like so many others: A man named Roberto Mata sued the airline Avianca, saying he was injured when a metal serving cart struck his knee during a flight to Kennedy International Airport in New York.

When Avianca asked a Manhattan federal judge to toss out the case, Mr. Mata’s lawyers vehemently objected, submitting a 10-page brief that cited more than half a dozen relevant court decisions. There was Martinez v. Delta Air Lines, Zicherman v. Korean Air Lines and, of course, Varghese v. China Southern Airlines, with its learned discussion of federal law and “the tolling effect of the automatic stay on a statute of limitations.”

There was just one hitch: No one — not the airline’s lawyers, not even the judge himself — could find the decisions or the quotations cited and summarized in the brief.

That was because ChatGPT had invented everything.

The lawyer who created the brief, Steven A. Schwartz of the firm Levidow, Levidow & Oberman, threw himself on the mercy of the court on Thursday, saying in an affidavit that he had used the artificial intelligence program to do his legal research — “a source that has revealed itself to be unreliable.”


Mr. Schwartz, who has practiced law in New York for three decades, told Judge P. Kevin Castel that he had no intent to deceive the court or the airline. Mr. Schwartz said that he had never used ChatGPT, and “therefore was unaware of the possibility that its content could be false.”

He had, he told Judge Castel, even asked the program to verify that the cases were real.

It had said yes.

Mr. Schwartz said he “greatly regrets” relying on ChatGPT “and will never do so in the future without absolute verification of its authenticity.”

Judge Castel said in an order that he had been presented with “an unprecedented circumstance,” a legal submission replete with “bogus judicial decisions, with bogus quotes and bogus internal citations.” He ordered a hearing for June 8 to discuss potential sanctions.

As artificial intelligence sweeps the online world, it has conjured dystopian visions of computers replacing not only human interaction, but also human labor. The fear has been especially intense for knowledge workers, many of whom worry that their daily activities may not be as rarefied as the world thinks — but for which the world pays billable hours.


Stephen Gillers, a legal ethics professor at New York University School of Law, said the issue was particularly acute among lawyers, who have been debating the value and the dangers of A.I. software like ChatGPT, as well as the need to verify whatever information it provides.

“The discussion now among the bar is how to avoid exactly what this case describes,” Mr. Gillers said. “You cannot just take the output and cut and paste it into your court filings.”


The real-life case of Roberto Mata v. Avianca Inc. shows that white-collar professions may have at least a little time left before the robots take over.

It began when Mr. Mata was a passenger on Avianca Flight 670 from El Salvador to New York on Aug. 27, 2019, when an airline employee bonked him with the serving cart, according to the lawsuit. After Mr. Mata sued, the airline filed papers asking that the case be dismissed because the statute of limitations had expired.

In a brief filed in March, Mr. Mata’s lawyers said the lawsuit should continue, bolstering their argument with references and quotes from the many court decisions that have since been debunked.


Soon, Avianca’s lawyers wrote to Judge Castel, saying they were unable to find the cases that were cited in the brief.

When it came to Varghese v. China Southern Airlines, they said they had “not been able to locate this case by caption or citation, nor any case bearing any resemblance to it.”

They pointed to a lengthy quote from the purported Varghese decision contained in the brief. “The undersigned has not been able to locate this quotation, nor anything like it in any case,” Avianca’s lawyers wrote.

Indeed, the lawyers added, the quotation, which came from Varghese itself, cited something called Zicherman v. Korean Air Lines Co. Ltd., an opinion purportedly handed down by the U.S. Court of Appeals for the 11th Circuit in 2008. They said they could not find that, either.

Judge Castel ordered Mr. Mata’s attorneys to provide copies of the opinions referred to in their brief. The lawyers submitted a compendium of eight; in most cases, they listed the court and judges who issued them, the docket numbers and dates.


The copy of the supposed Varghese decision, for example, is six pages long and says it was written by a member of a three-judge panel of the 11th Circuit. But Avianca’s lawyers told the judge that they could not find that opinion, or the others, on court dockets or legal databases.

Bart Banino, a lawyer for Avianca, said that his firm, Condon & Forsyth, specialized in aviation law and that its lawyers could tell the cases in the brief were not real. He added that they had an inkling a chatbot might have been involved.

Mr. Schwartz did not respond to a message seeking comment, nor did Peter LoDuca, another lawyer at the firm, whose name appeared on the brief.

Mr. LoDuca said in an affidavit this week that he did not conduct any of the research in question, and that he had “no reason to doubt the sincerity” of Mr. Schwartz’s work or the authenticity of the opinions.

ChatGPT generates realistic responses by making guesses about which fragments of text should follow other sequences, based on a statistical model that has ingested billions of examples of text pulled from all over the internet. In Mr. Mata’s case, the program appears to have discerned the labyrinthine framework of a written legal argument, but has populated it with names and facts from a bouillabaisse of existing cases.


Judge Castel, in his order calling for a hearing, suggested that he had made his own inquiry. He wrote that the clerk of the 11th Circuit had confirmed that the docket number printed on the purported Varghese opinion was connected to an entirely different case.

Calling the opinion “bogus,” Judge Castel noted that it contained internal citations and quotes that, in turn, were nonexistent. He said that five of the other decisions submitted by Mr. Mata’s lawyers also appeared to be fake.

On Thursday, Mr. Mata’s lawyers offered affidavits containing their version of what had happened.

Mr. Schwartz wrote that he had originally filed Mr. Mata’s lawsuit in state court, but after the airline had it transferred to Manhattan’s federal court, where Mr. Schwartz is not admitted to practice, one of his colleagues, Mr. LoDuca, became the attorney of record. Mr. Schwartz said he had continued to do the legal research, in which Mr. LoDuca had no role.

Mr. Schwartz said that he had consulted ChatGPT “to supplement” his own work and that, “in consultation” with it, found and cited the half-dozen nonexistent cases. He said ChatGPT had provided reassurances.

“Is varghese a real case,” he typed, according to a copy of the exchange that he submitted to the judge.

“Yes,” the chatbot replied, offering a citation and adding that it “is a real case.”

Mr. Schwartz dug deeper.

“What is your source,” he wrote, according to the filing.

“I apologize for the confusion earlier,” ChatGPT responded, offering a legal citation.
“Are the other cases you provided fake,” Mr. Schwartz asked.

ChatGPT responded, “No, the other cases I provided are real and can be found in reputable legal databases.”

But, alas, they could not be.

Sheelagh McNeil contributed research.

Benjamin Weiser is a reporter covering the Manhattan federal courts. He has long covered criminal justice, both as a beat and investigative reporter. Before joining The Times in 1997, he worked at The Washington Post. @BenWeiserNYT
 

Jimmy from Linkedin

All Star
Supporter
Joined
Oct 10, 2017
Messages
2,353
Reputation
1,576
Daps
7,074
Reppin
Managing Director at Breh & Breh, Inc
Really hallucinates how? So in other words I'd have to verify and read through them myself?
ChatGPT is basically a very advanced pattern matcher. I asked it once for some books by some black psychoanalysts. It gave me five authors and book titles. Two of those authors existed, one was a white guy who was not a psychoanalyst and the other book was black skin white masks. When I really looked at it, the pattern it matched was derivatives of black skin white masks, eg, The Child In The Storm: Perceptions of Black Male Agression in White America.

When I read the citations I was ready to clean my wallet out and really understood. You can ask ChatGPT to help you format stuff, help you synthesize an understanding of the material, but it cannot find stuff for you. Thats all you. Copy and paste the articles and the sections and use different plugins (ChatGPT+) so that you can "Chat with the PDF", and that is how you can research with ChatGPT.

But if you want the understanding you entered grad school to get. Still gotta wear down that proverbial shoe leather my g :ufdup:
 
Top