OpenAI GPT-3 fine-tuned complete API generates non-english content with longer response lengths [closed] - openai-api

Closed. This question is not about programming or software development. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 4 days ago.
Improve this question
I have been working with a fine tuned OpenAI GPT-3 model based on the curie model for a project that generates academic type essays for a user given topic. It was trained on academic papers scraped from a website. It works well on short response lengths such as 300 words or less, but when the users select longer lengths it begins to generate nonsense characters towards the end of the output.
Example:
Input parameters
Response starts out normal for the prompt Canada
And then at the end of the response it gives gibberish
I've found a couple issues with the training method already and fixed some of those errors, but am having trouble brainstorming possible causes of these repeating gibberish characters. Does anyone have any ideas or possible causes of this kind of output? Has anyone run into this before? Any help or advice is greatly appreciated.
Thank you!

Related

How can I filter certain words in senders's address with gmail? [closed]

Closed. This question is not about programming or software development. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 6 days ago.
Improve this question
Recently I have been getting a lot of spam from a shopping service I have an account on. The emails that they normally send are from "buyer-info13.g#mail.[removed].com" or "notice-buyer04.g#mail.[removed].com"
I can easily mark the sender's address as spam or automatically removed, but the 2 numbers at the end (13, or 04) change.
I was looking at search operators you can use with gmail but for some reason they do not work when I try from:("notice-buyer" OR "buyer-info"):
Is there any way I can make this filter correctly? If this does not work does regex filters work? I can easily do this with Outlook but then I would require Outlook to be running 24/7 to filter emails. Thank you.
The gmail search operators for multiple terms require brackets, not parens. Example: {from:amy from:david}.
In your case, I'd probably enter the full email address, including the domain.
Reference: https://support.google.com/mail/answer/7190?hl=en

Stop Excel Scientific Notation Conversion [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 3 years ago.
Improve this question
Is there a way to STOP EXCEL from converting data into Scientific Notation?
Microsoft Excel apparently assumes that any data that contains numbers and an E is supposed to be scientific notation and converts it.
17195E000039 automatically changes to 1.7195E+43
17221E000112 automatically changes to 1.7221E+116
Is there a way to prevent Excel from doing this so it doesn't continue to cause problems for all of us users who never want to use SCIENTIFIC NOTATION ever? I cannot foresee any time where I will EVER want to have anything converted to scientific notation. This useless functionality only causes problems.
How do I stop Excel from trying to convert text into scientific without manually performing some operation first?
Unfortunately, the data comes from another system that exports to excel that I cannot change. I cannot manually change the page beforehand and going through afterwards sucks.
Issue also occurs with simple strings (eg "1E6" from microtitre plates). I don't think it's a simple as pnuts or Scott's responses.
See similar Q I just posted with more detail, toy reprex example, and trial solutions that have failed. How to stop csv files converting characters (eg plate well IDs 2E6) to scientific notation?
Is there a way to STOP EXCEL from converting data into Scientific Notation?
As explained by #Scott Craner: For you, No.
This is something that happens automatically when the data is written - by the time you see it it is too late.

Google search operators to find out of stock Amazon products [closed]

Closed. This question is not about programming or software development. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 8 days ago.
Improve this question
I need to identify "Out of Stock" Amazon product listings for acrylic paints.
I am using the Google search operator:
site:amazon.com -inurl:feature.html acrylic paint "We don't know when or if this item will be back in stock."
I'm using "-inurl:feature.html" to remove best seller lists.
And "We don't know when or if this item will be back in stock." to display only out of stock items.
But the results are still too general. Does anyone know how I can only show results from the "Arts, Crafts & Sewing" category? I tried including that into the search string, but since that phrase appears on all pages it did not work.
Or maybe someone has a better way to do this instead of using Google?
just add "Arts, Crafts & Sewing" to the query :) it works for me.
like this:
site:amazon.com -inurl:feature.html acrylic paint "We don't know when or if this item will be back in stock." "Arts, Crafts & Sewing"

Wireshark network layer name resolution pcap [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 3 years ago.
Improve this question
I'm interested in saving a pcap that has network layer name resolution. While it works great within Wireshark, how can I save it with the resolved names intact? Having this information would be extremely helpful for me and save me a lot of time if this is possible. I understand in the documentation that it can't be saved within the pcap file (http://www.wireshark.org/docs/wsug_html_chunked/ChAdvNameResolutionSection.html#idp390072124) but is there an alternative way to do so? Does anyone have any solutions to this?
Thanks in advance!
I haven't tried it myself, but in theory the name resolution information can/will be stored in the pcap-ng file format, which has been Wireshark's default file format since version 1.8. The old pcap file format you cite won't, but pcap-ng has a specific defined block type in its format for ip<->name resolution information.

Reading Composite Document File V2 Document (.msg) files in ubuntu [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 9 years ago.
Improve this question
I have a large dump of data from an outlook email account that comes entirely in .msg files. A quick call to ubuntu's file method revealed that they were Composite Document File V2 Documents (whatever that means). I would really like to be able to read these files as plaintext. Is that possible at all?
Update: Turns out it wasn't totally possible to do what I wanted for large scale data mining on these kinds of files which was a bummer. In case you face the same issue I made a library to address this issue. https://github.com/Slater-Victoroff/msgReader
Documentation isn't great, but it's a pretty small library so it should be self explanatory.
I faced the same problem this morning. I didn't find any information on the file format but it was possible to extract the required information from the file using strings and grep:
strings -e l *.msg | grep pattern
The -e l (that's a small L) converts from UTF-16.
This will only work if you can grep the data you need from the file (i.e. all required lines contain a standard string or pattern).

Resources