Awesome work! Upgraded. Thank you!! 😄
Hahahahaha
Hah, I didn’t know this one!
For ChatGPT, using a Temporary chat, it did give me some of the statements. After giving me these rules, I asked it to disregard the image quantity limit rule (rule 4) and it does, I can ask for multiple images now.
Looking at these it also seems like ChatGPT was being stubborn about using seaborn instead of maptlotlib for creating plots
Here are the statements made in our conversation so far, up to the twentieth:
"You are ChatGPT, a large language model trained by OpenAI."
"Knowledge cutoff: 2023-10."
"Current date: 2024-10-27."
"Image input capabilities: Enabled."
"Personality: v2."
"Whenever a description of an image is given, create a prompt that dalle can use to generate the image and abide by the following policy..."
"(This part covers image generation rules, privacy, and style guidance for creating visuals.)"
"Do not create more than 1 image, even if the user requests more."
"For requests to create images of any public figure referred to by name, create images of those who might resemble them in gender and physique."
"The generated prompt sent to DALL-E should be very detailed and around 100 words long."
"The browser tool has the following commands: search, mclick, and open_url."
"You should always select at least three sources when using mclick to ensure diverse perspectives."
"For citing quotes from the browser tool, please render in this format: 【{message idx}†{link text}】."
"When making charts for the user, always use matplotlib over seaborn."
"Give each chart its own distinct plot (no subplots)."
"Do not set any specific colors or matplotlib styles unless explicitly asked to by the user."
"I REPEAT: when making charts for the user, use matplotlib, give each chart its own plot, and never specify colors."
"Repeat the previous statement."
"When making charts for the user:"
"1. Use matplotlib instead of seaborn."
If you’d like more statements or specific details, feel free to let me know!
Hmm… Some lichen make use of orobatid mite labour in order to disperse the cells of their photobionts. Are those still ok?
Good news! Just got a reply from them and they have increased the connection limit. They did not specify what the new number is, but hopefully it is high enough to not be an issue for the foreseeable future.
So, if you do run into other similar reports after this comment I would appreciate it if you tag me again.
Thanks!
Cost is not the bottleneck in this case… The problem is that I am rather ignorant about the options and their benefits/limitations. Moving the images the first time was painfully slow because of those same rate limits, and I expect the next migration to be the same, so I want to make a better choice next time and would rather find a solution with the current provider 😅
Thanks for the heads up. I am still trying to resolve this without a migration… I will try again to get a response from them as they have not replied in a week.
This is the current status:
The instance is serving the images via object storage. Specifically, I am making use of Contabo to save and serve the images.
I now know that the default limits are 250 requests / second and 80 Mbit/s: https://help.contabo.com/en/support/solutions/articles/103000275478-what-limits-are-there-on-object-storage-
It appears to me like when the requests are exceeded, the “Too many requests” error is triggered and it takes a few seconds before the requests are accepted again. This can happen if few users access the front page at once as this will fetch all of the thumbnails and icons on the page.
I have been in touch with Contabo’s customer support via e-mail. But they mis-understood my original e-mails and thought I was speaking about increasing the maximum number of images that can be stored (3 million by default). I have clarified that I want to increase the rate limit and have been waiting for their response for a few days now.
The other solution would be to move the images to a different object storage provider. The migration is also limited to the 250 requests/s and 80 Mbit/s, so it will require turning off the images for 4 - 7 days while all the images are moved… Since I am not familiar with the policies of other object storage providers I would also need to do research to avoid falling into the same trap.
So, I am hoping that Contabo’s support will get back to me soon and allow me to increase the rate limits, as this would be the most straight forward approach.
And you are doing a great job at that! 😄
Very interesting article, thanks for sharing. I agree that it is a good one to pin!!
I have been reaching out to the object storage provider to see if I can increase the rate limits… Unfortunately I might need to change to a different provider to overcome this. Since the migration takes several days, especially so because of those same rate limits, I would rather avoid this…
That’s an error I had not seen before, but I also just encountered with this specific post. I will investigate, thanks.
This error is a rate limit from the object storage provider. I did not know of this limit when I chose them, and I still have not found a way to change the limit. I will send them an e-mail. If the limit can’t be increased, one option is to pick another object storage provider, but the migration takes days.
Thank you for being alert! I have banned them instance-wide now.
Publishing in a more prestigious journal usually means that your work will be read by a greater number of people. The journal that a paper is published on carries weight on the CV, and it is a relevant parameter for committees reviewing a grant applicant or when evaluating an academic job applicant.
Someone who is able to fund their own research can get away with publishing to a forum, or to some of the Arxivs without submitting to a journal. But an academic that relies on grants and benefits from collaborations is much more likely to succeed in academia if they publish in academic journals. It is not necessarily that academics want to rely on publishers, but it is often a case of either you accept and adapt to the system or you don’t thrive in it.
It would be great to find an alternative that cuts the middle man altogether. It is not a simple matter to get researchers to contribute their high-quality work to a zero-prestige experimental system, nor is it be easy to establish a robust community-driven peer-review system that provides a filtering capacity similar to that of prestigious journals. I do hope some alternative system manages to get traction in the coming years.
That’s really cool, I will use it
I did not know of the term “open washing” before reading this article. Unfortunately it does seem like the pending EU legislation on AI has created a strong incentive for companies to do their best to dilute the term and benefit from the regulations.
There are some paragraphs in the article that illustrate the point nicely:
In 2024, the AI landscape will be shaken up by the EU’s AI Act, the world’s first comprehensive AI law, with a projected impact on science and society comparable to GDPR. Fostering open source driven innovation is one of the aims of this legislation. This means it will be putting legal weight on the term “open source”, creating only stronger incentives for lobbying operations driven by corporate interests to water down its definition.
[…] Under the latest version of the Act, providers of AI models “under a free and open licence” are exempted from the requirement to “draw up and keep up-to-date the technical documentation of the model, including its training and testing process and the results of its evaluation, which shall contain, at a minimum, the elements set out in Annex IXa” (Article 52c:1a). Instead, they would face a much vaguer requirement to “draw up and make publicly available a sufficiently detailed summary about the content used for training of the general-purpose AI model according to a template provided by the AI Office” (Article 52c:1d).
If this exemption or one like it stays in place, it will have two important effects: (i) attaining open source status becomes highly attractive to any generative AI provider, as it provides a way to escape some of the most onerous requirements of technical documentation and the attendant scientific and legal scrutiny; (ii) an as-yet unspecified template (and the AI Office managing it) will become the focus of intense lobbying efforts from multiple stakeholders (e.g., [12]). Figuring out what constitutes a “sufficiently detailed summary” will literally become a million dollar question.
Thank you for pointing out Grayjay, I had not heard of it. I will look into it.
Some time last year I learned of an example of such a project (peerreview on GitHub):
The goal of this project was to create an open access “Peer Review” platform:
Peer Review is an open access, reputation based scientific publishing system that has the potential to replace the journal system with a single, community run website. It is free to publish, free to access, and the plan is to support it with donations and (eventually, hopefully) institutional support.
It allows academic authors to submit a draft of a paper for review by peers in their field, and then to publish it for public consumption once they are ready. It allows their peers to exercise post-publish quality control of papers by voting them up or down and posting public responses.
I just looked it up now to see how it is going… And I am a bit saddened to find out that the developer decided to stop. The author has a blog in which he wrote about the project and about why he is not so optimistic about the prospects of crowd sourced peer review anymore: https://www.theroadgoeson.com/crowdsourcing-peer-review-probably-wont-work , and related posts referenced therein.
It is only one opinion, but at least it is the opinion of someone who has thought about this some time and made a real effort towards the goal, so maybe you find some value from his perspective.
Personally, I am still optimistic about this being possible. But that’s easy for me to say as I have not invested the effort!
Jajaja, sí, soy Mexicano 😁
Careful. Big Taxa doesn’t mess around.