@HistoPol@annaleen the original ChatGPT turned out to be a lot less prone to wild vengeful outbursts than whatever model it was that they plugged into Bing - it's a pretty interesting demo of how well the safety features of ChatGPT (which Bing seems not to have) have held up
Open source maintainers are often good at writing code but not good at asking for money
Companies aren't very good at giving money away, but they absolutely know how to hire consultants - and they often have a training budget set aside already
My closing suggestion was to support open source projects you depend on by reaching out to their maintainers and offering to pay them money to speak to your team! Ask them for an hour over Zoom or similar.
OK, I have a somewhat baffling (to me) Mastodon question. How do I link to a thread?
I want to link to a fantastic thread by @mhoye - but if I link to the first post in that thread - https://mastodon.social/@mhoye/111335603309582734 - I get a page with a single post on it and other people's replies, with no indication it's part of a larger thread from the same author
"We already know one major effect of AI on the skills distribution: AI acts as a skills leveler for a huge range of professional work. If you were in the bottom half of the skill distribution for writing, idea generation, analyses, or any of a number of other professional tasks, you will likely find that, with the help of AI, you have become quite good."
For me this is the single most positive potential future outcome of the current leap forward in terms of AI capabilities: tools that helps human beings level up, as opposed to replacing them
A big question here (as Ethan outlines further down in that post) is whether organizations will chose to do more with their existing teams, or will downsize them to try to achieve the same with less people
Early hints we're getting from stock photography, translation & illustration don't look great so far
I'm particularly interested in translation (between human languages) because that's been quietly impacted by leaps forward automation for 5-10 years at this point
I'd love to see some credible studies on what the impact of that has been on the human translator industry: has there been mass unemployment, or are existing translators getting more work but it's more "clean up this AI-generated version", or is it a mixture of both?
Here's a fun example of something you can now do with LLM: search for every README.md file in your home directory and store embeddings for all of them in a collection called "readmes": ``` llm embed readmes \ --model sentence-transformers/all-MiniLM-L6-v2 \ --files ~/ '**/README.md' ``` Then run a similarity search for "sqlite" like this: ``` llm similar readmes -c sqlite ```
Leaked Google document: “We Have No Moat, And Neither Does OpenAI”
The most interesting thing I've read recently about LLMs - a purportedly leaked document from a researcher at Google talking about the huge strategic impact open source models are having https://simonwillison.net/2023/May/4/no-moat/
@jimgar the ethics of this stuff is incredibly complicated
I'm very optimistic about the models being trained on the RedPajama data - there's one out already and evidently more to follow very shortly https://simonwillison.net/tags/redpajama/
@resing@jimgar I'm not convinced it's possible to train a usable LLM without including copyrighted material in they raw pretraining data
As such, personally think it's a necessary evil to avoid a monopoly on LLM technology belonging to organizations that are willing to train against crawler data
Open source developer building tools to help journalists, archivists, librarians and others analyze, explore and publish their data. https://datasette.io and many other #projects.