Notices by lain (lain@fediffusion.art)
-
Embed this notice
lain (lain@fediffusion.art)'s status on Wednesday, 04-Sep-2024 00:42:30 JST lain cc @hakui -
Embed this notice
lain (lain@fediffusion.art)'s status on Tuesday, 03-Sep-2024 15:13:43 JST lain @thendrix video games! -
Embed this notice
lain (lain@fediffusion.art)'s status on Saturday, 31-Aug-2024 18:36:30 JST lain > I'm shutting down http://therapistai.com/, my most controversial product ever I think
horrible domain name -
Embed this notice
lain (lain@fediffusion.art)'s status on Wednesday, 28-Aug-2024 01:35:24 JST lain flux dev really is the new king -
Embed this notice
lain (lain@fediffusion.art)'s status on Saturday, 10-Aug-2024 06:38:06 JST lain > run python main.py
> downloads 50gb of models
> crashes
> refuses to elaborate
i love python-based machine learning!! -
Embed this notice
lain (lain@fediffusion.art)'s status on Friday, 09-Aug-2024 04:49:45 JST lain prompting kaiabait with flux -
Embed this notice
lain (lain@fediffusion.art)'s status on Tuesday, 06-Aug-2024 20:00:35 JST lain Someone is selling a pleroma tan generation service
https://aieasypic.com/inspire/models/detail/fox-tan-pleroma-mascot-v10-42504 -
Embed this notice
lain (lain@fediffusion.art)'s status on Monday, 05-Aug-2024 06:01:09 JST lain > You might think--why would someone write an entire article justifying that language models are useful??! Isn't that obvious?!? But there seem to be a (large?) contingent of people out there---in the academic literature, in the software engineering space, and also in the media sphere---who proclaim widely that LLMs contribute nothing, are just another hype cycle, and in a few years will die having had no impact on the world. I will be arguing these people are wrong because current LLMs are already useful.
https://nicholas.carlini.com/writing/2024/how-i-use-ai.htmlIn conversation from fediffusion.art permalink Attachments
-
Embed this notice
lain (lain@fediffusion.art)'s status on Friday, 02-Aug-2024 02:20:30 JST lain new open weight image generation model: https://blackforestlabs.ai/ In conversation from fediffusion.art permalink Attachments
-
Embed this notice
lain (lain@fediffusion.art)'s status on Tuesday, 30-Jul-2024 22:12:59 JST lain are you thinking what i'm thinking? In conversation from fediffusion.art permalink Attachments
-
Embed this notice
lain (lain@fediffusion.art)'s status on Tuesday, 30-Jul-2024 04:13:57 JST lain @meso so you don't feel inadequate In conversation from fediffusion.art permalink -
Embed this notice
lain (lain@fediffusion.art)'s status on Monday, 29-Jul-2024 19:58:47 JST lain @kaia she's over 30 by now In conversation from gnusocial.jp permalink -
Embed this notice
lain (lain@fediffusion.art)'s status on Monday, 29-Jul-2024 19:57:44 JST lain i'm thinking miku In conversation from fediffusion.art permalink Attachments
-
Embed this notice
lain (lain@fediffusion.art)'s status on Saturday, 27-Jul-2024 04:59:49 JST lain new paper in nature that shows that the way that all current models are successfully using to generate training data "doesn't work". good that it didn't come out earlier. In conversation from fediffusion.art permalink Attachments
-
Embed this notice
lain (lain@fediffusion.art)'s status on Thursday, 25-Jul-2024 02:24:06 JST lain le based Mistral want to one-up Meta and release an open 120b model that also reaches gpt4 quality, at least in certain tasks. It's also highly multilingual, I think this model would be the perfect japanese learning partner. Non-commercial license, sadly.
https://mistral.ai/news/mistral-large-2407/In conversation from fediffusion.art permalink Attachments
-
Embed this notice
lain (lain@fediffusion.art)'s status on Wednesday, 24-Jul-2024 00:38:30 JST lain Meta releases Llama 3.1, an 8b, 70b and 405b model, all of them with a large context size of 128k.
The 405b model beats chatgpt 4o on nearly every benchmark.
https://ai.meta.com/blog/meta-llama-3-1/In conversation from fediffusion.art permalink Attachments
-
Embed this notice
lain (lain@fediffusion.art)'s status on Friday, 19-Jul-2024 01:28:38 JST lain https://mistral.ai/news/mistral-nemo/
> Today, we are excited to release Mistral NeMo, a 12B model built in collaboration with NVIDIA. Mistral NeMo offers a large context window of up to 128k tokens. Its reasoning, world knowledge, and coding accuracy are state-of-the-art in its size category. As it relies on standard architecture, Mistral NeMo is easy to use and a drop-in replacement in any system using Mistral 7B.In conversation from fediffusion.art permalink Attachments
-
Embed this notice
lain (lain@fediffusion.art)'s status on Friday, 19-Jul-2024 01:28:27 JST lain 128k plus good multilanguage ability is insane. that's like a hundred pages of text. You can put a whole short story in it to translate. In conversation from fediffusion.art permalink -
Embed this notice
lain (lain@fediffusion.art)'s status on Wednesday, 17-Jul-2024 01:17:46 JST lain @kaia and redpilled In conversation from gnusocial.jp permalink -
Embed this notice
lain (lain@fediffusion.art)'s status on Wednesday, 17-Jul-2024 01:08:09 JST lain probable future US vice president In conversation from fediffusion.art permalink Attachments