If the live version is already broken, there isn’t much to lose deploying the fix as soon as possible. Not sure what else they could have done here.
If the live version is already broken, there isn’t much to lose deploying the fix as soon as possible. Not sure what else they could have done here.
Of course, but I still find it remarkable that the task that was picked as an example for something extremely difficult is now trivially easy just a few years later
The example given in the comic has moved from one category to the other. Determining whether an image contains a bird is a fairly simple “two hour” task now.
Plot twist: The woman in the comic is Fei-Fei Li, she got the research team and five years and succeeded 🤯
The article just says that the account is suspended, there is no official statement from Twitter an no indication that they suspended the account on purpose. The most likely reason is that the account was mass reported by trolls and got suspended automatically.
The Github UX is amazing if you ever had to use gitlab or bitbucket
Yeah, 3D printers are everywhere. Both as a business and as a hobby, it’s bigger than it has ever been.
It’s not a different discipline, an LLM is an example of a machine learning model.
I think it’s reasonable to not short stocks. I just find it a bit weird to see people confidently proclaim that a company is overvalued, but than not shorting the stock, which would be the rational thing to do.
It’s hard to tell how much a platform is worth, arguably the value of Twitter was 44B, since someone was willing to pay that.
The good news is, if you’re really certain that Reddit is overvalued, you’ll soon be able to short it and get rich if you end up being right!
I don’t think the number of bots matters much, there are much more real people on Twitter than on Mastodon. It’s not an issue for Twitter because they already are the platform where everyone else is. I’m optimistic about Mastodon, it already has the better UX and the better business model and I think it will slowly attract more users over time and eventually reach the relevance that Twitter had at its peak.
It works as long as you don’t call list()
within that function.
The difficult thing is gaining users, not writing the code.
I’ve been on Mastodon for over a year and I never experienced anything that could be classified as a technical glitch. From a tech / UI perspective it feels very polished to me.
I guess the only exception would be that old posts are sometimes missing on profiles from different servers.
Haha, I completely missed that it’s a game.
I wonder how often someone walks in and tells them about the mistake. Do the baristas have a standard response?
GitKraken!
I’m not an electron hater, but a terminal in electron sounds like a parody.
This article is full of errors!
At its core, an LLM is a big (“large”) list of phrases and sentences
Definitely not! An LLM is the combination of an architecture and its model parameters. It’s just a bunch of numbers, no list of sentences, no database. (Seems like the author confused the word “LLM” with the dataset of the LLM???)
an LLM is a storage space (“database”) containing as many sample documents as possible
Nope. This applies to the dataset, not the model. I guess you can argue that memorization happens sometimes, so it might have some features of a database. But it isn’t one.
Additional data (like the topic, mood, tone, source, or any number of other ways to categorize the documents) can be provided
LLMs are trained in an unsupervised fashion. Just sequences of tokens, no labels.
Typically, an LLM will cover a single context, e.g. only social media
I’m not aware of any LLM that does this. What’s the “context” of GPT-4?
software developers have gone to great lengths to collect an unfathomable number of sample texts and meticulously categorize those samples in as many ways as possible
The closest real thing is the RLHF process that is used to fine tune an existing LLM for a specific application (like ChatGPT). The dataset for the LLM is not annotated or categorized in any way.
a GPT uses the words and proximity data stored in LLMs
This is confusing. “GPT” is the architecture of the LLM.
it is impossible for it to create something never seen before
This isn’t accurate, depending on the temperature setting, an LLM can output literally any word at any time with a non-zero probability. It can absolutely produce things it hasn’t seen.
Also I think it’s too simple to just assert that LLMs are not intelligent. It mostly depends on your definition of intelligence and there are lots of philosophical discussions to be had (see also the AI effect).
This looks like an embarrassing mistake. If someone were to try to “tank” Twitter, it wouldn’t really make sense to do this on purpose.