holomorphic

joined 2 years ago
[–] holomorphic@lemmy.world 1 points 1 week ago

For me it takes 4-7km of running or 15-25km cycling to get to the state where it suddenly feels like I can just go on forever. Of course that only works until the food I carry runs out or muscles start to hurt.

[–] holomorphic@lemmy.world 6 points 2 weeks ago

Why would a company want to do that for their own internal use? Models you can download are mostly just data. They don't do anything on their own. You can even write your own interpreter for them, if you feel like it.

[–] holomorphic@lemmy.world 3 points 2 weeks ago

Doug Dimmadome, the owner of the dimmsdale dimmadome?

[–] holomorphic@lemmy.world 8 points 2 weeks ago

xcq link stays blue

[–] holomorphic@lemmy.world 3 points 4 weeks ago (1 children)

Switzerland? In the French or Italien speaking parts maybe. The German speaking part is more like Germany with less bureaucracy.

[–] holomorphic@lemmy.world 2 points 1 month ago

Sure, let's pretend the field of AI, with that exact name, isn't at least 50 years old.
Facial recognition has probably been taught in AI courses for almost two decades now. "Artificial Intelligence: A Modern Approach" is 30 years old, and only about half of it is concerned with Statistics.

[–] holomorphic@lemmy.world 7 points 1 month ago

Aaron was about as much a cofounder of reddit as Musk is a cofounder of tesla. Source: I was on reddit when he joined and I remember his own yc-startup Infogami.

[–] holomorphic@lemmy.world 2 points 3 months ago

Actually I agree. I guess I was just still annoyed after reading just previously about how llms are somehow not neural networks, and in fact not machine learning at all...

Btw, you can absolutely finetune llms on classical regression problems if you have the required data (and care more about prediction quality than statistical guarantees.) The resulting regressors are often quite good.

[–] holomorphic@lemmy.world 2 points 3 months ago

I will admit didn't check because it was late and the article failed to load. I just remember reading several papers 1-2years ago on things like cancer-cell segmentation where the 'classical' UNet architecture was beaten by either pure transformers, or unets with added attention gates on all horizontal connections.

[–] holomorphic@lemmy.world 31 points 3 months ago (2 children)

Ah, yes. The one place that hasn't been corrupted by capitalism

[–] holomorphic@lemmy.world 3 points 3 months ago* (last edited 3 months ago) (6 children)

Those models will almost certainly be essentially the same transformer architecture as any of the llms use; simply because they beat most other architectures in almost any field people have tried them. An llm is, after all, just classifier with an unusually large set of classes (all possible tokens) which gets applied repeatedly

 
 
view more: next ›