

Forget that the real evil is that the first character of Int or Ganz in this case is capitalized


Forget that the real evil is that the first character of Int or Ganz in this case is capitalized


Why does it matter what is replaced?
I would say if it frees up a human from having to do it that it’s good news.


“must benefit everyone, not just a handful of billionaires”
It’s hard to disagree with that statement but all this upheaval about “killing jobs” seems really pointless to me.
A manual farm has far more employees then one that uses modern machinery, is the economy in decline if the farmer replaces human labor with machines?


Good stuff, Ukraine has proven again that having a nuclear arsenal is really the only way to ensure a country’s security when faced wirh a hostile nuclear armed neighbor.


I’m selfhosting Forgejo and i don’t really see the benefit of migrating to a container, i can easily install and update it via the package manager so what benefit does containerization give?


If there is a way to combine this with something like anubis this could be very interesting, then you are also not dependent on the honesty of reading the robots.txt
Have a RSL license? here you go scrape the content.
Trying to freeload? anubis time.


Having something to talk to is a massive improvement over bottling it all up.
AI is very beneficial to people who can’t afford the cost or are otherwise unable/unwilling to speak with a professional.


How is this not a “won’t somebody please think of the children” thing?
Yes it is terrible this has happend, but there is a way to prevent children from accessing AI and its called parenting.
Kids shouldn’t be using AI if it harms them, kids can’t make this choice so it should be made for them. Same with alcohol, same with porn, same with the other things restricted to children.
But that doesn’t mean responsible adults shouldn’t be able to use it, but “won’t somebody please think of the children” litigation will make that impossible.


I don’t care if openai lose all their money but this ruling would also effect open source AI.
If somebody releases a AI why would they be liable for how people decide to use it? Its software and like any other program its the user’s choice on how to use it.
If i decide to run rm / --no-preserve-root is gnu then responsible to fix it?
AI is already very censored, and if makers become liable for what people do with their AI they will become hyper censored and performance will go down the drain.


Hitler sues Mussolini


Why do core counts and memory type matter when the table includes memory bandwith and tflop16?
The H200 has HBM and alot of tensor cores which is reflected in its high stats in the table and the amd gpus don’t have cuda cores.
I know a major deterioration is to be expected but how major? Even in extreme cases with only 10% efficiency of the total power then its still competitive against the H200 since you can get way more for the price, even if you can only use 10% of that.


Thanks! Ill go check it out.


My target model is Qwen/Qwen3-235B-A22B-FP8. Ideally its maxium context lenght of 131K but i’m willing to compromise. I find it hard to give an concrete t/s awnser, let’s put it around 50. At max load probably around 8 concurrent users, but these situations will be rare enough that oprimizing for single user is probably more worth it.
My current setup is already: Xeon w7-3465X 128gb DDR5 2x 4090
It gets nice enough peformance loading 32B models completely in vram, but i am skeptical that a simillar system can run a 671B at higher speeds then a snails space, i currently run vLLM because it has higher peformance with tensor parrelism then lama.cpp but i shall check out ik_lama.cpp.


While I would still say it’s excessive to respond with “😑” i was too quick in waving these issues away.
Another commenter explained that residential power physically does not suppply enough to match high end gpus is why even for selfhosters they could be worth it.


Thanks, While I still would like to know thr peformance scaling of a cheap cluster this does awnser the question, pay way more for high end cards like the H200 for greater efficiency, or pay less and have to deal with these issues.


FP8 Tensor Core, the RTX pro 6000 datasheet keeps it vague with only mentioning AI TOPS, which they define as Effective FP4 TOPS with sparsity, and they didn’t even bother writing a datasheet for he 5090 only saying 3352 AI TOPS, which i suppose is fp4 then. the AMD datasheets only list fp16 and int8 matrix, whether int8 matrix is equal to fp8 i don’t know. So FP16 was the common denominator for all the cards i could find without comparing apples with oranges.

?


Well a scam for selfhosters, for datacenters it’s different ofcourse.
Im looking to upgrade to my first dedicated built server coming from only SBCs so I’m not sure how much of a concern heat will be, but space and power shouldn’t be an issue. (Within reason ofcourse)


As long as Russia is fighting China gets cheap oil
It’s almost like client side anti cheat doesn’t work and if proper server side anti cheat is made it wouldn’t matter what platform the client is on.