As Robin Hanson says, building the sheer variety of products we have is actually bad, because it increases unit costs. This is especially clear in laptops - there are far too many laptops with too little to distinguish them and too many nonsense minor issues. As such, I think we need a new streamlined and harmonized lineup of all laptops:
Cheapest Possible Technically Functional Laptop
Mediocre Office and Home Laptop (to be issued to most office workers and people who want to edit spreadsheets or emails and such)
CEO Laptop (reasonably fast, expensive, big battery for CEO activities)
Programmer Laptop (ThinkPad-like focused on CPU performance and reasonable portability)
Gamer Laptop (16" Legion-like with middling battery life and decently high-powered CPU/GPU)
Gamer Laptop (Big) (17"-18" desktop replacement)
Technician Laptop (smallish thick and rugged laptop with many ports)
Multimedia Laptop (Mediocre Office and Home Laptop with a nicer display and better graphics)
There would also be a version number updated whenever new components are available, of course. There can perhaps be two or three variants of each (with the same chassis, board, etc but different components) with different pricing, but no more.
Why do all three of the reasonably okay AI music tools (Udio, Suno, Riffusion) have fairly similar artifacts? Except for, I think, older versions of Udio, they all sound consistently off in some way I don't know enough music theory to explain, particularly in metal vocals and/or complex instrumentals. Do they all use the same autoencoders or something?
Street-Fighting Mathematics is not actually related to street fighting, but you should read it if you like estimating things. There is much power in being approximately right very fast, and it contains many clever tricks which are not immediately obvious but are very powerful. My favourite part so far is this exercise - you can uniquely (up to a dimensionless constant) identify this formula just from some ideas about what it should contain and a small linear algebra problem!
People are claiming (I don't know much RL) that DeepSeek-R1's training process is very simple (based on the paper: https://github.com/deepseek-ai/DeepSeek-R1/blob/main/DeepSeek_R1.pdf) - a boring standardish (for LLMs) RL algorithm optimizing for reward on some ground-truth-verifiable tasks (they don't say which). So why did o1 not happen until late 2024 (public release) or late 2023 (rumours of Q*)? "Do RL on useful tasks" is a very obvious idea. I think the relevant algorithms are older than that.
The paper says that they tried applying it to smaller models and it didn't work nearly as well, so "base models were bad then" is a plausible explanation, but it's clearly not true - GPT-4-base is probably a generally better (if costlier) model than 4o, which o1 is based on (could be distillation from a secret bigger one though); and LLaMA-3.1-405B used a somewhat similar postttraining process and is about as good a base model, but is not competitive with o1 or R1. So I don't think it's that.
What's going on here? The process is simple-sounding but filled with pitfalls DeepSeek don't mention? What has changed between 2022/23 and now which means we have at least three decent long-CoT reasoning models around?
there is a very large quantity of widely dispersed gods and you don't know about the vast majority of them
there are quite a few gods, but a bounded amount
there is exactly one god
there are exactly zero gods
By extrapolation, we can conclude that the next step is that humanity has negative one god, i.e. is in theological debt and must build a god to continue. This is where the EY-style "aligned singleton" came from. But people are now moving toward "we need everyone to have pocket gods" because they are insane, in line with the pattern. The next step is of course "we need to build gods and put them in everything".
Experiments
Various web projects I have put together over many years. Made with at least four different JS frameworks. Some of them are bad.
Collect Arbitrary Points and achievements by doing things on this website! See how many you have! Do nothing with them because you can't! This is the final form of gamification.
It is pitch black (if you ignore all of the lighting). You are likely to be eaten by Heavpoot's terrible writing skills, and/or lacerated/shot/[REDACTED]. Vaguely inspired by the SCP Foundation.
A Reverse Polish Notation (check wikipedia) calculator, version 2. Buggy and kind of unreliable. This updated version implements advanced features such as subtraction.
Your favourite* tic-tac-toe game in 3 dimensions, transplanted onto the main website via a slightly horrifically manual process! Technically this game is solved and always leads to player 1 winning with optimal play, but the AI is not good enough to do that without more compute!
Type websocket URLs in the top bar and hit enter; type messages in the bottom bar, and also hit enter. Probably useful for some weirdly designed websocket services.
In our Lenovo ThinkCentre M75q Tiny Gen5 review, we see how this 1L-class PC offers an improved AMD Ryzen SoC and we got 128GB of RAM working The post Lenovo ThinkCentre M75q Tiny Gen5 Review An AMD Ryzen Powered TinyMiniMicro appeared first on...
To celebrate the launch of the upcoming Mage Errant Illustrated Omnibus Edition Kickstarter, the first three books in the series are free until March 5th! Source
Just been making some adjustments to the Imperial Military Service Table of Ranks today (specifically, adjusting the enlisted ranks to correct some seniorities and fix the missing E-9 grade), so here's the current version, screenshotted from my...
Over on GitHub, Alejandro Martín has recently released his open-source 'rtl-sdr-analyzer' software, which is an RTL-SDR-based signal analyzer and automatic jamming detector. The software is based on Python and connects to the RTL-SDR via an rtl_tcp...