Hacker Newsnew | past | comments | ask | show | jobs | submit | shannifin's commentslogin

I've always thought the basilar membrane was a fascinating piece of biological engineering. Whether or not the difference between its behavior vs FT really matters depends on the context. Audio processing on a computer, FFT is often great. Trying to understand / model human sound perception, particularly in relation to time, FFT has weaknesses.


Learned about the tonnetz (among other models) from the book "Audacious Euphony"... Challenging (and unfortunately too expensive now), but fascinating stuff...

https://www.amazon.com/dp/019977269X


Others have mentioned Branta Games, but I first saw the effect here: https://youtu.be/TdTMeNXCnTs


This one is actually more sophisticated because it doesn't rely on scrolling pixels like the OP. So the object doesn't just disappear in screenshots, but also when the animation stops moving! So you can't actually display text that stands still, like the "hello" in the OP.


Yep. He tries text in another video by flipping pixels for one or more frames, so the words disappear very quickly. Definitely harder to read, especially longer words: https://youtu.be/EDQeArrqRZ4


I'm not sure I follow. Couldn't you display text that stands still by (re)drawing the outline of the text repeatedly? It would essentially be a two frame animation


I think the algorithm in the video is doing a very specific thing where there's a zero-width pixel-grid-clamped stroke (picture an etch-a-sketch-like seam carving "between" the bounds of pixels on the grid) moving about the grid, altering (with XOR?) anything it advances across.

So, sure, you could try to implement this by having a seam that is made to "reverberate" back and forth "across" the outlining pixels of a static shape on each frame. But that's not exactly the same thing as selecting the outline of the shape itself and having those pixels update each frame. Given the way this algorithm looks to work, pushing the seam "inwards" vs "outwards" across the same set of pixels forming the outline might gather an entirely different subset of pixels, creating a lot of holes or perhaps double-counting pixels.

And if you fix those problems, then you're not really using this algorithm any more; you're just doing the much-more-boring thing of taking a list of pixel positions forming the outline and updating them each frame. :)


I believe the algorithm in the video works by flipping the pixel color when the pixel changes from foreground (some shape) to background, or from background to foreground. If the shape doesn't move, there is no such change, so it disappears.

In the OP the foreground pixels continuously change (scrolling in this case) while the background doesn't change. That's a different method of separating background and foreground.


thanks, that's also the best explained one!


Problem is, even with symbolic logic, reasoning is not completely deterministic. Whether one can get to a set of given axioms from a given proposition is sometimes undecidable.


I don't think this is really a problem. The general problem of finding a proof from some axioms to some formula is undecidable (in e.g. first order logic). But that doesn't tell you anything about specific cases, in the same way that we can easily tell whether some specific program halts, like this one:

"return 1"


True, I was rather pointing out that being able to parse symbolic language deterministically doesn't imply that we could then "reason" deterministically in general; the reasoning would still need to involve some level of stochasticism. Whether or not that's a problem in practice depends on specifics.


Hmmm... Nothing made me cry. I will diminish and go into the west.


Godspeed


I had a similar reaction, probably because I don't care about dogs.


Ha! Fun stuff. While some ideas are actually intriguing, many if its suggestions seem to be overly vague jumbles of common phrases and technology. "AI-powered databases to leverage personalized accessibility for team management..." Lol. Still fun though.


No worse than a lot of actual companies. I would be totally unsurprised to see that description about a real startup.


Some little audio examples would also be nice so visitors don't have to scroll through the video to hear them.

Still, awesome work!


Yeah, all the cool 3D pictures should play demo videos on click!


Christmas 1996, I was 11. We finally got a modern computer with Windows 95, a CD-ROM drive, speakers, Oregon Trail 2... It was magical! The only time I ever wept with joy over a Christmas present.


I don't have much real use for celebrity voices (other than fun experimentation), but I'd love to be able to clone my own voice and character voices for the purposes of creating audiobooks / audioplays without having to pay monthly fees with monthly usage limits. So I'm excited by this sort of project!

P.S. Are there any tools for synthetic voice creation? Maybe melding two or more voices together, or just exploring latent space? Would be fun for character creation to create completely new voices.


I'd be interested as well. This is where I imagine the space is going - particularly as the potential for litigation increases around cloning.

Game studios will spin up a bunch of unique virtual voices for all the dialogue of extras. It'll probably be longer before we see replacements of main characters though. There's been some research in speech-to-speech transference as well - this means that company employee A records the character B's line with the appropriate emotional nuance (angry, sad, etc.) and the emotional aspect is copied on top of the generated TTS.


Have you tried eleven labs? I used that. Had to record 3 hours of training audio reading books and and news articles. But the result was really good.


They're great! They just cost too much for how much output I want.


How much did the training cost?


I’ve used tortoise tts before and trained it on my voice and a mix of voices. It’s not perfect but still impressive.


StyleTTSv2 is pretty good and open source, you can easily traverse its latent space for voice


Similarly, I’m not excited by “voice cloning” at all, but I’d like to have very high quality, natural sounding TTS. All of the projects that do that seem to be geared towards also allowing arbitrary voice cloning based on short audio clips I’ve noticed.


Was thinking the same thing. Very fun idea, nice work!


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: