Hacker Newsnew | past | comments | ask | show | jobs | submit | aeturnum's commentslogin

It is interesting to consider that disability may enable much higher academic performance as long as people get the proper accommodations. After all, wouldn't it be interesting if people we think of as disabled can - under the correct conditions - be more productive than 'able' people. An individuals' capability is generally pretty circumstantial and I think we should be open to asking questions about how optimal our current social structure is for productivity and capacity going forward. We may need to imagine new ways of living and structuring work and society to reach even higher levels of productivity.

^^This! Our current societal and cultural structure is not adequately set up to deal with, for example climate change and subsequent mass migrations, widening wealth gap, increasing authoritarianism, centrlising of control and power and a myriad of other problems, not to mention any second and third order effects; let alone improve to higher levels of productivity.

This is a weird response to a weird article. The original article doesn't define its terms and, as Robby points out, that makes it hard to critique. If a language is only "serious" if it can scale infinitely for all use cases then sure Ruby isn't serious - most languages aren't.

That said - this response and the critique seem to basically agree. The critique can be summed up as "Ruby doesn't work forever" (and so it should never be used) and this is saying "Ruby doesn't work forever" (which is fine). I could almost understand this post as saying: 'Ruby isn't serious and that's not a problem for anyone who uses it.'

I will say that I found it funny that the original article attacked Ruby for being all the way down at "18th place" (This is inaccurate - it's 14th in 2024) on the SO dev survey - while talking up Scala which is 9 places further down on the survey[1].

[1] https://survey.stackoverflow.co/2024/technology#most-popular...


> "Ruby doesn't work forever"

Where does the response even address this?

All I know is that Ruby code I wrote 10ish years ago is still going strong, for example a whole compiler https://github.com/WebKit/WebKit/tree/main/Source/JavaScript...


Here's some places I noticed it:

> critics love the Twitter example. But look closer. Ruby carried them further than most companies will ever reach. They outgrew their shoes. That’s not an indictment… that’s success.

> I’ve never seen a team fail because they chose Ruby. I have seen them fail because they chose complexity. Because they chose indecision.

> GitHub held the world’s source code together for years using Ruby.

There are many examples of companies that used Ruby at one point very successfully but moved on from it once it no longer fit their situation. This isn't a critique of Ruby! But it is agreeing that Ruby can be outgrown and that, if you are looking to start with a language your usecase probably won't ever outgrow, Ruby might not be the best choice.


GitHub is an example of something that worked better with ruby than react...it got much worse.

It also may have had a better time with more recent versions of Ruby.

> Ruby code I wrote 10ish years ago is still going strong, for example a whole compiler <https://github.com/WebKit/WebKit/tree/main/Source/JavaScript...>

Neat. How is offlineasm used? (Without going into the details about the background of LLInt, that is—what I mean is, how is the compiler invoked?) Is it just the reference compiler, corresponding to some other machinery inside JSC?


That’s how the interpreter in JavaScriptCore gets compiled. The interpreter is written in a macro assembly dialect I invented and this is the compiler for it.

(I say compiled, not assembled, because it’s higher level than normal assembly. There’s an actual pipeline of transformations that happens. Plus a Turing complete macro language)


Thanks, that's helpful. (I mistook the compiler as being one that deals with JSC bytecode, either as input or output.)

It's also funny he wisecracks Java and then loves Scala for it's robustness (much of which it owes to Java).

Email has been updated many times in the last 20 years. All of the major sender authentication protocols (SPF, DKIM, DMARC) were created and deployed over the last 20 years. Email is also famously insecure and lacking a standard way of managing encryption - so the reason you never see updates is because the features signal is changing do not exist in email at all.

SPF, DKIM, DMARC are all about server reputation. They don't count as any sort of update to email and don't affect the protocol. These days regular non E2EE email is as secure as any other messaging protocol that relies on trusted servers. Since it is federated over multiple servers it is better than systems with just one server. You can choose who to trust and can even host it yourself.

Compare with Signal where there is only one allowed server entity and hardly anyone verifies identities making man in the middle attacks trivial.


Any reference to the trivial mitm attacks which signal has suffered?

This is mostly about the usability issues that make such attacks work so well on Signal:

https://www.ndss-symposium.org/wp-content/uploads/2018/03/09...

This adds some detail about how Signal can do MITM attacks:

https://sequoia-pgp.org/blog/2021/06/28/202106-hey-signal-gr...

Some of the details might of changed since publication. My current understanding is that Signal doesn't even bring up the idea of identity verification if a user has not previously done it. So if anything, things have gotten worse.


PGP/GNUPG has worked well for me for nearly three decades.

Because amazon doesn't have a web search service but they do have a product recommendation service? Even if they do pay OpenAI they would certainly be competing with their own service and keeping prices down via that. OpenAI needs Amazon (or some other fulfillment company) to deliver products. Amazon does not need OpenAI - they can build their own recommendation engine or work with another.

> Because amazon doesn't have a web search service but they do have a product recommendation service?

That's the whole point here. People use web search as a product recommendation service, even though Amazon has one natively. What makes you think people won't (and they already are, in massive numbers) use chatbots for product recommendations and web search?


But OpenAI has the attention. It's where people ask for product recommendations, and it has context about the user. Surely Amazon doesn't need OpenAI, but OpenAI will be another valuable distribution channel for them - unless some other LLM takes the crown.

> Your kids are a bad influence for whatever reason? Not my problem to fix.

Not your problem to fix for sure - but it is your problem to equip your child to comfortably weather. There are bad influences out in the world and they generally have outsized effects on their social and professional scenes. In fact, the kind of curated, limited community you're advocating for is one where bad influences thrive.

> So much time and resources, to produce what exactly? A piece of paper and fancy picture to stare at?

I certainly agree the degree is whatever - but I think you're really under-valuing the social-gauntlet aspect of school. You will have classmates who kind of (or really) suck. You will need to do your work anyway. You will be incentivized to learn perseverance and a self-centered locus of control. These are valuable skills that only come from actual exposure to bad influences.

Someone who's perfect in perfect conditions is going to struggle because the world is not perfect. The aims you highlight here make me think less of homeschooling than I did before.


I think we have a mass re-assessment coming for how we think about data collected in public spaces. The realities of mass surveillance and mass data correlation come to very different outcomes than they did when we established our current rules about what is allowed in public spaces.

I don't really know what a better system looks like - but I suspect it has to do with the step where the info is provided to a third party. We can all exist in public and we can all take in whatever is happening in public - but it's not clear that passing that observation on to a third party who wasn't in public is an important freedom. Obviously this cuts both ways and we need to think carefully about preserving citizens rights to observe and report on the behavior of authorities (though also you could argue that reporting on people doing their jobs in the public space is different than reporting on private citizens).


> [I]t's not clear that passing that observation on to a third party who wasn't in public is an important freedom.

It's not hard to imagine a restriction on reporting one's observations failing any number of First Amendment challenges.


> we can all take in whatever is happening in public

People have the right to take in what is in public, but maybe cameras should not?

This could apply to everyone in public spaces. No video, audio or surveillance without obtaining permission. Better blur anything you share, or you might get busted. The least we could do is restrict corporations from possessing such data.

Similar to what Germany does with doorbell cameras, making it illegal to film anything outside of your property, like a public sidewalk or the neighbors house. It is my understanding that people there will confront someone taking pictures of them without their consent.


    > People have the right to take in what is in public
You write this as if it is a fundamental human right. I disagree. I could imagine this could be treated differently in different cultures. As an example, Google Maps has heavily censored their Street View in Germany to scrub any personal info (including faces). Another common issue that is handled very differently in different cultures: How to control video recording in public places.


> Google Maps has heavily censored their Street View in Germany to scrub any personal info

I remember when this first launched in the UK, automated face-scrubbing was in place. It was about 90% accurate on scrubbing faces from pictures. One of its best screwups was showing people's faces as they were standing outside a branch of KFC but blurring out the Colonel.


>You write this as if it is a fundamental human right. I disagree.

It's more common sense than any real sense of law. If something is a public space, how do you stop people from "taking it in"?

Recording is a different matter, but people existing is what comprises the "public".


> how do you stop people from "taking it in"

Please take a moment to draw for us detailed faces of all the people you've "taken in" today while you were outside. Use a sketch artist if you need to. Now compare those results with what you'd have if you did the same with a photocamera. And for good measure, add in the amount of effort it took you to recall, and the effort it will take you to describe to every reader on HN who you saw today.

Do you really not see any difference between the human process and what a digital camera can do?


I think we're agreeing but our frequencies are mixed. I was just saying "you can't stop people from using their eyes in public".photography and recording laws are very different.

for more context, the chain started with this:

>People have the right to take in what is in public, but maybe cameras should not?

and then the direct reply disagreed with this notion. I just wanted to distinguish between "taking in" and cameras, because it appears that user made a similar mistake.


I dunno - I think there are uses of surveillance in pursuit of enforcing laws that I don't think are harmful. Like...maybe you can record the public and pass it on to the police when there's a specific request for a time and place that a crime was allegedly committed? Like - if an organization has a legitimate interest in what happened there you can pass on your recording. But you can't just sell it to some random data broker, because they don't have a specific reason to want a recording of that place at that time.


My jaded AF crystal ball called history says that these things never change until the petite-bourgeoise (I'm no Marx fan, but I think he did a good job with that part of his social class classification system) are seriously harmed by it. The rulers don't care. The poor have real problems. This sorts of crap happens or doesn't happen at the behest of the materially comfortable people in the middle. And it seems like they never learn except the hard way.


It's not something I am excited about, but it is something I want my IDE to do well if I must engage with it. Other remote pair programming experiences are even worse and I appreciate Zed's capability in the area even if it's not what I prefer.

A lot of my IDE choices are about extensibility and flexibility more than perfection for my preferred coding approach. After all, until I only work for myself I need to be ready to accommodate the needs of others as part of my job.


The open source aspect seems worthy of attention if nothing else. Even if this is a middle-of-the-road scanner - the community being able to customize, improve and support it would be incredible. Especially considering your scanner is considered one of the best despite being over 20 years old.


The way I talk about is is that the value you deliver as a software "engineer" is: taste and good guesses. Anyone can bang out code given enough time. Anyone can read docs on how to implement an algorithm and implement it eventually. The way you deliver value is by having a feel for the service and good instincts about where to look first and how to approach problems. The only way to develop that taste and familiarity is to work on stuff yourself.

Once you can show, without doubt, what you should do software engineers have very little value. The reason they are still essential is that product choices are generally made under very ambiguous conditions. John Carmack said "If you aren't sure which way to do something, do it both ways and see which works better."[1] This might seem like it goes against what I am saying but actually narrowing "everything possible" to two options is huge value! That is a lot of what you provide as an engineer and the only way you are going to hone that sense is by working on your company's' product in production.

[1] https://aeflash.com/2013-01/john-carmack.html


I mean, if you click on the first link you will open the paper they are summarizing[1]. It's a meta analysis of 64 studies, so you could certainly go through the studies and look at each population.

However, the actual answer is that all population studies are only gross generalizations that may not apply to you. They are often quite useful because the odds are generally good that they do apply...but it's never certain. Even if you are a member of the studied population your specific circumstances may overwhelm your populations norms.

[1] https://pubmed.ncbi.nlm.nih.gov/38011755/


Yeah, but I am at the tail end of these statistical curves and my life was absolute hell and doctors (but one) did nothing for me. And they kept telling me to eat fiber cause my cholesterol was so high and HDL was too low (30)!

If the meta analysis showed population differences, why did the article not bring it up? This is what is wrong with nutrition research, then never account fro genetics despite the huge about of evidence that it is extremely importnat.

The truth is that fiber does not reduce mortality for everyone by 23%. I would rather not be guessing with science and health. I lived through that and it took me years to get out of it.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: