Hacker Newsnew | past | comments | ask | show | jobs | submit | jiberius's commentslogin

iOS-style runtime permissions (for things like location access) were introduced in Android 6.0, which was released back in 2015. That's quite the long play.


It dates all the way back to the Pixel 2 from 2017, still feels like a magical feature though.

I think the historical log of songs that the phone heard launched a little later, on the Pixel 3.


The biggest reason was that the Apollo program was able to return film from the surface of the Moon to Earth. Many of these photos weren't returned to Earth in film form. Instead they were developed onboard the spacecraft, scanned, transmitted via radio, and then recorded onto tape on the ground. Every step of that process incurs a decent bit of quality loss.

You can actually tell which probes had a reentry module because the photo quality is pretty stunning compared to the photos transmitted over radio, look at Zond 8 which returned film negatives to Earth.


I've never heard this. I'm not saying it's wrong.

I was told the all the Astronauts were trained on the Hasselblads, and Hasselblads are great cameras--even today.

The great high def pictures were developed on earth, but I'm not sure now.

I remember hearing they came home with hundreds if rolls of film.

The film footage was sent to earth via radio waves, but was grainy.


American videocameras were way superior as well.

Tube cameras Russia used up into 200x were inferior than RCA ones from sixties.

Consumer technology, and monetary incentives make mirracles.

Hard to fathom that today's "supercomputer in the pocket" came out of efforts to build superior gaming PCs, and frag more people in Q3.


> American videocameras were way superior as well.

I know you didn't phrase it exactly that way.

But we in Sweden are somewhat proud of supplying the Hasselblad cameras to the Apollo missions.


While I greatly dislike the nationalistic undertones in your comment, you are – unlike what some responses claim – not wrong in that the video cameras for the Apollo 11 mission [1] were indeed American and made by Westinghouse [2] and lenses by Fairchild [3]. Not sure how much I would argue that it was “consumer technology” though, at least not at the time.

[1]: https://en.wikipedia.org/wiki/Apollo_TV_camera#Westinghouse_...

[2]: https://en.wikipedia.org/wiki/Westinghouse_Electric_Corporat...

[3]: https://en.wikipedia.org/wiki/Fairchild_Camera_and_Instrumen...

As for the more famous non-video cameras, they were indeed made by Hasselblad [4] (Sweden) and the lenses by Zeiss [5] (West German). I am of course acutely aware of the latter as I shoot Zeiss lenses and prefer their aesthetics to pretty much anything else I have tried, although these days a chunk of them are made by Cosina [6] (Japan). If you want to dig deeper, I liked Hasselblad’s official homepage on the matter [7].

[4]: https://en.wikipedia.org/wiki/Hasselblad

[5]: https://en.wikipedia.org/wiki/Carl_Zeiss_AG

[6]: https://en.wikipedia.org/wiki/Cosina

[7]: https://www.hasselblad.com/inspiration/history/hasselblad-in...


The Russians got the Zeiss factory as spoils of WWII so they made at least some decent optics.


Except those american videocameras were probably made in Japan. Sony, Nikon, Ikegami, JVC, ...


There were no Japanese videocameras in sixties


A quick google says you're wrong: https://global.canon/en/c-museum/product/cine253.html

What is it with Internet commenters who are cocksure despite not knowing what they're talking about?


This is not a videocamera


Well, color me embarassed. But why the pivot of the discussion to the video cameras? Maybe to emphasize "America superior"?


What is it with Internet commenters who are cocksure despite not knowing what they're talking about?


I completely agree, I'm sick of people being cocksure about something despite not knowing what they're talking about.

https://en.wikipedia.org/wiki/Ikegami_Tsushinki#History


The delay after taking long exposures is caused by the camera taking a dark frame with the same exposure length as your actual image. When you take a long exposure, the sensor can exhibit noise caused by "hot pixels". The dark frame is taken with the shutter closed so that the camera can get an image of just the hot pixel noise, which it then subtracts from your actual long exposure.

You can turn this off in the settings, but I imagine it's way easier for the camera to correct for this specific kind of noise than for Lightroom to do so.

I believe it kicks in once the exposure length is >2" on my Sony A7ii.


On the one hand that's correct. On the other hand you could completely avoid that if you did long exposures the way a Google Pixel (2+ I think) does it (in software):

Take lots of short exposures and fuse them. If the device is handheld you get variability in positioning for free, if it's on a tripod, it will automatically wiggle the OIS slightly to achieve the same effect.


It seems like movement would be effective for averaging out random noise, but the described process isn’t for eliminating random noise, it’s for eliminating persistent hotter pixels on the sensor.


Indeed, and in fact the pixel does exactly the same dark frame trick to identify hot pixels. I don't know that it takes as long though - perhaps it's hidden by the fact you can keep doing other things while the photo is processing.


It cannot do the same thing as an a6000, as it lacks a physical shutter.


That's great information and makes me much less frustrated with my slow a6000.


You can also try turning off long exposure NR in the menus. I don't know if it makes enough of a difference on the a6000 to be worth leaving enabled and paying the extra cost in time, but that's a way to find out.


Oh thanks for letting me know! I have an A6000, so that makes sense!


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: