Jump to content


  • Posts

  • Joined

  • Last visited

Posts posted by no_connection

  1. No wonder I could not find any manual or info about it then, NDI sounds really interesting and although I have only played with it in OBS it seems to "just work". File transfer would make the Ninja V a lot more worth. But if they are not going to release it it does kinda put a damper on the enthusiasm or gas factor of it if you will.

    I live in Sweden so within EU could work.

    I have thought about doing some diy solution too with something like nVidia Jetson (just found out that was a thing) and a USB card or similar.

  2. We are kind of looking at the Atomos Ninja V, it have a lot of expansion available for lots of uses many of which we don't *need* atm. The ability to use it as a portable recorder it nice tho.

    We currently run 720p50 for live video through atem 1 m/e so the actual demand is a bit different.

    Do anyone have any ideas of a better recording solution at same or less price?

    It probably have to be ssd based since SD card just don't work (learned that the hard way with BMD HyperDeck Studio Mini that simply refused to record)

  3. The UHD have more detail at 100% than the HD at 100% which is a bit more muddy and seem to show more compression artifacts. And yes the upscale is not super great but nether was the HD it came from. You can't really get away from the plastic look when viewed up close and you rely on the source material to work decently for upscale. If there is not enough detail to work with it might not look good no matter what, if you scale up mud you are going to get mud, or a lot more mud, idk.

    If you have a HD or UHD screen does not really matter for comparison since you are watching a part of an image at 1:1 ratio, and as long as you have an idea how large the real image would be you can guesstimate how it would look. Sitting at the same pixel density as I do now but with a 4k screen would be painful to watch. I sit 75cm(30") from a 24" 1920x1200 screen if it matters.

    Here is a fun one then, the UHD scaled to 50% then back again.

    I used waifu2x to upscale.

    Oh and if it was not clear the source is @KnightsFan


    *edit* I should note that the downscale was maybe not the best but it also highlight one of the bigger limitations, you are not going to get thin lines and the finer/small details unless it happens to taper away in a manner the scaling algorithm picks it up, but that works mostly with 2d art.

  4. 3 hours ago, androidlad said:

    In terms of colour spectrum, the LCD panel has no effect on the backlight, the panel is literally millions of tiny windows that open or close to modulate the backlight.

    The panel is literally millions of Red Green and Blue windows.


    3 hours ago, androidlad said:

    Because our eyes can interpolate colours in between sharp RGB spikes and camera sensor can't, that's a big problem and it does matter.

    A camera have Red Green and Blue filters in front if the pixels, sounds similar to anything?

    I could add that some Tv do have white pixels to add brightness, like some projectors. Think some cameras have done this as well.

  5. Depending on model E5-26xx or E5-24xx. Difference being socket and number of PCIe lanes. I have been looking a bit at E5-2650 v2 that should be decently compared against 2600 but with 8 cores and you can have two of them.

    E5-2620 might be too slow with E3-2630 minimum entry depending on price. It's like E5-2650 but with 6 cores instead of 8. v2 have higher clock speed.

    It is easy to find the subpar garbage CPU tho, like E5-2407 or E5-2609. Those two are common entry level CPU but ridiculously slow to the point where I have no idea why they where every made. Even the "modern" generation of the same CPU is worse than the old stuff.

  6. I'm confused, you quote me with a Nah then say exactly what I said.

    I don't think anyone used RGB backlight, at least I don't think I saw one. R G and B components comes from the filtered array of pixels, you are going to have that no matter what. But old LCD ( as in old) had more washed out colors so therefore have larger humps instead of spikes, so "better" CRI but are still limited by the CCT driving it.

    39 minutes ago, androidlad said:

    High CRI/TLCI requires full spectrum light source and some high end white LED backlight can indeed provide.

    For the very reason that you need very saturated primaries to produce a wide gamut it will not have a full spectrum once it comes out the front end so I doubt they would waste power to produce it. That is why Quantom dot exist to create very controlled spikes that fit together with the RGB filters. Or rather R G and B Quantum dot cells.

  7. No it would have very bad CRI since it's literally a R G and B source fit together in a very tight space. Granted a camera sees R G and B so it might be fine.
    Light from a LCD is also polarized which can be useful.

    Don't be afraid to use low CRI lights if they give what you are after. And work around the times when they are not.


    I would theorize that older LCD might have better CRI since they are more washed out and probably covers more of the spectrum when white, on the flipside quantum dot probably is worse due to more saturated and narrow R G and B components.

  8. I would not say quickly, but if you keep it running 24/7 for 5 years it does add up. You are only adding power from fans and a little from using dual psu, you are getting less performance per watt due to older generations but it's not that bad.

    You might be looking at 80-150W depending on server vs maybe 50W. So at worst maybe 20-50$ a year if you do run it 24/7. I have no idea what you pay for electricity or cooling and if you only keep it for a year and sell it you might save more in price drop of a new system than electricity would cost. In a year you might even be editing in the cloud for all we know.

    Used workstations should be plenty and cheap too if you find them.

    What kind of servers have you found and for what cost?

  9. It depends, you can get dual 8 or 10 core that would probably out perform it in encoding power, and you can get a crapton of more RAM.

    But with Ryzen you will have an easier time getting fast storage due to NVMe and the RAM you get will also be faster, although not sure that matters. Power draw during it's life time it also a thing, although if you buy a server and have it for a year and then flip it and then Ryzen next gen will have dropped in price. Ryzen and what you need to run it is kind of expensive right now, especially if you look to have 32GB+ of RAM. I would want ECC RAM too, some ppl would not say it matters but I just seem to have things crash less on my computer than some other ppl experience. Or it might be cause I don't do dumb stuff, who knows. Then again modern RAM is way better than it used to be.

  10. I don't currently use a rack server as a workstation but I have experience with servers. I have used several tower servers as workstations over the years mainly HP.
    Dremel is your friend if you need to put a 16x card in a 8x slot, did that to a HP ML350 G5 and used that for a few years. It was loud but compared to the G4 it was a mouse in a corner. That thing have a 36W fan as case fan and 17W in the psu.

    I currently use a Lenovo TS140 as workstatoin and it has been great.

    I do intend to get a 2u server for ESXi and workstation use but it has to be good spec and price locally and not that much pop up here.

    I do like IBM servers as well as HP. You can get 2x 10G NIC to mount instead of 4x 1G Ethernet without taking up extra slot. HP G8 or IBM M4 would be the oldest to get and that will get you into sandy bridge generation I think. Not much have happened since then, just higher clock speed and less power draw. Well until newest gen AMD cpu. (don't bother with old AMD cpu systems, they where never good, even when new)

    Most servers you can get are not going to compare well in snappiness and single core performance so do factor that in your calculation. But since CPU and RAM price of a new system will get you a running server it might be a decent tradeof. Spectre and Meltdown mitigation is also a question mark.

  11. Depending on what you want to do it can be a decent deal.

    A few things to note that may or may not be applicable so I just made a list.

    DDR3 is less fun for electricity than DDR3L, so HP g8 or newer for example.
    Power draw is per RAM module so don't get one filled to the brim with 2GB sticks.
    Many server CPU have weird performance so look it up on PassMark so you don't get a 130W part that perform badly for no apparent reason.
    The difference between a v1 and v3 can be lower clocks and less power (or not) so again check the performance, a v1 might not be too bad but also might be.
    You often need two CPUs to use all PCIe slots on a 2u server, and most need a raiser card unless server was bought with it.
    Very few have PCIe power so you night need to get adapter cable or be creative.
    I have not run into any issued with HP when it comes to supplying power through PCIe port to power graphics card but I know some Dell might not want to supply more than 20W.
    Getting 64GB RAM should be pretty cheap, try that with DDR4 on Ryzen.
    You could probably make a caddy to mount SSD into the slots since you probably not going to get HDD with it.
    If you plan to get a cheaper server with v1 cpu and upgrade to v2 make sure bios support it.
    If you get a server with only 1 cpu it probably don't have all the fans to just put another cpu in it.
    As mentioned noise might be an issue especially for 1u but I think 2u is pretty decent depending on model, you could always make a box for it since servers have plenty of cooling power as long as you think of air flow.
    Startup times can be long since servers like to check everything, so 5min+ boot time is not uncommon.
    Also don't turn on/off a server every 5min or you will break it with time.

  12. One option if you want interactive is having instant slomo button that will play back the last set of time in slow motion. Or trigger a recording that then gets played back.

    As for audio a good place to start is Paul's Extreme Sound Stretch made to stretch audio out to extreme lengths (hence the name). Here linked in a portable form so you don't have to install anything. I'm pretty sure there would be some sort of version that exists with command line api to better suit your needs.


  13. I'm not arguing it's a great move by Sony, just sharing what ppl have told me, ppl that are asked by Sony because they are working in that field. 200 was for world cup (don't quote me on that number tho but I think it was something like that) Olympics probably have a lot more ppl than that.

    5 hours ago, gt3rs said:

    And me and many other colleagues that do both photos and video we are not the target?

    If it was it would have log and all those things right? I mean you are not wrong in being grumpy about Sony not releasing such a camera.

    5 hours ago, gt3rs said:

    Just out of curiosity how many A9 or 1Dx or D5 do you have and use for?

    I have nether, but my friend had a few Canon bodies, 1Dx MkII 400/2.8L 70-200/2.8L. That sort of thing.

    I'm just sharing what I know, I will not argue more about it.

  14. Are you telling me you sit at world cup with 200 of the worlds best sports photographers with that much $ in gear and can't afford a 2nd body for video use? You simply don't do video and many times aren't allowed to ether. I have had argument about this with someone that is in the intended audience for this camera and it is a different world. When you are at such an event getting imaged uploaded to the agency while you shoot is one of those make or break things that Sony never had to offer for example. I would argue that pro photographers covering the Olympics is what Sony want to target. Not just action and sports. I'm not sure if my friend is going to the Olympics but if he is you could argue with him there.

  15. On 10/4/2019 at 10:24 PM, gt3rs said:

    Why no picture profiles, log, 4k 60p etc..?

    Simply because the intended audience for it don't care or have use for it. You may not realize but features to cost to implement and since this is a pro camera just having it in there and probably working is not good enough, it needs extensive testing (that costs a lot) and if it impacts the reliability of any other function then there you go. Remember the blinking pixel from a7riii was it? That is simply not acceptable, might be to consumers but ppl who use it as a tool where you need something to work can't be beta testers like in the past. You might be sitting with 200 other photographers contending to get that one shot that will get published, things do take a different perspective. So it's not as simple as a "firmware update".

    And yes it does such for those that have been waiting or wanting the be all end all super tech thing that does all things but in a more pro body, maybe Sony could add a "creators" firmware with all the functions but with all the alpha testing (pun intended) that comes with it.

    So the a9ii is a giant upgrade to make it viable where the a9 didn't have the things it needed in that environment. Image quality or megapixels was never what held the a9 back in that regard, usability where it matters was.

  16. Pretty sure they could have done that with a tilty flippy screen as well.

    IIRC xpro2 had dual zoom level for the OVF.

    I played around with the xpro1 a few years ago and it was interesting, way too small rear lens tho that made it kinda bad to use unless you aimed your eyeball in there, it had some other downfalls like battery life and the reeeely noisy os zoom lens so I didn't end up buying it even tho it was pretty cheap used. I would not mind to see what the new generation have to offer.

  17. I would say rolling shutter is pretty important. did they say anything about that? Also it looks like the heatsink is pretty accessible to outside air, could that mean that if you add external forced air cooling and power it could be cranked up a bit?

    They could ditch internal recording to SD and just have 512GB NVME in there, and maybe external SD or similar "slot" to dump footage.

    I spot a s16 1080 mode in there, give me c-mount adapter and good frame rate and it would be reeely neat.

  18. You could take a decently bright LCD projector and aim it at a diffuser or backprojection(for a bright backdrop) depending on the need. Then make a fiery looking thing to play on it. Should be pretty cheap to find some older low res but bright one too as a dynamic light source. Black level probably not going to be an issue when used in this way. Btw I meant to use the projected image as a light source and not shining the projector directly at the subject, although that could be useful too.
    With some masking you could add reflectors and add "virtual" lights and many other neat things, imagination and time/skill it the limit.

  19. https://www.partha.com/

    I use this version if gimp as it's more like photoshop than regular gimp, and it comes with some very useful things like wavelet sharpen/decompose.

    I have not tried the newly released version so it might be even better. Grab the portable version and you don't even have to install it, just run it from anywhere.

    BTW portable is the best way to run anything (IMO) and I do it whenever I can, like web browser that I can just delete folder and open from backup if anything breaks or keep several version of depending on what I do with it.

  20. I don't watch game of thrones but I could definitely watch the bts.

    I would guess the biggest problem of it being too dark is clipping to video level that many phones do, even tho it should have been viewed in full range.

    I really don't think the guy can pull the "they have badly calib..." it just means he don't understand how ppl view the content or don't care.

    The source material on a good screen probably looked great tho even it it was dark, after usual compression (that murders anything dark by design) probably not so much.

  • Create New...