Jump to content

tupp

Members
  • Content Count

    856
  • Joined

  • Last visited

Everything posted by tupp

  1. On one feature I shot, we used a "Snow White"-themed rolling case, very similar to this: We put in a little padding in it, and it was very convenient in carrying the fully built camera. When we wanted the camera on set, we would call, "'Princess Case' on set!" When shooting in crowds, we could walk a fair distance away from the case, and nobody bothered it.
  2. I am primarily concerned about the risk of losing your camera and more so about the hazard that the rig poses for motorists behind the hero vehicle. At minimum, replace the open hooks with strong carabiners or with removable chain links. The top carabiner needs to completely encompass that runner on the luggage rack, so that it cannot fall off during a bounce. In addition, to prevent the bounce/wobble, solidity could be added with two extra ratchet straps (or motorcycle straps) -- one strap tensioned between the camera platform and to the top of the car and the other strap tensioned between the camera platform and the bottom of the car. These extra straps would also increase safety.
  3. Not familiar with Glidecam (as much as Steadicam), but there has to be a Glidecam model with enough capacity for a Sony F3. Be aware that Steadicam-style stabilizers are not something that the typical gimbal-kiddie can just pick up and instantly start shooting -- it takes a bit of practice and training. The best Steadicam operators have years of experience. On the other hand, I would bet that gimbals exist that could hold an F3 with an FD prime. I see two alarming problems with the rig pictured: That tag line with the hooks should be replaced with a solid strut (or, even better, two "triangulated" struts). Every significant bump will cause the camera to bounce up and down (ruining that part of the take). There is no backup "safety" portion of the rig -- a rig that uses suction cups and a tag line with open hooks. With each bounce, there is a possibility that one of those open hooks could fall off of its pick point, and if that happens, "that's all she wrote" for the camera (and possibly for a motorcyclist following the car). If they had to go with the tag line (instead of a strut), those open hooks should have been carabiners. Regardless, any car rig should have separate, properly-tensioned safety straps, and the pictured rig has nothing in that regard. A typical grip hostess tray with risers and a head would be more secure and would be easier to rig and adjust. Also, I am not an audio person, but why is that mic mounted like that on the camera with a car rig? I strongly urge you to go review several different tutorials on how to properly and safely rig car mounts, before trying to do so yourself.
  4. Meant to say, "The darker the image, the more saturation seems conspicuous -- unless, of course, the image gets so dark that it's mostly black."
  5. Generally, the darker the image, the more the saturation. Furthermore, most digital cameras give a lot of saturation in their non-raw files, and Canon cameras additionally boost the reds. Starting with the brightest sample posted, the image below was yielded merely by boosting the gamma/mid-tones, bringing the blacks down to zero, reducing the saturation and backing off the reds (for Canon): If one wants to keep it a little darker (and still have it look like daytime), be more gentle in boosting the mid-tones but further reduce the saturation, and keep the blacks at zero and keep the Canon reds reduced as in the image directly above: By the way, the fringing/chromatic-aberration doesn't look too bad, and a light touch with a CA/fringe filter should take care of it nicely.
  6. tupp

    bmp4k adventures

    Of course, all of the battery plates that I have linked/mentioned already have mounting holes/screws. Furthermore, each of the videos that I have linked show the plates mounted to cameras and cages.
  7. tupp

    bmp4k adventures

    The first battery plate that I linked (with the BMP4K power cable) is metal. Regardless, here is a video that gives a rundown on the various sizes of NP batteries.
  8. make sure your power switch is turned off when you plug in your light (or any other device); wear gloves if you are not familiar with how to handle hot lights; only mount gels to the side barndoors or to a gel frame made for the fixture.
  9. Tungsten lights are a good deal right now. In a recent interview, Roger Deakins talked about LEDs vs. tungsten.
  10. tupp

    bmp4k adventures

    NP batteries are a good way to go for most cameras, and there are plenty of options both more expensive and much less expensive than that small rig battery plate. Here is an NP battery plate with a BMP4K connector for US$38. If I had a BMP4K and I wanted to use NP batteries, I would just get a cheap plastic battery plate and wire it to a BMP4K connector (with pigtails), which would probably give a total cost of about US$15. Here is Chung Dha's inexpensive NP battery plate video. Keep in mind that some NP battery plates allow slight leeching of current from the battery by the connected equipment and by the plate's own LED indicators and circuitry. Here is The Frugal Filmmaker's video on how he installed a cut-off switch on plastic NP battery plates that prevents that current leeching.
  11. As you may recall, we have previously discussed the Topaz Labs software, including "JPEG to RAW AI" Again, the color depth of the original image cannot be increased unless something artificial is added. Likewise, detail lost in extreme highlights and shadows cannot be recreated unless something artificial is added. It seems that AI can provide more color depth and create details, as shown in the Topaz software. Removing artifacts such as banding does not indicate an increase in bit depth -- it just means that the artifacts have been removed. The resulting image without banding can have the same bit depth as the original image that suffers banding. COLOR DEPTH = RESOLUTION x BIT DEPTH. So, if one can increase the resolution and maintain the same bit depth, then the color depth increases. Similarly, if one can increase the bit depth and maintain the same resolution, then color depth increases. Of course, merely putting an 8-bit image into a 10-bit container will not increase the color depth of the original image, nor will merely up-ressing an image increase the color depth. Something artificial has to be introduced to increase color depth in a given image.
  12. I mentioned the Metabones G-mount Expander in an earlier GFX 100 thread. It gives a 1:1 crop factor with full frame lenses on the GFX 100. So, no vignetting worries. However, the only current Expander model accepts Nikkor lenses.
  13. I am sure. Kdenlive, Cinelerra (GG) or Blender is what I would recommend. I have heard good things about Shotcut and Olive. Openshot would also work for someone who doesn't need anything fancy. One of the the great things about Cinelerra is the Blue Banana plugin -- a very unique and powerful color grading interface. I wish that somebody would port it to MLV-app. If one is okay with proprietary software, then there is Resolve, Lightworks or Piranha. All of the open source NLEs work on most distros. On the other hand, there are special media distros that are worth considering, such as AV Linux and Ubuntu Studio. I tried Resolve once, and I as I recall it was distributed as a tarball, and I had no problem installing it on my Debian-based distro. Not sure where you are from, but, as I recall, my current Ebay machine cost US $145. It has an i7-3770, 3.4GHz cpu, and it came with 16 gigs of ram, two mediocre graphics cards and a 500GB drive. I put an SSD in it and loaded a non-systemd distro. It's fairly snappy.
  14. You can change the export settings in Keynote. Set it to 1920x1080 (or smaller) and to ProRes 422. Evidently, there is no way to adjust frame rate (defaults to 29.97) nor bitrate in the current version of Keynote. First of all, Mac is not the easiest OS to use. By the way, almost all OS's are point-and-click, so even a Mac user should have little trouble working with them . Furthermore, the free transcoding programs that I mentioned (ffmpeg, handbrake, mencoder) all work the same on any platform, be it Mac, Windows, Linux, BSD, etc. I suggested a Linux rendering box for... rendering (the proxies). In such a scenario, you would still use your current MacBook to edit the renders off of the aforementioned external SSD. On the other hand, you could easily do your entire post production (including presentation animation and editing) on the same Linux box, and it could cost you as little as $200 for a used machine that is adequately snappy. The software would all be free and open source. You still have not given any information on the number of clips per video that you edit nor on the length of those clips. If you use 5 clips per video and they are each 5 minutes in length, start the proxies rendering (in FCP?) on your current MacBook, and go make a cup of coffee. If you are making 3 videos per week with 2 hours of camera clips to edit for each video, a cheap, separate rendering box is likely more efficient. Your heightened teaching efforts are certainly welcome! However, it is doubtful that 4K is any more engaging to students than regular HD. Great!
  15. If the files coming off of those devices are compressed, your machine (and NLE) has to continually uncompress those files on the fly, while applying effects and filters. It's a huge demand on the computer's resources. "Keynote?" That sounds like a cute Apple name for a presentation app. Make those clips uncompressed at a low bitrate. How long and how many are your video clips? Try creating proxies with a couple of files, and see how long it takes. If you have a lot of clips to convert to proxies, you could also build a cheap Linux box and batch render your proxies (with ffmpeg, handbrake, mencoder, etc.) to a fast SSD drive. Then, edit off of that SSD. That workflow might payoff if you are doing three videos a week. Evidently, teaching has changed dramatically since I attended school. You are making more videos per week than a lot of pros make in a month. If this is for teaching, why do you need 4K? Try reducing the resolution to HD and reduce the bitrate wherever possible. What happened to chalkboards? US $2100 for a small 2Ghz laptop with a 512 SSD?! Before dropping that kind of money for a laptop of questionable power/quality, I would look into streamlining your workflow, as suggested above. Again, with any MacBook (or any Mac) that has a T2 chip, make sure that secure boot is disabled.
  16. Don't know much specifically about your gear, but merely using proxies should give a huge performance boost. Working with compressed camera files can slow things down to a crawl and cause discrepancies in effects and color grading. You shouldn't need high quality files until grading and rendering. Some graders transcode camera files to uncompressed and then work on them. Regardless, if you get a new MacBook, Louis Rossmann (who makes a living repairing MacBooks) warns folks to disable secure boot. If your current MacBook has a T2 chip, you should make sure that secure boot is disabled.
  17. In this lens re-greasing video, the vlogger used LiquiMoly LM47 MoS2. On the other hand, the same guy used a different grease in an earlier video.
  18. Interesting article and blog post! Many folks prefer the look of vintage lenses with digital sensors. It's good that Cooke has noticed this trend and reacted to it. Of course, they are not the only lens manufacturer to come out with brand-new "vintage" lines. It would be great if someone would test the character of the new Cooke "vintage" lenses against that of their old "Xtal Espress" anamorphics.
  19. Not sure how the highlight/shadow control could happen after encoding. By encoding, do you mean "conversion to 8-bit?" If so, I have no clue as to what stage in the camera's imaging process that the highlight/shadow control is applied, but I would guess that the 8-bit conversion happens early at a low level, before most other processes.
  20. I never experienced artifacts with Cinestyle. Are you referring to compression artifacts in the shadows or to posteriztion/banding? At any rate, I haven't noticed problems on the E-M10 III with my highlight/shadow settings (in my brief experience so far with the camera), but, again, I am using a light touch with those settings. I don't have any short clips, otherwise I would post them. When I get a chance, I will try to snip out a few seconds from one of the files for download -- I think that ffmpeg can do so without any transcoding.
  21. One thing that never gets mentioned about the E-M10 III is that, although it cannot employ custom picture profiles, it does share the highlight/shadow control feature found in other OMD cameras. This attribute allows changes in the camera's contrast curve over a large range of values. It's a powerful control, and one must use a light touch to avoid pushing the curve too far, as it can look unnatural. I set the highlights to "-1" and the shadows to "+1," which levels the contrast curve a bit. Additionally, I enable the "Muted" picture profile, with @TiJoBa's recommended "-2" setting for sharpness and with a "0" setting for saturation. This Imaging-Resource review gives examples of how the highlight/shadow control can affect those areas of the contrast curve. Scroll down to the "Highlight/Shadow Control" section and "mouse over" the different values to see how it changes the detail and brightness in those areas. By the way, the OMD highlight/shadow control also allows adjustment of the midrange values (at the "center cross" in the display). Eventually, I will test setting the shadows to "+2," the midrange to +1 and the highlights to "0."
  22. On the contrary, ML is thriving. You can't go by that nightly build page, as that is not where the action is. Most of the nightly builds that everyone uses are not official. To see the current activity, go to the main forum page, and scroll down to the bottom section titled, "Recently Updated Topics." After a brief scanning of just a few of the top messages, I see the following active developers: Danne, masc, cmh, Levas, ilia3101, reddeercity, 2blackbar, critix. Our own @ZEEK is active with ML and MLV-App instructional videos.
  23. Yep. Again, you don't necessarily need a separate clamp -- you could just bolt an L-bracket directly to the cage. Of course, using Arca-Swiss clamps or other quick-release system makes the changeovers faster (and adds height to the camera).
  24. I think most L-Brackets take 1/4-20 threads on both sides, so you might not even need a separate mounting clamp -- Arca-Swiss or otherwise.
  25. tupp

    How to edit H265?

    Most editors that I know don't edit camera files -- they use lower-quality, uncompressed proxies and/or transcode the entire project to an NLE-compatible, high quality format. Not only can using compressed camera files slow down the NLE and cause problems with effects, but grading compressed files can also cause discrepancies with the rendered look. My editor (and color-grading) friends usually grade with a high quality format after they have edited with proxies. Another trick when working on narrative projects -- use multiple drives to speed things up. For instance, in a two person dialog scene, put all of Character A's shots on one drive, and then put all of character B's shots on another drive. This will allow cutting between different drives instead of cutting within a single drive. Likewise, one could put close-ups on one drive, the medium shots on another drive and all of the wide shots on yet another drive.
×
×
  • Create New...