Jump to content

Banner.jpg.b83b14cd4142fe10848741bb2a14c66b.jpg

symmetal

Members
  • Posts

    2,406
  • Joined

  • Last visited

  • Days Won

    3

Everything posted by symmetal

  1. Great image. Nicely framed with the clouds. 😀 Alan
  2. I assume you're referring to using a field flattener or coma corrector, and the optimum correct distance between it and the image plane. You set the back distance to get round stars at the corners of the image and this should hopefully be the same distance to get best overall focus as well. You can get good focus near the centre of the image at any spacing distance, but as you move further from the image centre the spacing distance becomes more critical which is why the corners of the image show the spacing errors the most and are used after making small distance corrections to see if they look any better. Field flatteners give elongated stars at the image corners at incorrect spacing while coma correctors will still show some coma when incorrectly spaced. I've never used a coma corrector so assume that's what you'll see. 😉 If you have tilt in the imaging train then you'll find not all corners show the same star shape errors. One corner may look good while another has misshapen stars. Altering the distance will improve the misshapen stars corner, but the previously 'good' corner may then start showing errors. Alan
  3. Varying offset values for different gain settings is more trouble than it's worth. Just use one offset value that avoids black clipping at any of the gain values you may use. The gain in dynamic range by using a smaller offset value is negligible. Alan
  4. By USB Speed do you mean using USB2 vs USB3 or the USB Speed setting in the driver. The latest cameras use memory in the camera to read out and store the images quickly before transferring them to your PC via USB. Earlier cameras without this memory had issues with excessive amp glow using USB2, mainly with short exposure imaging. So now USB2 just means it takes longer to download the image and doesn't affect the quality. The USB speed setting in the driver is to avoid sending the data too fast that the PC can't cope. If set too high you may get dropped frames or no image at all. As many CMOS camera have 'amp-glow', (the later more expensive ones don't) bias frames are not very useful as you need to use darks to calibrate your lights and flats. The darks already contain the bias signal so using darks takes care of bias and amp-glow. The position of the peaks in the histogram don't need to match the lights. The light peaks will always be higher than the darks as extra light is entering the camera causing the peak to move to the right with increasing exposure. The offset needs to be the same setting for all your images, lights, darks, flats, and bias (if used). The offset value used is to avoid getting any image values at zero which means clipped data. If different offset values are used then you can't calibrate the data correctly. Alan
  5. Thanks David. I didn't notice the little platesolve and autofocus icons at the top right of the imaging tab. That certainly helps. 😀 Having a separate 'slew to target' button would be a convenient option for me to save having to go by CdC. The 'Framing' tab has a slew option but the coords need filling in beforehand as they are not necessarily related to the current target. Alan
  6. I've been looking at NINA for my Dual ZS-61 rig as it has synchronized dither while SGP still doesn't. It looks impressive, with comprehensive features but I can't see how you slew and centre on a target, or run auto-focus, without them being part of a sequence. Using SGP I normally load a sequence and then select slew, centre and auto-focus, before running the sequence. I tend to stick to one target for a session and if I stop and start the sequence for any reason I don't want it re-centreing or focusing each time. In NINA I can run a sequence with those options enabled and then stop the sequence, disable them and then restart the sequence, but it would be nice if I didn't have to do that each time. Alan
  7. Thanks for the replies and confirmation. 🙂 I'll have a go at soldering it on today. Alan
  8. Making my first surface mount PCB and the TSOT23-8 IC has these markings only. There are no beveled edges or cut-outs/indents to indicate pin 1 and the manufacturer states a dot is used. As you can see it's tiny 😬, with 0.65mm pin spacing and I can only assume the arrow indicates pin 1. Web searches haven't helped, so before I attempt to hand solder it on the board it would be nice to know for sure. If anyone can confirm it's pin 1 it would be helpful. 😀 It's a LTC4365, overvoltage, undervoltage and reverse supply protection controller and thought it's worth checking it out as protection for an imaging setup against faulty supplies which could be costly. It's only available in this package size. Alan
  9. Clutching at straws, but try unplugging your Lynx Astro cable from the PC and in Device Manager select 'View/Show Hidden Devices' and expand the Ports section. There will probably be some greyed out entries corresponding to allocations from previous devices that were plugged in, including the Lynx Astro Cable. Right mouse select the greyed out entries one by one and select 'Uninstall Device' and they should disappear. Then plug the Lynx Astro cable back in. It should re-install itself and still be allocated to COM3. When devices are unplugged the COM ports they used should be cleanly de-allocated, but that isn't always the case which is a reason the number of 'hidden devices' can grow and you get 'COM port in use or unavailable' messages even when they should be free. Perhaps a previous device on COM3 didn't clean itself up properly. Alan
  10. I've not tried it but according to this CN topic it looks like you can. 🙂 Alan
  11. With the Star Adventurer getting good polar alignment is important. If you have to polar align before mounting the scope and camera you can easily knock it off alignment while you're fitting the scope and camera. A big benefit of having a guidescope is being able to use Sharpcap's polar alignment tool with the Star Adventurer, which makes PA very easy, and can be done once everything is fitted to the mount. If your PA is good then Dec drift is not such a problem, though there will still be periodic error in the mount which could cause an issue using the 250mm Redcat with longer exposures. Guiding would also remove this periodic error problem. I use a Zwo Mini Finder with ASI120 on my dual ZS-61 scopes on an HEQ5 and it works perfectly with Sharpcap's PA tool. It will mount on the Redcat's finder shoe bracket easily. A further benefit of a guidescope is you can manually dither if you wish by looking at the PHD2 image and see how far the star actually moves for each dither, rather than guessing how much to move the mount Dec and RA controls. Can you not run 2 instances of PHD2 on the laptop to control both mounts? Alan
  12. Hi Graham, Happy to help. If you select the B.Cu layer before taking your screen shot it'll show the green tracks on top of the red. Alan PS. It's Gerber
  13. Rather than using a brute force approach to disable the power supply I would personably go for a more subtle approach of a separate unit on the PSU output which just cuts its output if the input voltage range exceeds preset limits. Something based on the LTC4365 or similar. With suitable mosfets the output current can then be higher than the 3A circuit shown in the examples. There are also modules available which use a relay to disable the output like this or perhaps this which may be worth looking into. Switch mode PSUs normally have over-voltage protection built in. The Mean Well ones certainly do. Alan
  14. The OPs power supplies are linear supplies using large transformers. The transformer has to supply sufficient voltage such that at maximum current draw the minimum voltage on the unregulated dc reservoir capacitors doesn't fall below the regulated output voltage plus the dropout voltage of the regulator stage, so an off load unregulated dc of 28V is fairly typical. On a basic regulated DC supply the only thing keeping the unregulated DC from the output are the transistors mounted on the large heat sinks on rear of the power supply. These transistors, along with their control circuitry, form a series regulator whose effective resistance varies depending on the current drawn to maintain the constant output voltage. These transistors dissipate the excess power due to the difference between the unregulated voltage and the stable output voltage. If one of them goes short circuit, often due to running too hot for too long, then the scenario experienced by the OP occurs. More expensive power supplies will incorporate over-voltage protection, current limiting etc. but there's no guarantee that the protection circuitry won't fail too. I too use Meanwell SMPSUs from Farnell. The original 100W versions I bought had been bench tested at full current by the manufacturer for 100 hours prior to sale and were about £45 each. The latest ones I bought were identical but hadn't been bench tested for 100 hours, and were £20 cheaper. I've never had any issues with them. Alan
  15. If you had noiseless images, (or very low noise) then the 16 bit image would be better. As soon as you stack a few images containing a "reasonable" amount of inherent noise then the inherent noise quickly becomes the dominant factor and overwhelms the quantization error 'noise'. From Craig Stark article for his test camera which had a noise Standard Deviation of 1.87, "The quantization noise in a 10-length stack of 8-bit images is smaller than the inherent noise in a 200-length stack of 16-bit images" From his graphs, as long as the Standard Deviation of the inherent noise is greater than 0.4 the inherent noise dominates the quantization noise for a stack of at least 50 frames when comparing 8 or 16 bit images. I suppose it's worth measuring the noise of your camera of a dark frame at your usual gain and exposure, for 8 and 16 bit and see what the Standard Deviation is. That would indicate whether it's worth using 16 bit images when stacking. 😉 Alan
  16. I've not found any difference in stacked 8 or 12 bit videos. Recording 12 bit just reduces your frame rate for full frame recordings as it is output from the camera as 16 bit, and also takes up twice the storage space. Alan
  17. If you used a USB2 cable that would account for a low frame rate. 4GB of ram is probably fine for a video of around 10s at 60fps before the frame rate dropped off. I've used an 4GB N3060 Celeron mini PC with SSD for Firecapture recording full frame for 30s or more and not had any problems. Alan
  18. You need to give a bit more information as to your setup, what spec PC you're using, ram available, USB3 cable length and/or extension and a reply to the question in the first paragraph I asked. Alan
  19. Does the Status actual FPS show 60 then change to 15FPS a few seconds or so after starting recording? This indicates you've run out of RAM and is having to store images directly to the disk resulting in a FPS drop. Does the Buffer bar drop right down and change to Red meaning no RAM left. If so, installing a SSD is the best option as writing to that is similar to the write to rwm speed. If there is still plenty of buffer space showing and the actual indicated FPS drops right down then there is a hardware holdup. Are you using a short USB3 cable, less than 3m, to enable the fastest data rates. If you're using an active USB3 extension this will seriously limit your data throughput. Ensure you don't have 16 bit enabled. 16 bit doubles your data throughput which can cause a slow down. Untick (disable) Gamma, or leave it at 50, as gamma processing each frame in camera takes significant extra time. Under Control/More check the USBTraffic setting and increase it if you can. If set too high it will cause you to lose frames or get no camera output. Disable the histogram during recording can gain some more FPS. Under Misc/Performance settings select SingleCore in Recommended Settings to reduce the amount of resources spent on the preview display. I'm assuming your Exposure time is less than 15mS as this is the longest exposure you can use to get 60 FPS. If your exposure is around 60mS then 15 FPS is the most you can achieve. Alan
  20. For lunar images stacking a series of DSLR stills would be likely to give a better image more easily than video recording. It's certainly worth trying for planets. It'll give you an idea of what can be achieved. Jupiter's rotation really only starts to be noticeable on high resolution images after a few minutes. For your size images it won't really cause any issues. If your image is good enough that only the planets rotation is causing it to look a bit blurry I think you'd be very happy. 😁 Alan
  21. Your uEye camera doesn't look like it has a ROI (Region of Interest) selection which would normally enable higher framerates. It's worth trying your DSLR. It's larger pixels means the on screen image will be less than half the size for a single frame, but should therefore give a brighter image as more planet area is covering a pixel Unfortunately, many DSLRs downscale the full frame image to video resolutions leading to a lower quality smaller video image. The drawback with DSLR videos is that they are lossy compressed, mp4 or similar, so stacking can't do such a good job as using uncompressed videos but should still give a reasonable image. If your DSLR can do centre cut-out videos where the central 1920 x 1080 pixels (or whatever the video resolution is) are used for the video instead of downscaling the full frame, you'll get a larger, higher quality image. Canon camera users can install the 'Magic Lantern' firmware which enables centre cut-out video recording. Alan
  22. I focus on a nearby star with a bahtinov mask. You may have to up the camera gain and exposure a lot to get 'useable 'spikes'. For the planet itself you want as low an exposure duration as you can manage to try and 'freeze' the seeing. 5mS is a good starting exposure. You may then have to put a lot of gain in the camera to get a bright enough image and it looks like a mass of noise on the preview, but once the frames are stacked the noise is 'magically' reduced a great deal. Take several videos at different exposures and gains and see which gives the best stacked image. A high frame rate just means you can get your 1000 frames or more quicker. It doesn't by itself make the image better. It just means there's less a chance of clouds etc. upsetting the video. With Jupiter however, too long a video, over several minutes, will lead to image blurring due to planet rotation. Alan
  23. I assume when you say your min ADU you mean the average sky background ADU which is close to the image median value. The read noise at unity gain is nearer 1.75 e- rather than the 1.55 e- you used to get 7.2075. This makes 3*RN^2 around 9.2 e-. Adding offset 21 and multiplying the result by 16 to convert from 12 to 16 bit gives a sky background ADU of 483. 3*RN^2 is 483 ADU, 5*RN^2 is 581 ADU and 10*RN^2 is 826 ADU. I tend to use the 10*RN^2 figures myself but it's just a personal preference. I've tried 15m and 20m exposures but for some reason I run the chance of slightly eggy stars over 10m so I tend to limit it to 10m. It's an AZ-EQ6 which I thought had a worm period of 8m. Alan
  24. Yes, amp glow is very noticeable at those exposures but a set of matching duration and temperature darks will take care of it. Any exposure over around 30s will have amp glow creeping in. It takes a while building a set of darks at those durations but once done they will be good to use for a long time. A new set of darks can be taken every year or so if you wish. Alan
×
×
  • Create New...

Important Information

We have placed cookies on your device to help make this website better. You can adjust your cookie settings, otherwise we'll assume you're okay to continue. By using this site, you agree to our Terms of Use.