IPB

Welcome Guest ( Log In | Register )

Perseverance Imagery, technical discussion of processing, cameras, etc.
Greenish
post Feb 22 2021, 02:26 AM
Post #1


Member
***

Group: Members
Posts: 219
Joined: 14-November 11
From: Washington, DC
Member No.: 6237



Putting this here, for reference as the payload of the JSON feed link for Perseverence raws (see source in other thread here)
https://mars.nasa.gov/rss/api/?feed=raw_ima...;&extended=
Looks like there's a ton of good data in addition to just the (PNG! Bayer color separated!) images biggrin.gif

This is for the first image shown on the page at the moment.
CODE
"images":
    [
        {
        "extended":
            {
            "mastAz":"UNK",
            "mastEl":"UNK",
            "sclk":"667129493.453",
            "scaleFactor":"4",
            "xyz":"(0.0,0.0,0.0)",
            "subframeRect":"(1,1,1280,960)",
            "dimension":"(1280,960)"
            },
        "sol":2,
        "attitude":"(0.415617,-0.00408664,-0.00947025,0.909481)",
        "image_files":
            {
            "medium":"https://mars.nasa.gov/mars2020-raw-images/pub/ods/surface/sol/00002/ids/edr/browse/rcam/RRB_0002_0667129492_604ECM_N0010052AUT_04096_00_2I3J01_800.jpg",
            "small":"https://mars.nasa.gov/mars2020-raw-images/pub/ods/surface/sol/00002/ids/edr/browse/rcam/RRB_0002_0667129492_604ECM_N0010052AUT_04096_00_2I3J01_320.jpg",
            "full_res":"https://mars.nasa.gov/mars2020-raw-images/pub/ods/surface/sol/00002/ids/edr/browse/rcam/RRB_0002_0667129492_604ECM_N0010052AUT_04096_00_2I3J01.png",
            "large":"https://mars.nasa.gov/mars2020-raw-images/pub/ods/surface/sol/00002/ids/edr/browse/rcam/RRB_0002_0667129492_604ECM_N0010052AUT_04096_00_2I3J01_1200.jpg"
            },
        "imageid":"RRB_0002_0667129492_604ECM_N0010052AUT_04096_00_2I3J01",
        "camera":
            {
            "filter_name":"UNK",
            "camera_vector":"(-0.7838279435884001,0.600143487448691,0.15950407306054173)",
            "camera_model_component_list":"2.0;0.0;(46.176,2.97867,720.521);(-0.701049,0.00940617,0.713051);(8.39e-06,0.0168764,-0.00743155);(-0.00878744,-0.00869157,-0.00676256);(-1.05782,-0.466472,-0.724517);(-0.702572,0.0113481,0.711523);(-448.981,-528.002,453.359)",
            "camera_position":"(-1.05782,-0.466472,-0.724517)",
            "instrument":"REAR_HAZCAM_RIGHT",
            "camera_model_type":"CAHVORE"
            },
        "caption":"NASA's Mars Perseverance rover acquired this image of the area in back of it using its onboard Rear Right Hazard Avoidance Camera. \n\n This image was acquired on Feb. 21, 2021 (Sol 2) at the local mean solar time of 15:37:11.",
        "sample_type":"Full",
        "date_taken_mars":"Sol-00002M15:37:11.985",
        "credit":"NASA/JPL-Caltech",
        "date_taken_utc":"2021-02-21T02:16:26Z",
        "json_link":"https://mars.nasa.gov/rss/api/?feed=raw_images&category=mars2020&feedtype=json&id=RRB_0002_0667129492_604ECM_N0010052AUT_04096_00_2I3J01",
        "link":"https://mars.nasa.gov/mars2020/multimedia/raw-images/?id=RRB_0002_0667129492_604ECM_N0010052AUT_04096_00_2I3J01",
        "drive":"52",
        "title":"Mars Perseverance Sol 2: Rear Right Hazard Avoidance Camera (Hazcam)",
        "site":1,
        "date_received":"2021-02-21T23:12:58Z"
        },


(with syntax color):
Attached Image


Here's hoping that one of you skilled characters can make good use...
Go to the top of the page
 
+Quote Post
12 Pages V   1 2 3 > »   
Start new topic
Replies (1 - 99)
phase4
post Feb 23 2021, 12:17 PM
Post #2


Junior Member
**

Group: Members
Posts: 95
Joined: 11-January 07
From: Amsterdam
Member No.: 1584



A small Perseverance update for Marslife, implementing the camera pointing data is a bit troublesome.
It suffers from a hit-and-miss accuracy. biggrin.gif

Attached Image


--------------------
Go to the top of the page
 
+Quote Post
Greenish
post Feb 23 2021, 02:40 PM
Post #3


Member
***

Group: Members
Posts: 219
Joined: 14-November 11
From: Washington, DC
Member No.: 6237



So cool!

Maybe they're still refining which way is up... lots of local coordinate frames to sort out and probably need some time (and the sun shots it looks like they just took, and some 3d processing of the local area, and some radio ranging) to get absolute position & orientation info.

The instrument position details will certainly be different vs MSL. And in my skim of the data it looked to me like they re-jiggered some of the pointing data formats, did you see that too? Not sure which parts you're using.
Go to the top of the page
 
+Quote Post
djellison
post Feb 23 2021, 08:39 PM
Post #4


Founder
****

Group: Chairman
Posts: 14432
Joined: 8-February 04
Member No.: 1



For the less technically minded among us...Ryan Kinnet has put up a page that grabs a listing with links to the PNG files that you can then use any browser-plugin-batch-downloader with

https://twitter.com/rover_18/status/1364309922167488512

I tried Firefox with 'DownloadThemAll' and it worked perfectly.

Meanwhile THIS GUY has python code to also grab the data
https://twitter.com/kevinmgill/status/1364311336000258048




Go to the top of the page
 
+Quote Post
ugordan
post Feb 23 2021, 08:45 PM
Post #5


Senior Member
****

Group: Members
Posts: 3648
Joined: 1-October 05
From: Croatia
Member No.: 523



Speaking of raw images, are these weird colorations due to something wrong with the pipeline or the cameras themselves (I'm thinking and hoping it's the former)?
Examples attached, one is a NavCam-L from yesterday (seems to have been pulled since), the other an RDC frame.
Attached thumbnail(s)
Attached Image
Attached Image
 


--------------------
Go to the top of the page
 
+Quote Post
mcaplinger
post Feb 23 2021, 08:56 PM
Post #6


Senior Member
****

Group: Members
Posts: 2517
Joined: 13-September 05
Member No.: 497



QUOTE (ugordan @ Feb 23 2021, 12:45 PM) *
Speaking of raw images, are these weird colorations due to something wrong with the pipeline...

Yes.


--------------------
Disclaimer: This post is based on public information only. Any opinions are my own.
Go to the top of the page
 
+Quote Post
ugordan
post Feb 23 2021, 09:01 PM
Post #7


Senior Member
****

Group: Members
Posts: 3648
Joined: 1-October 05
From: Croatia
Member No.: 523



QUOTE (mcaplinger @ Feb 23 2021, 09:56 PM) *
Yes.

That's reassuring to hear. One could sort of expect this from (well, almost) off-the shelf commercial EDL cameras, but the navcams are on a whole other level.

QUOTE (Steve G @ Feb 23 2021, 09:57 PM) *
New raw images just trickling in.

Yep, looks like they started download on the full quality, Bayered RDC camera views.

On a side note, I found it interesting how the color balance between RDC and DDC was so different. Yes, different detector resolutions, but same vendor.


--------------------
Go to the top of the page
 
+Quote Post
phase4
post Feb 23 2021, 09:35 PM
Post #8


Junior Member
**

Group: Members
Posts: 95
Joined: 11-January 07
From: Amsterdam
Member No.: 1584



Yes, I hope this erronous positioning will improve when the rover is calibrated properly. Otherwise it is my code to blame... biggrin.gif
I use the rover attitude, mastAz and mastEl values for camerapointing, I don't know if that has changed much since Curiosity.


--------------------
Go to the top of the page
 
+Quote Post
nprev
post Feb 24 2021, 01:30 AM
Post #9


Merciless Robot
****

Group: Admin
Posts: 8784
Joined: 8-December 05
From: Los Angeles
Member No.: 602



Hey, all. This thread is for the imagewizards among us & will focus on the abundance of data Mars 2020 will provide. Please post your products, methods, and tips, and use this thread to share information & likewise learn from others. smile.gif


--------------------
A few will take this knowledge and use this power of a dream realized as a force for change, an impetus for further discovery to make less ancient dreams real.
Go to the top of the page
 
+Quote Post
bdshort
post Feb 24 2021, 05:01 AM
Post #10


Newbie
*

Group: Members
Posts: 6
Joined: 6-August 12
Member No.: 6475



Okay, I hope this is the right forum for this question - one of the things I really like about both rovers having the ability to take stereo imagery is the ability to see stuff in 3D. Is there a utility out there that lets us view images in 3D using a VR headset? I've briefly searched around but haven't seen much. Anaglyphs are cool and everything but it leaves a bit to be desired.
Go to the top of the page
 
+Quote Post
Cargo Cult
post Feb 24 2021, 01:44 PM
Post #11


Junior Member
**

Group: Members
Posts: 36
Joined: 28-May 08
Member No.: 4152



QUOTE (bdshort @ Feb 24 2021, 07:01 AM) *
Okay, I hope this is the right forum for this question - one of the things I really like about both rovers having the ability to take stereo imagery is the ability to see stuff in 3D. Is there a utility out there that lets us view images in 3D using a VR headset? I've briefly searched around but haven't seen much. Anaglyphs are cool and everything but it leaves a bit to be desired.

Stereo imagery viewed in a VR headset is a bit underwhelming - reconstructing geometry using photogrammetry to create a fully 3D representation of a landscape is much more interesting.

I did this with some Curiosity imagery a few years ago, with fascinating results. If you have a SteamVR capable VR headset, you can have a look here: https://steamcommunity.com/sharedfiles/file...s/?id=928142301 - I typed up some fairly detailed notes in the description which will broadly apply to Perseverance imagery.

(Full disclosure: I work for Valve, creators of SteamVR. The Mars stuff was a fun personal project which turned into something a bit larger...)

Getting right back on topic for this thread - some notes on photogrammetry involving Perseverance imagery!

Some decent camera parameters* to start with in Agisoft Metashape (formerly PhotoScan):

Navcam
Camera type: Fisheye
Pixel size (mm): 0.0255 x 0.0255 (for 1280x960 images)
Focal length (mm) 19.1

Hazcam
Camera type: Fisheye
Pixel size (mm): 0.0255 x 0.0255 (for 1280x960 images)
Focal length (mm) 14

Using separate calibration profiles for left and right cameras may make sense - stuff worked better for Curiosity's navcams when I did this. (They're beautifully hand-made one-off scientific instruments, after all.) Metashape will further refine camera parameters once given these reasonable starting points.

It's all looking like really exciting data to play around with - navcam imagery in high-resolution, full colour after a bit of processing. I'm starting to figure out debayering stuff - I'm sure that this thread will come in great use!



* derived from The Mars 2020 Engineering Cameras and Microphone on the Perseverance Rover: A Next-Generation Imaging System for Mars Exploration
: Table 2, Perseverance Navcam, Hazcam, and Cachecam characteristics
Go to the top of the page
 
+Quote Post
bdshort
post Feb 24 2021, 04:22 PM
Post #12


Newbie
*

Group: Members
Posts: 6
Joined: 6-August 12
Member No.: 6475



Great! Thanks for the answer. I have an Index, so I'll check that out!

Edit: Just tried it - that's awesome! Gives a great sense of scale to the rover, and I liked all the point of interest spots. I'm surprised there isn't more stuff out there like this, it would be a fantastic educational tool and it's just fun to stand or sit around in for awhile.

QUOTE (Cargo Cult @ Feb 24 2021, 04:44 AM) *
Stereo imagery viewed in a VR headset is a bit underwhelming - reconstructing geometry using photogrammetry to create a fully 3D representation of a landscape is much more interesting.

I did this with some Curiosity imagery a few years ago, with fascinating results. If you have a SteamVR capable VR headset, you can have a look here: https://steamcommunity.com/sharedfiles/file...s/?id=928142301 - I typed up some fairly detailed notes in the description which will broadly apply to Perseverance imagery.

(Full disclosure: I work for Valve, creators of SteamVR. The Mars stuff was a fun personal project which turned into something a bit larger...)

Getting right back on topic for this thread - some notes on photogrammetry involving Perseverance imagery!

Some decent camera parameters* to start with in Agisoft Metashape (formerly PhotoScan):

Navcam
Camera type: Fisheye
Pixel size (mm): 0.0255 x 0.0255 (for 1280x960 images)
Focal length (mm) 19.1

Hazcam
Camera type: Fisheye
Pixel size (mm): 0.0255 x 0.0255 (for 1280x960 images)
Focal length (mm) 14

Using separate calibration profiles for left and right cameras may make sense - stuff worked better for Curiosity's navcams when I did this. (They're beautifully hand-made one-off scientific instruments, after all.) Metashape will further refine camera parameters once given these reasonable starting points.

It's all looking like really exciting data to play around with - navcam imagery in high-resolution, full colour after a bit of processing. I'm starting to figure out debayering stuff - I'm sure that this thread will come in great use!



* derived from The Mars 2020 Engineering Cameras and Microphone on the Perseverance Rover: A Next-Generation Imaging System for Mars Exploration
: Table 2, Perseverance Navcam, Hazcam, and Cachecam characteristics
Go to the top of the page
 
+Quote Post
scalbers
post Feb 24 2021, 11:41 PM
Post #13


Senior Member
****

Group: Members
Posts: 1639
Joined: 5-March 05
From: Boulder, CO
Member No.: 184



QUOTE
There are what i assume to be quaternions (?) regarding attitude in the search/query api. see https://mars.nasa.gov/rss/api/?feed=raw_ima...04096_034085J01 for example. There are altazimuth compass bearings and elevation for the regular people as well, which could be used to project the images to a compass ball according to the metadata, rather than manually overlapping and tie-pointing the images. I do not know of software or libraries that can help with this, if you do please tell.


Regarding the API "quaternion" coordinate info with the raw images, a LINUX program 'jq' is good for manipulating these. While this is just a starting point it would be interesting to see how an automated program might work with putting together a mosaic. I would just need some time to hook up some of my Fortran code with some scripts.

CargoCult - can your SteamVR creation be viewed somehow with a regular computer? With Curiosity, sittingduck (YouTube link below) had some really nice videos moving through a 3D landscape that might be interesting to see in VR. Sittingduck back in 2016 had used a Blender Plug-In obtained from phase4.

https://www.youtube.com/watch?v=7zW9yISB01Y...eature=youtu.be


--------------------
Steve [ my home page and planetary maps page ]
Go to the top of the page
 
+Quote Post
fredk
post Feb 25 2021, 01:13 AM
Post #14


Senior Member
****

Group: Members
Posts: 4247
Joined: 17-January 05
Member No.: 152



About image pointing, the algorithm I use for MSL, based on this post, seems to work fine for Percy, with the "rover_attitude" quaternion field replaced with "attitude". So for the "shiny rock" image:
https://mars.nasa.gov/mars2020-raw-images/p...6_034085J01.png
I get elevation,azimuth = 1.2, 251.5 degrees, which looks about right. That presumably is for the centre of the FOV, which happens to correspond almost exactly with the shiny rock.
Go to the top of the page
 
+Quote Post
djellison
post Feb 25 2021, 02:04 AM
Post #15


Founder
****

Group: Chairman
Posts: 14432
Joined: 8-February 04
Member No.: 1



It's not great - but this is my agisoft metashape results with the Navcam images so far
https://sketchfab.com/3d-models/m2020-landi...7215aa7db2fb0c8
Go to the top of the page
 
+Quote Post
scalbers
post Feb 25 2021, 06:28 PM
Post #16


Senior Member
****

Group: Members
Posts: 1639
Joined: 5-March 05
From: Boulder, CO
Member No.: 184



QUOTE (djellison @ Feb 25 2021, 02:04 AM) *
It's not great - but this is my agisoft metashape results with the Navcam images so far
https://sketchfab.com/3d-models/m2020-landi...7215aa7db2fb0c8


Looks pretty nice given the available vantage points (one rover location). Is there a possibility the "1st Person" navigation mode would work with this model?


--------------------
Steve [ my home page and planetary maps page ]
Go to the top of the page
 
+Quote Post
MarT
post Feb 25 2021, 08:11 PM
Post #17


Newbie
*

Group: Members
Posts: 13
Joined: 21-December 19
Member No.: 8729



QUOTE (Cargo Cult @ Feb 24 2021, 02:44 PM) *
Stereo imagery viewed in a VR headset is a bit underwhelming - reconstructing geometry using photogrammetry to create a fully 3D representation of a landscape is much more interesting.

I did this with some Curiosity imagery a few years ago, with fascinating results. If you have a SteamVR capable VR headset, you can have a look here: https://steamcommunity.com/sharedfiles/file...s/?id=928142301 - I typed up some fairly detailed notes in the description which will broadly apply to Perseverance imagery.

(Full disclosure: I work for Valve, creators of SteamVR. The Mars stuff was a fun personal project which turned into something a bit larger...)



Thats awsome! Great to see you here. I just ran into that two days ago. It is something I always wanted to do, but never fully found the time to work for a longer period of time. I have tried Photogrammetry on InSight, this is the result two years ago: https://www.youtube.com/watch?v=cBYAwTm_ArE...eature=youtu.be

Big thanks for the info here an on steam! It will be surely very helpfull to others too.

To stay on topic: I had a run with a short sequence of the true raw data from the down-look cams: https://www.youtube.com/watch?v=l4WKIoTjE4c...p;pbjreload=101
I will just wait for all the EDL data to download and will have a run with the data in Agisoft. I tried with the MARDI cam. While it was capable of clearly registering the images, I must have done something wrong, as the last image ended up "under the ground". Anyway the Perseverance data surely look promising in that matter.

I am working on a stabilized 360 video as with the MARDI cam. This is a quick version of the underlaying layer simulating approximately the view from 10km above the surface. I made it from the debayered data. Propably impacted by the FFMPEG compression.
Attached thumbnail(s)
Attached Image
 
Go to the top of the page
 
+Quote Post
phase4
post Feb 25 2021, 09:00 PM
Post #18


Junior Member
**

Group: Members
Posts: 95
Joined: 11-January 07
From: Amsterdam
Member No.: 1584



QUOTE (fredk @ Feb 25 2021, 02:13 AM) *
About image pointing, the algorithm I use for MSL, based on this post, seems to work fine for Percy, with the "rover_attitude" quaternion field replaced with "attitude".

Thank you for this hint, fredk. I didn't realize the Spice toolkit works just as fine without kernels.
For future use, does anyone know how to obtain the zoom value from the JSON information? Should it be derived from the CAHVOR data?

Attached Image


--------------------
Go to the top of the page
 
+Quote Post
mcaplinger
post Feb 25 2021, 09:56 PM
Post #19


Senior Member
****

Group: Members
Posts: 2517
Joined: 13-September 05
Member No.: 497



QUOTE (phase4 @ Feb 25 2021, 01:00 PM) *
For future use, does anyone know how to obtain the zoom value from the JSON information? Should it be derived from the CAHVOR data?

Should be if the CAHVOR model is set correctly, which they may or may not be at this point. See https://github.com/bvnayak/CAHVOR_camera_model and https://agupubs.onlinelibrary.wiley.com/doi...29/2003JE002199


--------------------
Disclaimer: This post is based on public information only. Any opinions are my own.
Go to the top of the page
 
+Quote Post
John Whitehead
post Feb 25 2021, 11:03 PM
Post #20


Junior Member
**

Group: Members
Posts: 99
Joined: 17-September 07
Member No.: 3901



Do color calibration targets ever fade (change color) from solar (and worse) radiation on Mars?
How are they tested on Earth and proven to not fade?
Go to the top of the page
 
+Quote Post
mcaplinger
post Feb 25 2021, 11:14 PM
Post #21


Senior Member
****

Group: Members
Posts: 2517
Joined: 13-September 05
Member No.: 497



QUOTE (John Whitehead @ Feb 25 2021, 03:03 PM) *
Do color calibration targets ever fade (change color) from solar (and worse) radiation on Mars?
How are they tested on Earth and proven to not fade?

https://mastcamz.asu.edu/mars-in-full-color/
QUOTE
At the University of Winnipeg, the effect of intense Mars-like ultraviolet (UV) light on the colors of the eight materials was studied, confirming that the materials will only change very little with UV-exposure through a long mission at the Martian surface.

Typically, they get dusty before fading would be an issue. Hopefully these magnets work better than the last time.


--------------------
Disclaimer: This post is based on public information only. Any opinions are my own.
Go to the top of the page
 
+Quote Post
JohnVV
post Feb 26 2021, 12:56 AM
Post #22


Member
***

Group: Members
Posts: 890
Joined: 18-November 08
Member No.: 4489



just came across a paper on the cameras and mic

"The Mars 2020 Engineering Cameras and Microphone on the Perseverance Rover: A Next-Generation Imaging System for Mars Exploration"

https://link.springer.com/article/10.1007/s11214-020-00765-9
Go to the top of the page
 
+Quote Post
phase4
post Feb 26 2021, 12:32 PM
Post #23


Junior Member
**

Group: Members
Posts: 95
Joined: 11-January 07
From: Amsterdam
Member No.: 1584



QUOTE (mcaplinger @ Feb 25 2021, 10:56 PM) *
Should be if the CAHVOR model is set correctly, which they may or may not be at this point. See https://github.com/bvnayak/CAHVOR_camera_model and https://agupubs.onlinelibrary.wiley.com/doi...29/2003JE002199


Thanks for the links. I tried the code but it gave very low values and fluctuating results for the Mastcam-Z focal length.
Will try again when new images arrive.




--------------------
Go to the top of the page
 
+Quote Post
mcaplinger
post Feb 26 2021, 04:44 PM
Post #24


Senior Member
****

Group: Members
Posts: 2517
Joined: 13-September 05
Member No.: 497



QUOTE (phase4 @ Feb 26 2021, 04:32 AM) *
Thanks for the links. I tried the code but it gave very low values and fluctuating results for the Mastcam-Z focal length.

Seems like, for the one case I looked at, that the values in the JSON are in the order VORCAH instead of CAHVOR as one would expect. I don't know if this is intentional or a bug. And I'm not sure if these models are actually correct anyway. But you could take a look.


--------------------
Disclaimer: This post is based on public information only. Any opinions are my own.
Go to the top of the page
 
+Quote Post
djellison
post Feb 26 2021, 05:30 PM
Post #25


Founder
****

Group: Chairman
Posts: 14432
Joined: 8-February 04
Member No.: 1



Jim has put out a Mastcam-Z filename decode guide

https://mastcamz.asu.edu/decoding-the-raw-p...mage-filenames/

It includes digits in there that describe the focal length in mm as three digits.

All the stereo pan images report 34mm for that - which matches the widest angle of Mastcam-Z MSL Mastcam Left.
Go to the top of the page
 
+Quote Post
mcaplinger
post Feb 26 2021, 05:35 PM
Post #26


Senior Member
****

Group: Members
Posts: 2517
Joined: 13-September 05
Member No.: 497



QUOTE (djellison @ Feb 26 2021, 09:30 AM) *
It includes digits in there that describe the focal length in mm as three digits.

Oh yeah, duh. Well, it was much more interesting to extract it from the camera model unsure.gif

FYI, widest angle of MCZ is 26mm.


--------------------
Disclaimer: This post is based on public information only. Any opinions are my own.
Go to the top of the page
 
+Quote Post
djellison
post Feb 26 2021, 05:45 PM
Post #27


Founder
****

Group: Chairman
Posts: 14432
Joined: 8-February 04
Member No.: 1



Ahh ok - they miss spoke at yesterdays thing then. This matches MSL Mastcam Left then smile.gif
Go to the top of the page
 
+Quote Post
mcaplinger
post Feb 26 2021, 06:15 PM
Post #28


Senior Member
****

Group: Members
Posts: 2517
Joined: 13-September 05
Member No.: 497



QUOTE (djellison @ Feb 26 2021, 09:45 AM) *
This matches MSL Mastcam Left then smile.gif

Correct. Original MCZ spec was for 34mm to 100mm but we ended up with some extra credit range and can do 26mm to 110mm. Typically values 26, 34, 48, 63, 79, 100, and 110 will be used, but AFAIK nobody is sure yet how they'll be chosen. Zoom has to be stowed at 26mm for driving, so that might motivate use of that setting sometimes.


--------------------
Disclaimer: This post is based on public information only. Any opinions are my own.
Go to the top of the page
 
+Quote Post
phase4
post Feb 26 2021, 06:51 PM
Post #29


Junior Member
**

Group: Members
Posts: 95
Joined: 11-January 07
From: Amsterdam
Member No.: 1584



QUOTE (mcaplinger @ Feb 26 2021, 05:44 PM) *
Seems like, for the one case I looked at, that the values in the JSON are in the order VORCAH instead of CAHVOR as one would expect.

So close. Processing the CAHVOR as VORCAH gave more plausible results, although one decimal off.
It was a fun exercise but no longer nescessary... thanks for posting the decode guide Doug!

Attached Image


--------------------
Go to the top of the page
 
+Quote Post
mcaplinger
post Feb 26 2021, 07:00 PM
Post #30


Senior Member
****

Group: Members
Posts: 2517
Joined: 13-September 05
Member No.: 497



QUOTE (phase4 @ Feb 26 2021, 10:51 AM) *
So close. Processing the CAHVOR as VORCAH gave more plausible results, although one decimal off.

Pixel size is 7.4e-6 m = 7.4e-3 mm, is that what you were using?


--------------------
Disclaimer: This post is based on public information only. Any opinions are my own.
Go to the top of the page
 
+Quote Post
phase4
post Feb 26 2021, 07:10 PM
Post #31


Junior Member
**

Group: Members
Posts: 95
Joined: 11-January 07
From: Amsterdam
Member No.: 1584



QUOTE (mcaplinger @ Feb 26 2021, 08:00 PM) *
Pixel size is 7.4e-6 m = 7.4e-3 mm, is that what you were using?


Ah thank you, I used 0.074 mm instead of 0.0074. Things are obviously ok now. biggrin.gif


--------------------
Go to the top of the page
 
+Quote Post
fredk
post Feb 26 2021, 07:21 PM
Post #32


Senior Member
****

Group: Members
Posts: 4247
Joined: 17-January 05
Member No.: 152



In a previous post I estimated the horizontal FOV to be 19.5 degrees for an image with a "34" focal length filename field. I said there that that was near the maximum FOV of 19.2 deg, but I mistakenly used vertical FOV, not horizontal. My measured 19.5 deg corresponds very closely to the max horizontal 25.6 scaled from 26 to 34 mm, so everything is consistent with a 34 mm focal length (apart from the statement yesterday as Doug mentioned).

And we seem to have a little bonus of around 1608 horizontal pixels, vs the 1600 photoactive stated in Bell etal.
Go to the top of the page
 
+Quote Post
MahFL
post Feb 27 2021, 11:19 PM
Post #33


Forum Contributor
****

Group: Members
Posts: 1372
Joined: 8-February 04
From: North East Florida, USA.
Member No.: 11



Is there a way to tell the zoom used from the MastCamZ file names ?
Go to the top of the page
 
+Quote Post
mcaplinger
post Feb 27 2021, 11:31 PM
Post #34


Senior Member
****

Group: Members
Posts: 2517
Joined: 13-September 05
Member No.: 497



QUOTE (MahFL @ Feb 27 2021, 03:19 PM) *
Is there a way to tell the zoom used from the MastCamZ file names ?

Yes. See post #25 in this thread.


--------------------
Disclaimer: This post is based on public information only. Any opinions are my own.
Go to the top of the page
 
+Quote Post
Greenish
post Mar 1 2021, 04:36 AM
Post #35


Member
***

Group: Members
Posts: 219
Joined: 14-November 11
From: Washington, DC
Member No.: 6237



It appears the JSON data for the Sol 9 images has corrected the CAHVORE formatting issue we saw in the earlier image metadata (i.e. in correct order).
I wonder if it was a different version effect from the cruise software load.
CODE
            "camera": {
                "filter_name": "UNK",
                "camera_vector": "(0.6929857012250769,-0.7042489328574726,-0.1542862873579498)",
                "camera_model_component_list": "(1.10353,-0.008945,-0.729124);(0.884935,-0.17181,0.432865);(2606.49,1675.44,1273.54);(747.819,-326.9,2757.19);(0.885293,-0.171227,0.432363);(1e-06,0.00889,-0.006754);(-0.006127,0.010389,0.004541);2.0;0.0",
                "camera_position": "(1.10353,-0.008945,-0.729124)",
                "instrument": "FRONT_HAZCAM_LEFT_A",
                "camera_model_type": "CAHVORE"


Go to the top of the page
 
+Quote Post
djellison
post Mar 1 2021, 04:43 AM
Post #36


Founder
****

Group: Chairman
Posts: 14432
Joined: 8-February 04
Member No.: 1



The front hazcams have quite a significant toe-out between them. ~20deg.
page 17
See https://www.ncbi.nlm.nih.gov/pmc/articles/P...Article_765.pdf
Go to the top of the page
 
+Quote Post
MarkL
post Mar 1 2021, 08:14 PM
Post #37


Member
***

Group: Members
Posts: 240
Joined: 18-July 06
Member No.: 981



QUOTE (JohnVV @ Feb 26 2021, 12:56 AM) *
just came across a paper on the cameras and mic



Thank you!
Go to the top of the page
 
+Quote Post
PDP8E
post Mar 1 2021, 11:07 PM
Post #38


Member
***

Group: Members
Posts: 808
Joined: 10-October 06
From: Maynard Mass USA
Member No.: 1241



The cameras used for rover lookdown and lookup are AMS CMV20000

here is the datasheet from AMS

https://ams.com/documents/20143/36005/CMV20...e1-428cb363ab0a

Maybe the lookup camera can be used for stargazing, cloud studies, tau....

My debayer program is still a little wonky green

(are there R and B multiplier 'factors' known for this camera?)

I downsized it to fit in 3MB here
Attached Image


--------------------
CLA CLL
Go to the top of the page
 
+Quote Post
Andreas Plesch
post Mar 2 2021, 03:19 AM
Post #39


Member
***

Group: Members
Posts: 248
Joined: 25-February 21
From: Waltham, Massachussetts, U.S.A.
Member No.: 8974



Since I believe this may not have been mentioned, it is useful to know that the rss API json feed allows for just returning a single record based on id. It has an id query parameter.
So one can browse to find an image, say

CODE
https://mars.nasa.gov/mars2020-raw-images/pub/ods/surface/sol/00002/ids/edr/browse/edl/EDF_0002_0667111022_758ECV_N0010052EDLC00002_0010LUJ01.png


and then use the id (filename without suffix) to get the record:

CODE
https://mars.nasa.gov/rss/api/?feed=raw_images&category=mars2020&feedtype=json&id=EDF_0002_0667111022_758ECV_N0010052EDLC00002_0010LUJ01

For example, this is interesting pointing out that the downlook rover camera does not have any interesting additional data.

[edit:] If the filename for the image has trailing digits as in https://mars.nasa.gov/mars2020-raw-images/p...2_06_0LLJ01.png, for some cameras (not sure which ones) it is necessary to omit the trailing digits (here "01") for the json id: https://mars.nasa.gov/rss/api/?feed=raw_ima...AZ00102_07_0LLJ

EDF_0002_0667111022_758ECV_N0010052EDLC00002_0010LUJ01 (for EDL) vs.
FLE_0009_0667754529_115ECM_N0030000FHAZ00102_07_0LLJ (for FHAZ)



--------------------
--
Andreas Plesch, andreasplesch at gmail dot com
Go to the top of the page
 
+Quote Post
Greenish
post Mar 2 2021, 03:23 AM
Post #40


Member
***

Group: Members
Posts: 219
Joined: 14-November 11
From: Washington, DC
Member No.: 6237



Good info from Emily Lakdawalla on twitter:
QUOTE
@elakdawalla 8:54 PM · Mar 1, 2021
- I just got off WebEx with Justin Maki, who leads of the Perseverance engineering camera team. I've learned a lot and gotten a lot of confused questions sorted out. I'll try to bang out a blog entry with lots of techy detail about raw images tomorrow.
- The TL;DR: of the interview was: a lot of the things that are weird and confusing in the raw image metadata from sols 1-4 have to do with the rover being on the cruise flight software at the time.
- For example, the cruise flight software did not "know" how to automatically create image thumbnails. So they had to instruct the rover computer with separate commands to make thumbnails for each image, which is why sequence IDs don't match up between thumbnail and full-res.
- Many of the more confusing issues were solved by the flight software update. They're going to continue to tweak parameters over the next week or two, testing to see what modes they like best for returning their data, but before long they'll settle into some routines.
- It's SO FUN to see this process working out in real time. They *could* hold all the images back until they're happy with their tweaking, but they're not. They're just shunting the images out, never mind the temporarily wonky metadata.
Go to the top of the page
 
+Quote Post
JohnVV
post Mar 2 2021, 03:26 AM
Post #41


Member
***

Group: Members
Posts: 890
Joined: 18-November 08
Member No.: 4489



QUOTE
My debayer program is still a little wonky green


have you tried the debayer in G'Mic
https://gmic.eu/
CODE
gmic input.png bayer2rgb -o output.png

Go to the top of the page
 
+Quote Post
fredk
post Mar 2 2021, 06:02 PM
Post #42


Senior Member
****

Group: Members
Posts: 4247
Joined: 17-January 05
Member No.: 152



Yeah, gmic gives the same wonky green/yellow cast. The particular deBayering interpolation algorithm shouldn't determine the overall hue (but may affect pixel-scale chroma details). What a deBayering by itself gives is known as "raw colour", and won't generally look right because the relative sensitivities of the RGB channels differ from those of the eye. DeBayered images released have similar casts, eg:
https://mars.nasa.gov/mars2020-raw-images/p...0_01_295J02.png

A simple relative scaling between RGB channels, ie a whitebalance, should help a lot with these images.
Go to the top of the page
 
+Quote Post
fredk
post Mar 2 2021, 07:48 PM
Post #43


Senior Member
****

Group: Members
Posts: 4247
Joined: 17-January 05
Member No.: 152



QUOTE (fredk @ Mar 2 2021, 07:02 PM) *
A simple relative scaling between RGB channels, ie a whitebalance, should help a lot with these images.

If the black level isn't maintained during the autostretch which is done on the public engineering cam frames, then such a simple fixed RGB scaling won't work for all frames and we're left with trial and error. So it could be that different tiles of a full frame navcam would need different colour adjustment.
Go to the top of the page
 
+Quote Post
lingyuk
post Mar 3 2021, 07:58 AM
Post #44


Newbie
*

Group: Members
Posts: 3
Joined: 28-February 21
Member No.: 8978



So, as I understand it - Perseverance has the ability compress frames into .MP4 files? And the landing videos posted by NASA on youtube were uploads of those MP4 files? And the rover will later send all of the full resolution frames (1000s of them) of the landing?

My question is: are those MP4 files available anywhere to download? Has NASA made them available?

Because Youtube compresses videos a lot, and the original files would have a lot more detail. Thanks!
Go to the top of the page
 
+Quote Post
MarkL
post Mar 3 2021, 05:54 PM
Post #45


Member
***

Group: Members
Posts: 240
Joined: 18-July 06
Member No.: 981



QUOTE (lingyuk @ Mar 3 2021, 08:58 AM) *
My question is: are those MP4 files available anywhere to download? Has NASA made them available?

Great question. JPL ws able to get video footage very quickly and there are still a lot of individual frames to be published so these must have been videos created by the cameras and uplinked on Sol 1.

Where are the raw video files do you suppose?

Can we get our hands on them?
Go to the top of the page
 
+Quote Post
lingyuk
post Mar 4 2021, 10:57 AM
Post #46


Newbie
*

Group: Members
Posts: 3
Joined: 28-February 21
Member No.: 8978



QUOTE (MarkL @ Mar 3 2021, 05:54 PM) *
Great question. JPL ws able to get video footage very quickly and there are still a lot of individual frames to be published so these must have been videos created by the cameras and uplinked on Sol 1.

Where are the raw video files do you suppose?

Can we get our hands on them?


I found this file:

https://mars.nasa.gov/system/downloadable_i..._deployment.mp4

It looks like it's the original MP4 file. I couldn't find any others though.
Go to the top of the page
 
+Quote Post
fredk
post Mar 4 2021, 03:37 PM
Post #47


Senior Member
****

Group: Members
Posts: 4247
Joined: 17-January 05
Member No.: 152



That video is slowed down and has duplicated frames, so must've been re-encoded from the original.
Go to the top of the page
 
+Quote Post
Osvold
post Mar 8 2021, 04:05 PM
Post #48


Newbie
*

Group: Members
Posts: 2
Joined: 7-March 21
Member No.: 8986



Hello guys, I am new here and I have no expirience with pictures from Mars, so could be, this is obvious to you, but I struggle to find out how this works. I would like to know, if there is any way how to find out direction the rover is looking from a picture. As an example, i would love to know, if the hill
Attached Image
on the attached photo is the red encircled hill on the map
Attached Image
. How is possible to figure this out, apart form guessing from the map, please? Thanks a lot for any help.
Go to the top of the page
 
+Quote Post
john_s
post Mar 8 2021, 05:19 PM
Post #49


Member
***

Group: Members
Posts: 700
Joined: 3-December 04
From: Boulder, Colorado, USA
Member No.: 117



One easy way to get oriented is to look for Phil Stooke's circular projections which he posts regularly to support his mapping efforts. These always have north at the top, and show vertically-exaggerated images of features in the distance. The one linked here confirms your hunch about the identity of that mesa.
Go to the top of the page
 
+Quote Post
Pando
post Mar 8 2021, 08:56 PM
Post #50


The Insider
***

Group: Members
Posts: 669
Joined: 3-May 04
Member No.: 73



QUOTE (Osvold @ Mar 8 2021, 08:05 AM) *
i would love to know, if the hill on the attached photo is the red encircled hill on the map


Yes, it is.
Go to the top of the page
 
+Quote Post
Floyd
post Mar 9 2021, 01:48 AM
Post #51


Member
***

Group: Members
Posts: 910
Joined: 4-September 06
From: Boston
Member No.: 1102



The Raw page has several types of images and cameras. I did not see listed the SuperCam (Perseverence ChemCam). Will these images eventually find their way to the raw page?


--------------------
Go to the top of the page
 
+Quote Post
MahFL
post Mar 9 2021, 04:15 AM
Post #52


Forum Contributor
****

Group: Members
Posts: 1372
Joined: 8-February 04
From: North East Florida, USA.
Member No.: 11



QUOTE (Floyd @ Mar 9 2021, 01:48 AM) *
The Raw page has several types of images and cameras. I did not see listed the SuperCam (Perseverence ChemCam). Will these images eventually find their way to the raw page?


I would assume so, as the list has grown as images come down from the different cameras.
Go to the top of the page
 
+Quote Post
Osvold
post Mar 9 2021, 07:09 AM
Post #53


Newbie
*

Group: Members
Posts: 2
Joined: 7-March 21
Member No.: 8986



@john_s @Pando thanks a lot guys
Go to the top of the page
 
+Quote Post
MarkL
post Mar 9 2021, 09:57 AM
Post #54


Member
***

Group: Members
Posts: 240
Joined: 18-July 06
Member No.: 981



QUOTE (lingyuk @ Mar 4 2021, 10:57 AM) *
I found this file:

https://mars.nasa.gov/system/downloadable_i..._deployment.mp4

It looks like it's the original MP4 file. I couldn't find any others though.


Thanks for finding it. I hope they release all the raw videos that were uplinked from the rover on Sols 0-1. Please JPL. This is historic stuff being first use of a legendary open source utility on another planet. Ffmpeg rocks the (solar) system.
Go to the top of the page
 
+Quote Post
Explorer1
post Mar 9 2021, 02:48 PM
Post #55


Senior Member
****

Group: Members
Posts: 2086
Joined: 13-February 10
From: Ontario
Member No.: 5221



A silly clickbait video where someone zooms in on empty sky (https://www.youtube.com/watch?v=dQHYA0-Tfrk) got me thinking; how good would the Mastcam-Z be at astronomy? It would presumably be able to catch both moons if pointed at the right place and time (either during the day or at night, if power allows)?
I remember the Phobos-rise image taken early in Curiosity's mission ( http://www.midnightplanets.com/web/MSL/ima...58E01_DXXX.html ); Maztcam-Z could do a much better job now.
Go to the top of the page
 
+Quote Post
fredk
post Mar 9 2021, 03:48 PM
Post #56


Senior Member
****

Group: Members
Posts: 4247
Joined: 17-January 05
Member No.: 152



The MCZ specs are pretty similar to MSL mastcam, apart from the ability to zoom. The long end is 110mm at f/9.5 vs 100mm at f/10 for MR, and the sensor resolutions are very similar. So moon imaging should be very similar.

I thought the main advantages of MCZ were the ability to do stereo imaging with the same FOV in L and R cameras, as well as the intermediate zoom focal lengths.
Go to the top of the page
 
+Quote Post
mcaplinger
post Mar 9 2021, 04:48 PM
Post #57


Senior Member
****

Group: Members
Posts: 2517
Joined: 13-September 05
Member No.: 497



QUOTE (fredk @ Mar 9 2021, 07:48 AM) *
So moon imaging should be very similar.

Fred is correct, MCZ will not improve significantly on images like https://photojournal.jpl.nasa.gov/catalog/PIA17350


--------------------
Disclaimer: This post is based on public information only. Any opinions are my own.
Go to the top of the page
 
+Quote Post
tolis
post Mar 9 2021, 06:52 PM
Post #58


Member
***

Group: Members
Posts: 149
Joined: 18-June 08
Member No.: 4216



For astronomical imaging, there is also SkyCam.
Go to the top of the page
 
+Quote Post
djellison
post Mar 9 2021, 08:35 PM
Post #59


Founder
****

Group: Chairman
Posts: 14432
Joined: 8-February 04
Member No.: 1



The MEDA SkyCam is a repurposed MER/MSL heritage Hazcam. They are INCREDIBLY optically dark. We're tried ( and failed ) to image night time features with MSL NavCams - even Earth and Venus were not visible. You're not going to get much, if anything, astronomical, through the MEDA SkyCam
Go to the top of the page
 
+Quote Post
tolis
post Mar 9 2021, 09:18 PM
Post #60


Member
***

Group: Members
Posts: 149
Joined: 18-June 08
Member No.: 4216



Pity. It would have been a great edu resource.
Go to the top of the page
 
+Quote Post
PaulH51
post Mar 9 2021, 10:01 PM
Post #61


Senior Member
****

Group: Members
Posts: 2429
Joined: 30-January 13
From: Penang, Malaysia.
Member No.: 6853



Sol 18: Our first image from the SHERLOC/WATSON camera with its lens cap open

I did Bayer reconstruction of the raw image using GIMP with the add-on filter 'G'MIC-Qt'

The image data drop out was seen on some of the earlier images with the lens cap closed, but in a different place, so hopefully not a problem.

I noted that no focus motor count was issued on the raw page so it's not possible to establish the distance to the target or estimate the scale like we could do for MAHLI. Maybe that data will be in the JSON details?
Attached Image


Raw image
Go to the top of the page
 
+Quote Post
fredk
post Mar 9 2021, 10:15 PM
Post #62


Senior Member
****

Group: Members
Posts: 4247
Joined: 17-January 05
Member No.: 152



I see no motor count at the json details page for that image:
https://mars.nasa.gov/rss/api/?feed=raw_ima...LC08001_0000LUJ
The filter name field just says "open".
Go to the top of the page
 
+Quote Post
Andreas Plesch
post Mar 9 2021, 11:47 PM
Post #63


Member
***

Group: Members
Posts: 248
Joined: 25-February 21
From: Waltham, Massachussetts, U.S.A.
Member No.: 8974



Here is the full json in case something jumps out:
CODE
{
"image": [
{
"extended": {
"mastAz": "307.733",
"mastEl": "-20.2427",
"sclk": "668555815.086",
"scaleFactor": "1",
"xyz": "(39.3213,34.1703,-0.00444177)",
"subframeRect": "(1,1,1648,1200)",
"dimension": "(1648,1200)"
},
"link_related_sol": "https://mars.nasa.gov/mars2020/multimedia/raw-images/?sol=18",
"sol": 18,
"attitude": "(0.812911,-0.0114458,0.0116122,0.582159)",
"json_link_related_sol": "https://mars.nasa.gov/rss/api/?feed=raw_images&category=mars2020&feedtype=json&sol=18",
"image_files": {
"medium": "https://mars.nasa.gov/mars2020-raw-images/pub/ods/surface/sol/00018/ids/edr/browse/shrlc/SI0_0018_0668555815_008ECM_N0030578SRLC08001_0000LUJ02_800.jpg",
"small": "https://mars.nasa.gov/mars2020-raw-images/pub/ods/surface/sol/00018/ids/edr/browse/shrlc/SI0_0018_0668555815_008ECM_N0030578SRLC08001_0000LUJ02_320.jpg",
"full_res": "https://mars.nasa.gov/mars2020-raw-images/pub/ods/surface/sol/00018/ids/edr/browse/shrlc/SI0_0018_0668555815_008ECM_N0030578SRLC08001_0000LUJ02.png",
"large": "https://mars.nasa.gov/mars2020-raw-images/pub/ods/surface/sol/00018/ids/edr/browse/shrlc/SI0_0018_0668555815_008ECM_N0030578SRLC08001_0000LUJ02_1200.jpg"
},
"imageid": "SI0_0018_0668555815_008ECM_N0030578SRLC08001_0000LUJ",
"camera": {
"filter_name": "OPEN",
"camera_vector": "(0.5740789607042541,-0.7421743420005331,0.3458476441336543)",
"camera_model_component_list": "(2.08195,-0.111422,-1.69411);(0.577631,-0.741337,0.341707);(-1801.91,-2368.23,291.421);(946.46,-1240.2,-2497.62);(0.580824,-0.747997,0.321164);(0.000473,-0.032613,0.136778)",
"camera_position": "(2.08195,-0.111422,-1.69411)",
"instrument": "SHERLOC_WATSON",
"camera_model_type": "CAHVOR"
},
"caption": "NASA's Mars Perseverance rover acquired this image using its SHERLOC WATSON camera, located on the turret at the end of the rover's robotic arm. \n\nThis image was acquired on Mar. 9, 2021 (Sol 18) at the local mean solar time of 17:13:24.",
"sample_type": "Full",
"date_taken_mars": "Sol-00018M17:13:24.434",
"credit": "NASA/JPL-Caltech",
"date_taken_utc": "2021-03-09T17:56:16.000",
"link": "https://mars.nasa.gov/mars2020/multimedia/raw-images/?id=SI0_0018_0668555815_008ECM_N0030578SRLC08001_0000LUJ",
"link_related_camera": "https://mars.nasa.gov/mars2020/multimedia/raw-images/?camera=SHERLOC_WATSON&sol=18",
"drive": "578",
"title": "Mars Perseverance Sol 18: WATSON Camera ",
"site": 3,
"date_received": "2021-03-09T18:26:55Z"
}
],
"type": "mars2020-imagedetail-1.1",
"mission": "mars2020"
}


--------------------
--
Andreas Plesch, andreasplesch at gmail dot com
Go to the top of the page
 
+Quote Post
Greenish
post Mar 10 2021, 03:12 AM
Post #64


Member
***

Group: Members
Posts: 219
Joined: 14-November 11
From: Washington, DC
Member No.: 6237



If "camera_position" is correct at (2.08195,-0.111422,-1.69411) then WATSON's eyeball would be a fairly high 1.7m off the ground, and out ~1m in front of the body, just left (port) of centerline.
(For ref, in RNAV frame z=0 is ~ground level, positive down; hazcams are at x~ +1.1, z~ -0.7 and navcam at z~ -1.9.)

An FOV of 37 deg diag would give about 1.1m diagonal in the image if pointed straight down. However if the axis in CAHVOR is correct then it's pointed ~20 deg below horizontal (it works out to match the "mast" azimuth & elevation values). Not convinced either of these is right - or my math, for that matter.
Go to the top of the page
 
+Quote Post
lingyuk
post Mar 10 2021, 10:41 AM
Post #65


Newbie
*

Group: Members
Posts: 3
Joined: 28-February 21
Member No.: 8978



What is the current best method for downloading all of the raw png files from all cameras?
Go to the top of the page
 
+Quote Post
mcaplinger
post Mar 10 2021, 03:17 PM
Post #66


Senior Member
****

Group: Members
Posts: 2517
Joined: 13-September 05
Member No.: 497



QUOTE (Greenish @ Mar 9 2021, 07:12 PM) *
If "camera_position" is correct at (2.08195,-0.111422,-1.69411) then WATSON's eyeball would be a fairly high 1.7m off the ground...

That looks wrong to me, I suspect a bug somewhere.


--------------------
Disclaimer: This post is based on public information only. Any opinions are my own.
Go to the top of the page
 
+Quote Post
xflare
post Mar 10 2021, 05:23 PM
Post #67


Member
***

Group: Members
Posts: 282
Joined: 18-June 04
Member No.: 84



Ah I didnt realise we wouldn't be seeing SUPERCAM RAW images like we did with Curiosity
Go to the top of the page
 
+Quote Post
MarkL
post Mar 10 2021, 09:50 PM
Post #68


Member
***

Group: Members
Posts: 240
Joined: 18-July 06
Member No.: 981



QUOTE (lingyuk @ Mar 10 2021, 10:41 AM) *
What is the current best method for downloading all of the raw png files from all cameras?


https://rkinnett.github.io/roverpics/

Firefox/DownThemAll is a great combo and you can get the actual JSON files or just the URLs. Great service!
Go to the top of the page
 
+Quote Post
fredk
post Mar 10 2021, 11:50 PM
Post #69


Senior Member
****

Group: Members
Posts: 4247
Joined: 17-January 05
Member No.: 152



QUOTE (Greenish @ Mar 10 2021, 04:12 AM) *
if the axis in CAHVOR is correct then it's pointed ~20 deg below horizontal (it works out to match the "mast" azimuth & elevation values)

Using the camera_vector and attitude quaternion gives el,az = -19.3,19.5 deg in the LL frame (so az is relative to north), so the azimuth is quite different from the mastAz value. My guess is that mastEl and mastAz are for the RNAV frame (elevations are pretty close since the rover was close to level).
Go to the top of the page
 
+Quote Post
htspace
post Mar 11 2021, 08:02 AM
Post #70


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



How to batch debayer images?
What software is there? Fast and easy?
When I download these pictures, I don’t know how to batch process……
Thank you!
Attached Image
Go to the top of the page
 
+Quote Post
htspace
post Mar 11 2021, 08:06 AM
Post #71


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



Hello, there is no response when clicking Export selected image URLs. sad.gif
Go to the top of the page
 
+Quote Post
htspace
post Mar 11 2021, 08:07 AM
Post #72


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



QUOTE (MarkL @ Mar 11 2021, 05:50 AM) *
https://rkinnett.github.io/roverpics/

Firefox/DownThemAll is a great combo and you can get the actual JSON files or just the URLs. Great service!

Hello, there is no response when clicking Export selected image URLs. sad.gif
Go to the top of the page
 
+Quote Post
JohnVV
post Mar 11 2021, 01:07 PM
Post #73


Member
***

Group: Members
Posts: 890
Joined: 18-November 08
Member No.: 4489



QUOTE
What software is there? Fast and easy?


i use Gmic's built in program " bayer2rgb"
https://gmic.eu/
it is a commandline tool
CODE
gmic InputImage.png bayer2rgb 1,1,1 -o OutputImage.png

now they will be a bit "yellow/green" do to the contrast stretch that these png's have had done to them


Go to the top of the page
 
+Quote Post
JRehling
post Mar 11 2021, 05:22 PM
Post #74


Senior Member
****

Group: Members
Posts: 2530
Joined: 20-April 05
Member No.: 321



The goals of a panoramic color camera and astrophotography are in many ways diametrically opposed. (That said, that image of Phobos and Deimos is simply awesome.) For imaging Phobos, Earth, Jupiter, etc. from Mars and getting the most impressive results, you'd want an instrument very different from Mastcam-Z. And for obvious reasons, that's not what this mission is about.

To do astrophotography from Mars (where, I imagine, the "seeing" would almost always be excellent due to the low pressure) the ideal instrument would be the same as on Earth – a big aperture on the telescope and a small pixel count, just big enough to capture the object of interest. Lots of area means your system is capturing meaningless data when you're looking at a small object surrounded by empty sky. (Some camera software allows you to capture just a small window of a larger frame; I have one with that capability, but I don't think there's any reason why you'd want that on Mars.) Color via a monochrome sensor and the ability to move different filters in front of it… Bayer color automatically reduces your resolution by about 1.5.

It'd be fun to see what a ~20 cm aperture telescope could do from Mars, but I think we can imagine… take pictures of Jupiter as seen by amateurs on Earth and improve the resolution about 2x. It'd be amazing. Almost Voyager quality. Pictures of Earth would have about half the resolution of pictures of Mars as seen from Earth (because you'd only be looking at Earth's night side when the planets are closest).

But I don't see this sort of add-on justifying the cost anytime soon.
Go to the top of the page
 
+Quote Post
djellison
post Mar 11 2021, 06:36 PM
Post #75


Founder
****

Group: Chairman
Posts: 14432
Joined: 8-February 04
Member No.: 1



QUOTE (JohnVV @ Mar 11 2021, 05:07 AM) *
i use Gmic's built in program " bayer2rgb"


I must admit - I've found Gmic's debayer performance to be pretty ugly - especially at the edge of images.

I've been using PIPP which can batch debayer images and I've been very pleased with the results

https://sites.google.com/site/astropipp/
Go to the top of the page
 
+Quote Post
Ant103
post Mar 11 2021, 07:29 PM
Post #76


Senior Member
****

Group: Members
Posts: 1619
Joined: 12-February 06
From: Bergerac - FR
Member No.: 678



As for me, I'm using ImageJ with Debayer plugin. Alongside with a batch script, I can do hundred of pictures in a very short time.


--------------------
Go to the top of the page
 
+Quote Post
James Sorenson
post Mar 11 2021, 08:00 PM
Post #77


Member
***

Group: Members
Posts: 691
Joined: 21-December 07
From: Clatskanie, Oregon
Member No.: 3988



I'm using Pixinsight which is an astrophotography processing suite, it has a real nice debayer script. I use it for my astrophotography and it is expensive, 300 USD, but for those that can't afford that, the Planetary Image Pre-processor (PIPP) has a great batch debayer function in it which I've also used and it works great. On the plus side it's a free and easy program. smile.gif
Go to the top of the page
 
+Quote Post
phase4
post Mar 11 2021, 08:49 PM
Post #78


Junior Member
**

Group: Members
Posts: 95
Joined: 11-January 07
From: Amsterdam
Member No.: 1584



The Marslife website uses an open source GPU shader to debayer images 'on-the-fly' which is blazingly fast.
While not really useful for non-coders I thought to share this solution as a reference. Pretty well documented, too.

Efficient, high-quality Bayer demosaic filtering on GPUs by Morgan McGuire.


--------------------
Go to the top of the page
 
+Quote Post
fredk
post Mar 12 2021, 12:37 AM
Post #79


Senior Member
****

Group: Members
Posts: 4247
Joined: 17-January 05
Member No.: 152



QUOTE (phase4 @ Mar 11 2021, 09:49 PM) *
The Marslife website uses an open source GPU shader to debayer images 'on-the-fly' which is blazingly fast.

Thanks; I was wondering how you deBayered the mastcam sequences so fast.

Incidentally the Malvar etal interpolation scheme you use is the same as that used onboard MSL to deBayer when jpegs are sent down; see the MSSS camera specification. (And that's the scheme I use in my own code.)

Anyway, deBayering M20 frames is a breeze - dealing well with the jpeged Bayered images for MSL is hard work!
Go to the top of the page
 
+Quote Post
htspace
post Mar 12 2021, 01:32 AM
Post #80


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



QUOTE (JohnVV @ Mar 11 2021, 09:07 PM) *
i use Gmic's built in program " bayer2rgb"
https://gmic.eu/
it is a commandline tool
CODE
gmic InputImage.png bayer2rgb 1,1,1 -o OutputImage.png

now they will be a bit "yellow/green" do to the contrast stretch that these png's have had done to them

Thank you!
Go to the top of the page
 
+Quote Post
htspace
post Mar 12 2021, 01:48 AM
Post #81


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



QUOTE (djellison @ Mar 12 2021, 02:36 AM) *
I must admit - I've found Gmic's debayer performance to be pretty ugly - especially at the edge of images.

I've been using PIPP which can batch debayer images and I've been very pleased with the results

https://sites.google.com/site/astropipp/

I don't know how to perform batch operations, only one picture can be output.can you teach me? Thank you! tongue.gif
Attached Image
Go to the top of the page
 
+Quote Post
htspace
post Mar 12 2021, 01:49 AM
Post #82


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



QUOTE (James Sorenson @ Mar 12 2021, 04:00 AM) *
I'm using Pixinsight which is an astrophotography processing suite, it has a real nice debayer script. I use it for my astrophotography and it is expensive, 300 USD, but for those that can't afford that, the Planetary Image Pre-processor (PIPP) has a great batch debayer function in it which I've also used and it works great. On the plus side it's a free and easy program. smile.gif

Yes, this software is expensive.
I don't know how to perform batch operations, only one picture can be output.can you teach me? Thank you! laugh.gif
Go to the top of the page
 
+Quote Post
James Sorenson
post Mar 12 2021, 03:02 AM
Post #83


Member
***

Group: Members
Posts: 691
Joined: 21-December 07
From: Clatskanie, Oregon
Member No.: 3988



For PIPP, all you need to do is open up all of the images that you are going to debayer, then on the input tab select the option "Debayer Monochrome frames" and the pattern to RGGB. You can also choose the processing algorithm as well if you want VNG or Bilinear works good, you can experiment. To see a preview before final processing, you can click the "Test Options" button in the top right. If it looks good, you can move to the output tab and choose your output file options. Once ready, just go to the "Do Processing" tab and click "Start Processing". All the images that you loaded should be batch debayerd into a new folder.
Go to the top of the page
 
+Quote Post
Greenish
post Mar 12 2021, 04:58 AM
Post #84


Member
***

Group: Members
Posts: 219
Joined: 14-November 11
From: Washington, DC
Member No.: 6237



QUOTE (Greenish @ Mar 9 2021, 10:12 PM) *
If "camera_position" is correct at (2.08195,-0.111422,-1.69411) then WATSON's eyeball would be a fairly high 1.7m off the ground, ...
... Not convinced ...


Well, maybe I'm not totally nuts. These pictures are before & after the WATSON test images. Looks pretty high to me.
Attached Image

Implies roughly the scale bar shown below, so the larger rocks are ~2cm.
Attached Image
Go to the top of the page
 
+Quote Post
htspace
post Mar 12 2021, 06:54 AM
Post #85


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



QUOTE (James Sorenson @ Mar 12 2021, 11:02 AM) *
For PIPP, all you need to do is open up all of the images that you are going to debayer, then on the input tab select the option "Debayer Monochrome frames" and the pattern to RGGB. You can also choose the processing algorithm as well if you want VNG or Bilinear works good, you can experiment. To see a preview before final processing, you can click the "Test Options" button in the top right. If it looks good, you can move to the output tab and choose your output file options. Once ready, just go to the "Do Processing" tab and click "Start Processing". All the images that you loaded should be batch debayerd into a new folder.

Thank you! The previous steps were successful, and finally failed.I don't know how to solve it.
Attached Image
Go to the top of the page
 
+Quote Post
htspace
post Mar 12 2021, 09:17 AM
Post #86


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



QUOTE (Ant103 @ Mar 12 2021, 03:29 AM) *
As for me, I'm using ImageJ with Debayer plugin. Alongside with a batch script, I can do hundred of pictures in a very short time.

Oh it's you! Thank you for your work! cool.gif
How to create a batch script?I am a novice……
Go to the top of the page
 
+Quote Post
htspace
post Mar 12 2021, 10:14 AM
Post #87


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



QUOTE (James Sorenson @ Mar 12 2021, 11:02 AM) *
For PIPP, all you need to do is open up all of the images that you are going to debayer, then on the input tab select the option "Debayer Monochrome frames" and the pattern to RGGB. You can also choose the processing algorithm as well if you want VNG or Bilinear works good, you can experiment. To see a preview before final processing, you can click the "Test Options" button in the top right. If it looks good, you can move to the output tab and choose your output file options. Once ready, just go to the "Do Processing" tab and click "Start Processing". All the images that you loaded should be batch debayerd into a new folder.

I found the reason, it is the system language problem. biggrin.gif
Go to the top of the page
 
+Quote Post
Art Martin
post Mar 12 2021, 03:40 PM
Post #88


Member
***

Group: Members
Posts: 122
Joined: 19-June 07
Member No.: 2455



I just installed PIPP and it seems to be a very full featured program but I'm certainly lacking a great deal of info. I downloaded a Watson image of the underbelly that to my untrained eye seems like it's a candidate for debayering since there are lines and dots across the image that makes it look blurry and unprocessed. However, when processing it through PIPP it didn't seem like it did anything. The output looked pretty much like the input and it certainly was not in color. There were tons of options to choose but unfortunately the examples on the website were for some Jupiter video file not some single image. Any shortcuts someone can point to?

Edit: I did manage to get a color image debayered using Gmic by just trial and error. I'll post it in the helicopter area.
Go to the top of the page
 
+Quote Post
mcaplinger
post Mar 12 2021, 04:54 PM
Post #89


Senior Member
****

Group: Members
Posts: 2517
Joined: 13-September 05
Member No.: 497



QUOTE (Greenish @ Mar 9 2021, 07:12 PM) *
If "camera_position" is correct at (2.08195,-0.111422,-1.69411) then WATSON's eyeball would be a fairly high 1.7m off the ground...

I take it back, perhaps this is correct. Should be possible to find the WATSON footprint in the front Hazcam image.

I've asked if the focus position can be put back into the public data, but we don't have much control over that.


--------------------
Disclaimer: This post is based on public information only. Any opinions are my own.
Go to the top of the page
 
+Quote Post
fredk
post Mar 14 2021, 06:13 PM
Post #90


Senior Member
****

Group: Members
Posts: 4247
Joined: 17-January 05
Member No.: 152



Has anyone figured out how to sort the images at mars.nasa.gov/mars2020/multimedia/raw-images/ by time taken? Selecting a sort in the "sort by" box gives images out of order according to the listed LMST. (Eg, the first few images are out of order when you filter on sol 18.) The corresponding operation on the MSL site works fine.

The corresponding api call
CODE
https://mars.nasa.gov/rss/api/?feed=raw_images&category=mars2020&feedtype=json&num=50&page=0&order=sol+desc&&condition_2=18:sol:gte&condition_3=18:sol:lte&extended=sample_type::full,
similarly gives an out of order json file, as do any alternatives I've tried for the "order=" option.

Maybe this is related to the timestamp problems that are being worked on?
Go to the top of the page
 
+Quote Post
phase4
post Mar 14 2021, 08:17 PM
Post #91


Junior Member
**

Group: Members
Posts: 95
Joined: 11-January 07
From: Amsterdam
Member No.: 1584



QUOTE (fredk @ Mar 14 2021, 07:13 PM) *
Has anyone figured out how to sort the images at mars.nasa.gov/mars2020/multimedia/raw-images/ by time taken?

Not from the website directly. For the arm animation I made earlier I sorted the JSON entries on the spacecraft clock time which is given as "extended/sclk". (and can also be found as part of the image filename)


--------------------
Go to the top of the page
 
+Quote Post
Hynee
post Mar 15 2021, 10:57 AM
Post #92


Newbie
*

Group: Members
Posts: 6
Joined: 20-April 05
Member No.: 265



Does anybody know this plate's purpose?



Here are some other pictures:



Go to the top of the page
 
+Quote Post
PaulH51
post Mar 15 2021, 11:22 AM
Post #93


Senior Member
****

Group: Members
Posts: 2429
Joined: 30-January 13
From: Penang, Malaysia.
Member No.: 6853



QUOTE (Hynee @ Mar 15 2021, 06:57 PM) *
Does anybody know this plate's purpose?

It's a 'Bellypan' that protected the Adaptive Caching Assembly (ACA) since launch, but especially during the final part of EDL...

Surface operations on Mars require the Sample Handling Assembly (SHA) to extend approximately 200 mm (~8 inches) below the Rover’s bellypan. Therefore, an ejectable belly-pan was implemented directly below the Adaptive Caching Assembly (ACA) volume, which was released after landing to provide the SHA with an unobstructed volume to extend into during operations. Surface features are then assessed via Rover imaging prior to SHA motion to prevent contact with potential obstacles below the Rover.

Extracted from this PDF titled: Mars 2020 Rover Adaptive Caching Assembly: Caching Martian Samples for Potential Earth Return
Go to the top of the page
 
+Quote Post
Hynee
post Mar 15 2021, 11:25 AM
Post #94


Newbie
*

Group: Members
Posts: 6
Joined: 20-April 05
Member No.: 265



QUOTE (PaulH51 @ Mar 15 2021, 10:22 PM) *
It's a 'Bellypan' that protected the Adaptive Caching Assembly (ACA) since launch, but especially during the final part of EDL...

Surface operations on Mars require the Sample Handling Assembly (SHA) to extend approximately 200 mm (~8 inches) below the Rover’s bellypan. Therefore, an ejectable belly-pan was implemented directly below the Adaptive Caching Assembly (ACA) volume, which was released after landing to provide the SHA with an unobstructed volume to extend into during operations. Surface features are then assessed via Rover imaging prior to SHA motion to prevent contact with potential obstacles below the Rover.

Extracted from this PDF titled: Mars 2020 Rover Adaptive Caching Assembly: Caching Martian Samples for Potential Earth Return

Thanks!
Go to the top of the page
 
+Quote Post
ChrisC
post Mar 15 2021, 09:19 PM
Post #95


Junior Member
**

Group: Members
Posts: 59
Joined: 4-July 08
Member No.: 4251



QUOTE (Hynee @ Mar 15 2021, 05:57 AM) *
Does anybody know this plate's purpose?

FYI this event is discussed in the regular thread:
http://www.unmannedspaceflight.com/index.p...=8608&st=60

I think this thread here is supposed to be focused on imagery, e.g. technical discussion of processing, cameras, metadata etc.

ADMIN: Chris is correct. Let's please post mission event pictures such as the panel deployment in the real-time progress threads (Perseverance Early Drives being the current one) and reserve this thread for technical discussion of the imagery systems and methods.
Go to the top of the page
 
+Quote Post
htspace
post Mar 16 2021, 12:02 PM
Post #96


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



After I installed the required plugins,how to perform batch debayer in ImageJ and GIMP software?

Need an installation script, right? How to do it?
unsure.gif
Go to the top of the page
 
+Quote Post
Adam Hurcewicz
post Mar 16 2021, 12:39 PM
Post #97


Member
***

Group: Members
Posts: 102
Joined: 29-January 10
From: Poland
Member No.: 5205



I use FitsWork to debayer images. It's great!


--------------------
Adam Hurcewicz from Poland
Go to the top of the page
 
+Quote Post
htspace
post Mar 16 2021, 01:45 PM
Post #98


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



QUOTE (Adam Hurcewicz @ Mar 16 2021, 08:39 PM) *
I use FitsWork to debayer images. It's great!

Can it be processed in batches? How to do it? tongue.gif
Attached Image
Go to the top of the page
 
+Quote Post
PDP8E
post Mar 16 2021, 11:39 PM
Post #99


Member
***

Group: Members
Posts: 808
Joined: 10-October 06
From: Maynard Mass USA
Member No.: 1241



use the command-line interface to your tool (if it has one) and a batch file.


--------------------
CLA CLL
Go to the top of the page
 
+Quote Post
htspace
post Mar 17 2021, 12:13 AM
Post #100


Junior Member
**

Group: Members
Posts: 20
Joined: 6-August 20
Member No.: 8852



QUOTE (PDP8E @ Mar 17 2021, 07:39 AM) *
use the command-line interface to your tool (if it has one) and a batch file.

I get it, it's great! biggrin.gif
Go to the top of the page
 
+Quote Post

12 Pages V   1 2 3 > » 
Reply to this topicStart new topic

 



RSS Lo-Fi Version Time is now: 23rd May 2024 - 12:51 PM
RULES AND GUIDELINES
Please read the Forum Rules and Guidelines before posting.

IMAGE COPYRIGHT
Images posted on UnmannedSpaceflight.com may be copyrighted. Do not reproduce without permission. Read here for further information on space images and copyright.

OPINIONS AND MODERATION
Opinions expressed on UnmannedSpaceflight.com are those of the individual posters and do not necessarily reflect the opinions of UnmannedSpaceflight.com or The Planetary Society. The all-volunteer UnmannedSpaceflight.com moderation team is wholly independent of The Planetary Society. The Planetary Society has no influence over decisions made by the UnmannedSpaceflight.com moderators.
SUPPORT THE FORUM
Unmannedspaceflight.com is funded by the Planetary Society. Please consider supporting our work and many other projects by donating to the Society or becoming a member.