Conversation
Codecov Report✅ All modified and coverable lines are covered by tests. Additional details and impacted files@@ Coverage Diff @@
## main #81 +/- ##
=======================================
Coverage 71.93% 71.93%
=======================================
Files 11 11
Lines 2423 2423
Branches 364 364
=======================================
Hits 1743 1743
Misses 583 583
Partials 97 97 ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
|
@davidhassell @bnlawrence as you folks saw on last Thu, we do have a few bits and bobs on docs for Pyfive (also, it's all nicely deployed to RTD), but this PR is here to add more - pls feel free to edit and contribute via this PR when you folks have a minute 🍺 |
kmuehlbauer
left a comment
There was a problem hiding this comment.
I've just had a look here. Thanks @valeriupredoi for these efforts in documentation.
Co-authored-by: Kai Mühlbauer <kmuehlbauer@wradlib.org>
…into add_documentation
…including showing the code works).
|
@valeriupredoi Apart from the failing check (dunno why, it builds locally) I think this is close to a minimal set of useable documentation. It'd be good to check my parallel example actually works, but apart from that, I think we'd be good to go. Please have a good look at this now. |
mega! Very cool @bnlawrence - RTD is fussing about these type of warnings (a few of them): I can fix those tomorrow 🍺 |
valeriupredoi
left a comment
There was a problem hiding this comment.
some minor comments, but a bigger one includes fixing the S3 example 🍺
Co-authored-by: Valeriu Predoi <valeriu.predoi@gmail.com>
Co-authored-by: Valeriu Predoi <valeriu.predoi@gmail.com>
Co-authored-by: Valeriu Predoi <valeriu.predoi@gmail.com>
Co-authored-by: Valeriu Predoi <valeriu.predoi@gmail.com>
Co-authored-by: Valeriu Predoi <valeriu.predoi@gmail.com>
Co-authored-by: Valeriu Predoi <valeriu.predoi@gmail.com>
Yes, not sure the caching is understood by botocore, I had conflated a couple of different things, best to be explicit that we are providing params for s3fs itself. Co-authored-by: Valeriu Predoi <valeriu.predoi@gmail.com>
Co-authored-by: Valeriu Predoi <valeriu.predoi@gmail.com>
Co-authored-by: Valeriu Predoi <valeriu.predoi@gmail.com>
Co-authored-by: Valeriu Predoi <valeriu.predoi@gmail.com>
…ather than just the original docstring copied from h5py during implementation.
| return data.min() | ||
|
|
||
| with ThreadPoolExecutor() as executor: | ||
| results = list(executor.map(get_min_of_variable, variable_names)) |
There was a problem hiding this comment.
this works! But - I have not noticed any time improvements, it's very possible since I had to cut the size of the data because my RAM is not big enough, so what I was loading would be as fast, if not faster, in single thread process
There was a problem hiding this comment.
As discussed, I'm slightly concerned by your RAM problems, it's probably ok to leave this (contrived) example in the docs as all our real examples are too complex to use as examplars, but we should do some due diligence on why this is happening.
There was a problem hiding this comment.
this works, as I said - the issues I have with the buffer could well be because I am running out of both RAM and actual disk, so those may be very user-specific, that's why I said I need to look a lot closer at it
|
superb work @bnlawrence @kmuehlbauer @davidhassell 🍺 x 3 |
kmuehlbauer
left a comment
There was a problem hiding this comment.
Very neat, very nice docs. I've just went through it, added some suggestions.
| The data storage complexities arise from two main factors: the use of chunking, and the way attributes are stored in the files | ||
|
|
There was a problem hiding this comment.
| The data storage complexities arise from two main factors: the use of chunking, and the way attributes are stored in the files | |
| The data storage complexities arise from two main factors: the use of chunking, and the way attributes are stored in the files. | |
| ------------------------------- | ||
|
|
||
| Optimal access to data occurs when the data is chunked in a way that matches the access patterns of your application, and when the | ||
| b-tree indexes and attributess are stored contiguously in the file. |
There was a problem hiding this comment.
| b-tree indexes and attributess are stored contiguously in the file. | |
| b-tree indexes and attributes are stored contiguously in the file. |
| import pyfive | ||
|
|
||
| with pyfive.File("data.h5", "r") as f: | ||
| variables = [f for var in f] |
There was a problem hiding this comment.
| variables = [f for var in f] | |
| variables = [var for var in f] |
| print("Results:", results) | ||
|
|
||
|
|
||
| You can do the same thing to parallelise manipuations within the variables, by for example using, ``Dask``, but that is beyond the scope of this document. |
There was a problem hiding this comment.
| You can do the same thing to parallelise manipuations within the variables, by for example using, ``Dask``, but that is beyond the scope of this document. | |
| You can do the same thing to parallelise manipulations within the variables, by for example using ``Dask``, but that is beyond the scope of this document. |
| return self._index | ||
| #### The following method can be used to set pseudo chunking size after the | ||
| #### file has been closed and before data transactions. This is pyfive specific | ||
| def set_psuedo_chunk_size(self, newsize_MB): |
There was a problem hiding this comment.
| def set_psuedo_chunk_size(self, newsize_MB): | |
| def set_pseudo_chunk_size(self, newsize_MB): |
There was a problem hiding this comment.
There might be more occurences?
There was a problem hiding this comment.
Sadly you're probably right, but we are hoping to do a wee internal hackathon in autumn, where we might deal with both enums and our test coverage ... so if they're not picked up before then, hopefully proper coverage tests will find the rest.
|
Ah, missed by 7 minutes ;-) Please have a look at my comments, whether or not they are useful. |
|
@valeriupredoi Can we add these in please? |
|
thanks @kmuehlbauer - I'll pop those into a new PR I'll open now 🍺 |
Description
We now have a fully working Readthedocs setup with a doc stub that builds well; it's time to add actual documentation. I have created the Install section so far, but we need more stuffs.
Checklist