Stowage is a bloat-free .NET cloud storage kit that supports at minimum THE major β providers.
- Independent π. Provides an independent implementation of the β storage APIs. Because you can't just have official corporate SDKs as a single source of truth.
- Readable. Official SDKs like the ones for AWS, Google, or Azure are overengineered and unreadable. Some are autogenerated and look just bad and foreign to .NET ecosystem. Some won't even compile without some custom rituals.
- Beautiful π¦. Designed to fit into .NET ecosystem, not the other way around.
- Rich π°. Provide maximum functionality. However, in addition to that, provide humanly possible way to easily extend it with new functionality, without waiting for new SDK releases.
- Embeddable π±. Has zero external dependencies, relies only on built-in .NET API. Often official SDKs have a very deep dependency tree causing a large binary sizes and endless conflicts during runtime. This one is a single .NET .dll with no dependencies whatsoever.
- Cross Cloud π₯. Same API. Any cloud. Best decisions made for you. It's like iPhone vs Windows Phone.
- Cross Tested β. It's not just cross cloud but also cross tested (I don't know how to call this). It tests that all cloud providers behave absolutely the same on various method calls. They should validate arguments the same, throw same exceptions in the same situations, and support the same set of functionality. Sounds simple, but it's rare to find in a library. And it important, otherwise what's the point of a generic API if you need to write a lot of
if()
s? (or pattern matching).
This library originally came out from being frustrated on working on my another library - Storage.Net. While it's OK, most of the time I had to deal with SDK incompatibilities, breaking changes, oddnesses, and slowness, whereas most of the time users needs something simple that just works.
Right, time to gear up. We'll do it step by step. First, you need to install the package.
Simplest case, using the local π½ and writing text "I'm a page!!!" to a file called "pagefile.sys" at the root of disk C::
using Stowage;
using(IFileStorage fs = Files.Of.LocalDisk("c:\\"))
{
await fs.WriteText("pagefile.sys", "I'm a page!!!!");
}
This is local disk, yeah? But what about cloud storage, like Azure Blob Storage? Piece of cake:
using Stowage;
using(IFileStorage fs = Files.Of.AzureBlobStorage("accountName", "accountKey", "containerName"))
{
var entries = await fs.Ls();
}
Streaming is a first-class feature. This means the streaming is real with no workarounds or in-memory buffers, so you can upload/download files of virtually unlimited sizes. Most official SDKs do not support streaming at all - surprisingly even the cloud leader's .NET SDK doesn't. Each requires some sort of crippled down version of stream - either knowing length beforehand, or will buffer it all in memory. I don't. I stream like a stream.
Proper streaming support also means that you can transform streams as you write to them or read from them - something that is not available in the native SDKs. For instance gzipping, encryption, anything else.
Streaming is also truly compatible with synchronous and asynchronous API.
Whenever a method appears here, I assume it belongs to IFileStorage
interface, unless specified.
Use .Ls()
(short for list) - very easy to remember! Everyone knows what ls does, right? Optionally allows to list entries recursively.
The core method for reading is Stream OpenRead(IOPath path)
- this returns a stream from file path. Stream is the lowest level data structure. There are other helper methods that by default rely on this method, like ReadText
etc. Just have a quick look:
IFileStorage fs = ...;
Stream target = ...;
// copy to another stream
using Stream s = await fs.OpenRead("/myfile.txt");
// synchronous copy:
s.CopyTo(target);
// or alternatively, asynchronous copy (preferred):
await s.CopyToAsync(target);
// if you just need text:
string content = await fs.ReadText("/myfile.txt");
Of course there are more overloaded methods you can take advantage of.
The main method Stream OpenWrite(IOPath path, ...)
opens(/creates?) a file for writing. It returns a real writeable stream you can write to and close afterwards. It behaves like a stream and is a stream.
There are other overloads which support writing text etc.
Rm(IOPath path)
trashes files or folders (or both) with options to do it recursively!
There are other useful utility methods:
bool Exists(IOPath path)
that checks for file existence. It supposed to be really efficient, hence a separate method.Ren
renames files and folders.- and more are coming - check
IFileStorage
interface to be up to date.
- Local Disk Directory (
Files.Of.LocalDisk(...)
). - In-Memory (
Files.Of.InternalMemory(...)
). - AWS S3 (
Files.Of.AmazonS3(...)
). - Azure Blob Storage / Data Lake Gen 2 (
Files.Of.AzureBlobStorage(...)
). - Google Cloud Storage (
Files.Of.GoogleCloudStorage(...)
). - Databricks DBFS (
Files.Of.DatabricksDbfs(...)
).
Instantiation instructions are in the code documentation (IntelliSense?) - I prefer this to writing out here locally.
There are many ways to extend functionality:
- Documentation. You might think it's not extending anything, however if user is not aware for some functionality it doesn't exist. Documenting it is making it available, hence extending. You must be slightly mad to follow my style of writing though.
- New functionality. Adding utility methods like copying files inside or between accounts, automatic JSON serialisation etc. is always good. Look
IFileStorage
interface andPolyfilledFileStorage
. In most cases these two files are enough to add pure business logic. Not counting unit tests. Which you must write. Otherwise it's easier to do the whole thing by myself. Which is what will happen according to my experience. - Native optimisations. Some functionality is generic, and some depends on a specific cloud provider. For instance, one can copy a file by downloading it locally, and uploading with a new name. Or utilise a native REST call that accepts source and target file name, if it exists. Involves digging deeper into specific provider's API.
When contributing a new provider, it's way more preferrable to embed it's code in the library, provided that:
- there are no extra nuget dependencies.
- it's cross-platform.
I'm a strong opponent of simplicity and not going to repeat the mistake of turning this into a nuget tree dependency hell!
- Used by:
- databricks-sql-cli - Unofficial Databricks SQL management console.
- Pocket Bricks - Databricks client for Android.
- Featured in The .NET MAUI Podcast, episode 98.
- RCLONE - cross-platform open-source cloud sync tool.
- Storage.Net - the roots of this project.
You are welcome to contribute in any form, however I wouldn't bother, especially financially. Don't bother buying me a β, I can do it myself real cheap during COVID! Why? During my years of OSS development everyone I know (including myself) have only lost money. Why I'm still doing this? Probably because it's just cool and I'm enjoying it.