This documentation is for Stowage v2 which is a major redesign. Version 1 documentation can be found here.
Stowage is a bloat-free .NET cloud storage kit that supports at minimum THE major β providers.
- Independent π. Provides an independent implementation of the β storage APIs. Because you can't just have official corporate SDKs as a single source of truth.
- Readable. Official SDKs like the ones for AWS, Google, or Azure are overengineered and unreadable. Some are autogenerated and look just bad and foreign to .NET ecosystem. Some won't even compile without some custom rituals.
- Beautiful π¦. Designed to fit into .NET ecosystem, not the other way around.
- Rich π°. Provide maximum functionality. However, in addition to that, provide humanly possible way to easily extend it with new functionality, without waiting for new SDK releases.
- Embeddable π±. Has zero external dependencies, relies only on built-in .NET API. Often official SDKs have a very deep dependency tree causing a large binary sizes and endless conflicts during runtime. This one is a single .NET .dll with no dependencies whatsoever.
- Cross Cloud π₯. Same API. Any cloud. Best decisions made for you. It's like iPhone vs Windows Phone.
- Cross Tested β. It's not just cross cloud but also cross tested (I don't know how to call this). It tests that all cloud providers behave absolutely the same on various method calls. They should validate arguments the same, throw same exceptions in the same situations, and support the same set of functionality. Sounds simple, but it's rare to find in a library. And it important, otherwise what's the point of a generic API if you need to write a lot of
if()
s? (or pattern matching).
This library originally came out from being frustrated on working on my another library - Storage.Net. While it's OK, most of the time I had to deal with SDK incompatibilities, breaking changes, oddnesses, and slowness, whereas most of the time users needs something simple that just works.
Right, time to gear up. We'll do it step by step. First, you need to install the package.
Simplest case, using the local π½ and writing text "I'm a page!!!" to a file called "pagefile.sys" at the root of disk C::
using Stowage;
using(IFileStorage fs = Files.Of.LocalDisk("c:\\")) {
await fs.WriteText("pagefile.sys", "I'm a page!!!!");
}
This is local disk, yeah? But what about cloud storage, like Azure Blob Storage? Piece of cake:
using Stowage;
using(IFileStorage fs = Files.Of.AzureBlobStorage("accountName", "accountKey", "containerName")) {
var entries = await fs.Ls();
}
Streaming is a first-class feature. This means the streaming is real with no workarounds or in-memory buffers, so you can upload/download files of virtually unlimited sizes. Most official SDKs do not support streaming at all - surprisingly even the cloud leader's .NET SDK doesn't. Each requires some sort of crippled down version of stream - either knowing length beforehand, or will buffer it all in memory. I don't. I stream like a stream.
Proper streaming support also means that you can transform streams as you write to them or read from them - something that is not available in the native SDKs. For instance gzipping, encryption, anything else.
Streaming is also truly compatible with synchronous and asynchronous API.
Whenever a method appears here, I assume it belongs to IFileStorage
interface, unless specified.
Use .Ls()
(short for list) - very easy to remember! Everyone knows what ls does, right? Optionally allows to list entries recursively.
The core method for reading is Stream OpenRead(IOPath path)
- this returns a stream from file path. Stream is the lowest level data structure. There are other helper methods that by default rely on this method, like ReadText
etc. Just have a quick look:
IFileStorage fs = ...;
Stream target = ...;
// copy to another stream
using Stream s = await fs.OpenRead("/myfile.txt");
// synchronous copy:
s.CopyTo(target);
// or alternatively, asynchronous copy (preferred):
await s.CopyToAsync(target);
// if you just need text:
string content = await fs.ReadText("/myfile.txt");
Of course there are more overloaded methods you can take advantage of.
The main method Stream OpenWrite(IOPath path, ...)
opens(/creates?) a file for writing. It returns a real writeable stream you can write to and close afterwards. It behaves like a stream and is a stream.
There are other overloads which support writing text etc.
Rm(IOPath path)
trashes files or folders (or both) with options to do it recursively!
There are other useful utility methods:
bool Exists(IOPath path)
that checks for file existence. It supposed to be really efficient, hence a separate method.Ren
renames files and folders.- and more are coming - check
IFileStorage
interface to be up to date.
-
Local Disk Directory (
Files.Of.LocalDisk(...)
). -
In-Memory (
Files.Of.InternalMemory(...)
). -
AWS S3 (
Files.Of.AmazonS3(...)
).- Minio (
Files.Of.Minio(...)
). - DigitalOcean Spaces (
Files.Of.DigitalOceanSpaces(...)
).
- Minio (
-
Azure Blob Storage / Data Lake Gen 2 (
Files.Of.AzureBlobStorage(...)
). -
Google Cloud Storage (
Files.Of.GoogleCloudStorage(...)
). -
Databricks DBFS (
Files.Of.DatabricksDbfs(...)
).
Instantiation instructions are in the code documentation (IntelliSense?) - I prefer this to writing out here locally.
Below are some details worth mentioning.
It's just what it says - local disk, which by default maps to an entire filesystem. In Windows, which has drive letters, it will map to an entire disk of where the application's current directory's drive is (i.e. if you are in c:/my/app
it it will map to c:/
).
IFileStorage storage = Files.Of.LocalDisk();
// or
IFileStorage storage = Storage.Of.ConnectionString("disk://")
Optionally, you can specify the root directory:
IFileStorage storage = Files.Of.LocalDisk("/a/folder");
// or
IFileStorage storage = Storage.Of.ConnectionString("disk://path=/a/folder")
do not forget there is
path
keyword in the connection string, it's so tempting to write "disk://a/folder" instead of "disk://path=/a/folder"!
In AWS, the path addressing style is the following:
/bucket/path/object
Ls
on the root folder returns list of buckets in the AWS account, whether you do have access to them or not.
The most usual way to authenticate with S3 is to use the following method:
IFileStorage storage = Files.Of.AmazonS3(key, secret, region);
// or
IFileStorage storage = Storage.Of.ConnectionString("s3://keyId=<key>;key=<secret>;region=<region>")
These are what Amazon calls "long-term" credentials. If you are using STS, the same method overload allows you to pass sessionToken
.
IFileStorage storage = Files.Of.AmazonS3(key, secret, region, sessionToken);
//or
IFileStorage storage = Storage.Of.ConnectionString("s3://keyId=<key>;key=<secret>;region=<region>;sessionToken=<session token>")
Another way to authenticate is using CLI profile. This is useful when you machine is already authenticated using aws cli, awsume or similar tools that write credentials and configuration to ~/.aws/credentials
and ~/.aws/config
.
You only need to pass the profile name (and only if it's not a default one):
// ---- using default profile ----
IFileStorage storage = Files.Of.AmazonS3FromCliProfile();
//or
IFileStorage storage = Storage.Of.ConnectionString("s3://");
// ----- using specific profile, like "myprofile" ----
IFileStorage storage = Files.Of.AmazonS3FromCliProfile("myprofile");
//or
IFileStorage storage = Storage.Of.ConnectionString("s3://profile=myprofile");
This method has other default parameters, such as regionName
which can be specified or overridden if not found in CLI configuration, i.e. Files.Of.AmazonS3FromCliProfile();
has optional region
parameters, and connection string has optional region=
keyword.
Minio is essentially using the standard S3 protocol, but addressing style is slightly different. There is a helper extension that somewhat simplifies Minio authentication:
IFileStorage storage = Files.Of.Minio(endpoint, key, secret);
In Azure Blob Storage, path addressing style is the following:
/container/path/object
Note that there is no storage account in the path, mostly because Shared Key authentication is storage account scoped, not tenant scoped.
Ls
on the root folder returns list of containers in the storage account.
Azure provider supports authentication with Shared Key:
IFileStorage storage = Files.Of.AzureBlobStorage(accountName, sharedKey);
// or
IFileStorage storage = Storage.Of.ConnectionString("az://account=<account name>;key=<shared key>")
IFileStorage storage = Files.Of.AzureBlobStorage(
accountName,
new ClientSecretCredential(tenantId, clientId, clientSecret));
Managed identities are not yet supported due to low demand, but watch this space.
Azure emulator is supported, just use AzureBlobStorageWithLocalEmulator()
method to connect to it.
This storage essentially wraps around another storage to provide content caching capabilities. Example:
IFileStorage storage = Files.Of.AzureBlobStorage(accountName, sharedKey);
IFileStorage cachedStorage = Files.Of.LocalDiskCacheStorage(storage);
When using cachedStorage
, all the operations are forwarded to storage
as is, except for OpenRead
which downloads content locally and opens a stream to the local file.
There are many ways to extend functionality:
- Documentation. You might think it's not extending anything, however if user is not aware for some functionality it doesn't exist. Documenting it is making it available, hence extending. You must be slightly mad to follow my style of writing though.
- New functionality. Adding utility methods like copying files inside or between accounts, automatic JSON serialisation etc. is always good. Look
IFileStorage
interface andPolyfilledFileStorage
. In most cases these two files are enough to add pure business logic. Not counting unit tests. Which you must write. Otherwise it's easier to do the whole thing by myself. Which is what will happen according to my experience. - Native optimisations. Some functionality is generic, and some depends on a specific cloud provider. For instance, one can copy a file by downloading it locally, and uploading with a new name. Or utilise a native REST call that accepts source and target file name, if it exists. Involves digging deeper into specific provider's API.
When contributing a new provider, it's way more preferrable to embed it's code in the library, provided that:
- there are no extra nuget dependencies.
- it's cross-platform.
I'm a strong advocate of simplicity and not going to repeat the mistake of turning this into a nuget tree dependency hell!
- Featured in The .NET MAUI Podcast, episode 98.
- Blog post Exploring Stowage.
Raise a PR to appear here.
- RCLONE - cross-platform open-source cloud sync tool.
- Storage.Net - the roots of this project.
You are welcome to contribute in any form, however I wouldn't bother, especially financially. Don't bother buying me a β, I can do it myself real cheap. During my years of OSS development everyone I know (including myself) have only lost money. Why I'm still doing this? Probably because it's just cool and I'm enjoying it.